Federal Register 2010, 2011, 2012, 2013, 2014
2011-12-01
... Statistics Relating to Competitive Need Limitations AGENCY: Office of the United States Trade Representative. ACTION: Notice. SUMMARY: This notice is to inform the public of the availability of import statistics for... System of Preferences (GSP) program. These import statistics identify some articles for which the 2011...
15 CFR 30.51 - Statistical information required for import entries.
Code of Federal Regulations, 2011 CFR
2011-01-01
... 15 Commerce and Foreign Trade 1 2011-01-01 2011-01-01 false Statistical information required for import entries. 30.51 Section 30.51 Commerce and Foreign Trade Regulations Relating to Commerce and... § 30.51 Statistical information required for import entries. The information required for statistical...
15 CFR 30.51 - Statistical information required for import entries.
Code of Federal Regulations, 2010 CFR
2010-01-01
... 15 Commerce and Foreign Trade 1 2010-01-01 2010-01-01 false Statistical information required for import entries. 30.51 Section 30.51 Commerce and Foreign Trade Regulations Relating to Commerce and... § 30.51 Statistical information required for import entries. The information required for statistical...
Self-Regulated Learning Strategies in Relation with Statistics Anxiety
ERIC Educational Resources Information Center
Kesici, Sahin; Baloglu, Mustafa; Deniz, M. Engin
2011-01-01
Dealing with students' attitudinal problems related to statistics is an important aspect of statistics instruction. Employing the appropriate learning strategies may have a relationship with anxiety during the process of statistics learning. Thus, the present study investigated multivariate relationships between self-regulated learning strategies…
Determining the Statistical Significance of Relative Weights
ERIC Educational Resources Information Center
Tonidandel, Scott; LeBreton, James M.; Johnson, Jeff W.
2009-01-01
Relative weight analysis is a procedure for estimating the relative importance of correlated predictors in a regression equation. Because the sampling distribution of relative weights is unknown, researchers using relative weight analysis are unable to make judgments regarding the statistical significance of the relative weights. J. W. Johnson…
Definitions and Models of Statistical Literacy: A Literature Review
ERIC Educational Resources Information Center
Sharma, Sashi
2017-01-01
Despite statistical literacy being relatively new in statistics education research, it needs special attention as attempts are being made to enhance the teaching, learning and assessing of this sub-strand. It is important that teachers and researchers are aware of the challenges of teaching this literacy. In this article, the growing importance of…
Is GAISE Evident? College Students' Perceptions of Statistics Classes as "Almost Not Math"
ERIC Educational Resources Information Center
Hedges, Sarai; Harkness, Shelly Sheats
2017-01-01
The connection between mathematics and statistics is an important aspect in understanding college students' learning of statistics because studies have shown relationships among mathematics attitudes and performance and statistics attitudes. Statistics attitudes, in turn, are related to performance in statistics courses. Little research has been…
Fostering Self-Concept and Interest for Statistics through Specific Learning Environments
ERIC Educational Resources Information Center
Sproesser, Ute; Engel, Joachim; Kuntze, Sebastian
2016-01-01
Supporting motivational variables such as self-concept or interest is an important goal of schooling as they relate to learning and achievement. In this study, we investigated whether specific interest and self-concept related to the domains of statistics and mathematics can be fostered through a four-lesson intervention focusing on statistics.…
ERIC Educational Resources Information Center
Sullivan, Sharon G.; Grabois, Andrew; Greco, Albert N.
2003-01-01
Includes six reports related to book trade statistics, including prices of U.S. and foreign materials; book title output and average prices; book sales statistics; book exports and imports; book outlets in the U.S. and Canada; and numbers of books and other media reviewed by major reviewing publications. (LRW)
2013-05-02
REPORT Statistical Relational Learning ( SRL ) as an Enabling Technology for Data Acquisition and Data Fusion in Video 14. ABSTRACT 16. SECURITY...particular, it is important to reason about which portions of video require expensive analysis and storage. This project aims to make these...inferences using new and existing tools from Statistical Relational Learning ( SRL ). SRL is a recently emerging technology that enables the effective 1
Study Designs and Statistical Analyses for Biomarker Research
Gosho, Masahiko; Nagashima, Kengo; Sato, Yasunori
2012-01-01
Biomarkers are becoming increasingly important for streamlining drug discovery and development. In addition, biomarkers are widely expected to be used as a tool for disease diagnosis, personalized medication, and surrogate endpoints in clinical research. In this paper, we highlight several important aspects related to study design and statistical analysis for clinical research incorporating biomarkers. We describe the typical and current study designs for exploring, detecting, and utilizing biomarkers. Furthermore, we introduce statistical issues such as confounding and multiplicity for statistical tests in biomarker research. PMID:23012528
Sajobi, Tolulope T; Lix, Lisa M; Singh, Gurbakhshash; Lowerison, Mark; Engbers, Jordan; Mayo, Nancy E
2015-03-01
Response shift (RS) is an important phenomenon that influences the assessment of longitudinal changes in health-related quality of life (HRQOL) studies. Given that RS effects are often small, missing data due to attrition or item non-response can contribute to failure to detect RS effects. Since missing data are often encountered in longitudinal HRQOL data, effective strategies to deal with missing data are important to consider. This study aims to compare different imputation methods on the detection of reprioritization RS in the HRQOL of caregivers of stroke survivors. Data were from a Canadian multi-center longitudinal study of caregivers of stroke survivors over a one-year period. The Stroke Impact Scale physical function score at baseline, with a cutoff of 75, was used to measure patient stroke severity for the reprioritization RS analysis. Mean imputation, likelihood-based expectation-maximization imputation, and multiple imputation methods were compared in test procedures based on changes in relative importance weights to detect RS in SF-36 domains over a 6-month period. Monte Carlo simulation methods were used to compare the statistical powers of relative importance test procedures for detecting RS in incomplete longitudinal data under different missing data mechanisms and imputation methods. Of the 409 caregivers, 15.9 and 31.3 % of them had missing data at baseline and 6 months, respectively. There were no statistically significant changes in relative importance weights on any of the domains when complete-case analysis was adopted. But statistical significant changes were detected on physical functioning and/or vitality domains when mean imputation or EM imputation was adopted. There were also statistically significant changes in relative importance weights for physical functioning, mental health, and vitality domains when multiple imputation method was adopted. Our simulations revealed that relative importance test procedures were least powerful under complete-case analysis method and most powerful when a mean imputation or multiple imputation method was adopted for missing data, regardless of the missing data mechanism and proportion of missing data. Test procedures based on relative importance measures are sensitive to the type and amount of missing data and imputation method. Relative importance test procedures based on mean imputation and multiple imputation are recommended for detecting RS in incomplete data.
A Bifactor Approach to Model Multifaceted Constructs in Statistical Mediation Analysis
ERIC Educational Resources Information Center
Gonzalez, Oscar; MacKinnon, David P.
2018-01-01
Statistical mediation analysis allows researchers to identify the most important mediating constructs in the causal process studied. Identifying specific mediators is especially relevant when the hypothesized mediating construct consists of multiple related facets. The general definition of the construct and its facets might relate differently to…
Book Trade Research and Statistics.
ERIC Educational Resources Information Center
Sullivan, Sharon G.; Ink, Gary; Grabois, Andrew; Barr, Catherine
2001-01-01
Includes six articles that discuss research and statistics relating to the book trade. Topics include prices of U.S. and foreign materials; book title output and average prices; book sales statistics; book exports and imports; book outlets in the U.S. and Canada; and books and other media reviewed. (LRW)
A Study on the Application of Creative Problem Solving Teaching to Statistics Teaching
ERIC Educational Resources Information Center
Hu, Ridong; Xiaohui, Su; Shieh, Chich-Jen
2017-01-01
Everyone would encounter the life issue of solving complicated problems generated by economic behaviors among all activities for making a living. Various life problems encountered therefore could be generalized by economic statistics. In other words, a lot of important events in daily life are related to economic statistics. For this reason,…
The Relative Importance of Low Significance Level and High Power in Multiple Tests of Significance.
ERIC Educational Resources Information Center
Westermann, Rainer; Hager, Willi
1983-01-01
Two psychological experiments--Anderson and Shanteau (1970), Berkowitz and LePage (1967)--are reanalyzed to present the problem of the relative importance of low Type 1 error probability and high power when answering a research question by testing several statistical hypotheses. (Author/PN)
Temporal distribution of alcohol related facial fractures.
Lee, Kai H; Qiu, Michael; Sun, Jiandong
2017-11-01
This study aimed to address 2 important aspects of temporal pattern in alcohol-related facial fractures: (1) comparison of temporal pattern of alcohol-related facial fracture (alcohol group) presentation with non-alcohol-related fracture (non-alcohol group) presentation; (2) temporal pattern of patient demographic characteristics, injury characteristics, and surgical management in the alcohol group presentation. This study retrospectively examined the Victorian admitted episodes data set (VAED) for the years 2010 to 2013. VAED is a standardized set of data collected during all hospital presentations in Victoria. The study found higher incidence of alcohol-related facial fracture presentations during weekends and during the summer and spring months compared with non-alcohol-related fractures (statistically significant). Alcohol-related facial fractures are more likely to involve male patients in the 20- to 29-year age group, occur as a result of interpersonal violence, and require shorter hospital stays during weekend admissions (statistically significant). No statistically significant relationship has been observed in seasonal variation across all variables. This study found distinct characteristics in temporal distribution of alcohol-related facial fractures. These characteristics are, in particular, significant in weekend trauma admissions. Such information is important in workforce planning, resource distribution, and implementation of injury prevention programs. Copyright © 2017 Elsevier Inc. All rights reserved.
Gòmez, Miguel-Ángel; Lorenzo, Alberto; Ortega, Enrique; Sampaio, Jaime; Ibàñez, Sergio-José
2009-01-01
The aim of the present study was to identify the game-related statistics that allow discriminating between starters and nonstarter players in women’s basketball when related to winning or losing games and best or worst teams. The sample comprised all 216 regular season games from the 2005 Women’s National Basketball Association League (WNBA). The game-related statistics included were 2- and 3- point field-goals (both successful and unsuccessful), free-throws (both successful and unsuccessful), defensive and offensive rebounds, assists, blocks, fouls, steals, turnovers and minutes played. Results from multivariate analysis showed that when best teams won, the discriminant game-related statistics were successful 2-point field-goals (SC = 0.47), successful free-throws (SC = 0.44), fouls (SC = -0.41), assists (SC = 0.37), and defensive rebounds (SC = 0.37). When the worst teams won, the discriminant game-related statistics were successful 2-point field- goals (SC = 0.37), successful free-throws (SC = 0.45), assists (SC = 0.58), and steals (SC = 0.35). The results showed that the successful 2-point field-goals, successful free-throws and the assists were the most powerful variables discriminating between starters and nonstarters. These specific characteristics helped to point out the importance of starters’ players shooting and passing ability during competitions. Key points The players’ game-related statistical profile varied according to team status, game outcome and team quality in women’s basketball. The results of this work help to point out the different player’s performance described in women’s basketball compared with men’s basketball. The results obtained enhance the importance of starters and nonstarters contribution to team’s performance in different game contexts. Results showed the power of successful 2-point field-goals, successful free-throws and assists discriminating between starters and nonstarters in all the analyses. PMID:24149538
Johnson, Karen A.
2013-01-01
Background and Aims Convergent floral traits hypothesized as attracting particular pollinators are known as pollination syndromes. Floral diversity suggests that the Australian epacrid flora may be adapted to pollinator type. Currently there are empirical data on the pollination systems for 87 species (approx. 15 % of Australian epacrids). This provides an opportunity to test for pollination syndromes and their important morphological traits in an iconic element of the Australian flora. Methods Data on epacrid–pollinator relationships were obtained from published literature and field observation. A multivariate approach was used to test whether epacrid floral attributes related to pollinator profiles. Statistical classification was then used to rank floral attributes according to their predictive value. Data sets excluding mixed pollination systems were used to test the predictive power of statistical classification to identify pollination models. Key Results Floral attributes are correlated with bird, fly and bee pollination. Using floral attributes identified as correlating with pollinator type, bird pollination is classified with 86 % accuracy, red flowers being the most important predictor. Fly and bee pollination are classified with 78 and 69 % accuracy, but have a lack of individually important floral predictors. Excluding mixed pollination systems improved the accuracy of the prediction of both bee and fly pollination systems. Conclusions Although most epacrids have generalized pollination systems, a correlation between bird pollination and red, long-tubed epacrids is found. Statistical classification highlights the relative importance of each floral attribute in relation to pollinator type and proves useful in classifying epacrids to bird, fly and bee pollination systems. PMID:23681546
Statistical genetics concepts and approaches in schizophrenia and related neuropsychiatric research.
Schork, Nicholas J; Greenwood, Tiffany A; Braff, David L
2007-01-01
Statistical genetics is a research field that focuses on mathematical models and statistical inference methodologies that relate genetic variations (ie, naturally occurring human DNA sequence variations or "polymorphisms") to particular traits or diseases (phenotypes) usually from data collected on large samples of families or individuals. The ultimate goal of such analysis is the identification of genes and genetic variations that influence disease susceptibility. Although of extreme interest and importance, the fact that many genes and environmental factors contribute to neuropsychiatric diseases of public health importance (eg, schizophrenia, bipolar disorder, and depression) complicates relevant studies and suggests that very sophisticated mathematical and statistical modeling may be required. In addition, large-scale contemporary human DNA sequencing and related projects, such as the Human Genome Project and the International HapMap Project, as well as the development of high-throughput DNA sequencing and genotyping technologies have provided statistical geneticists with a great deal of very relevant and appropriate information and resources. Unfortunately, the use of these resources and their interpretation are not straightforward when applied to complex, multifactorial diseases such as schizophrenia. In this brief and largely nonmathematical review of the field of statistical genetics, we describe many of the main concepts, definitions, and issues that motivate contemporary research. We also provide a discussion of the most pressing contemporary problems that demand further research if progress is to be made in the identification of genes and genetic variations that predispose to complex neuropsychiatric diseases.
Index of Economic Freedom: Unrealized Pedagogical Opportunities
ERIC Educational Resources Information Center
Maier, Mark; Miller, John A.
2017-01-01
Although the Index of Economic Freedom appears in many economic textbooks, their coverage of the index misses opportunities to teach statistical and policy-related concepts important for the principles course. The standard textbook presentation passes up an opportunity to examine the statistical issues of weighting in composite index numbers and…
Computer Assisted Problem Solving in an Introductory Statistics Course. Technical Report No. 56.
ERIC Educational Resources Information Center
Anderson, Thomas H.; And Others
The computer assisted problem solving system (CAPS) described in this booklet administered "homework" problem sets designed to develop students' computational, estimation, and procedural skills. These skills were related to important concepts in an introductory statistics course. CAPS generated unique data, judged student performance,…
NASA Technical Reports Server (NTRS)
Murphy, Kyle R.; Mann, Ian R.; Rae, I. Jonathan; Sibeck, David G.; Watt, Clare E. J.
2016-01-01
Wave-particle interactions play a crucial role in energetic particle dynamics in the Earths radiation belts. However, the relative importance of different wave modes in these dynamics is poorly understood. Typically, this is assessed during geomagnetic storms using statistically averaged empirical wave models as a function of geomagnetic activity in advanced radiation belt simulations. However, statistical averages poorly characterize extreme events such as geomagnetic storms in that storm-time ultralow frequency wave power is typically larger than that derived over a solar cycle and Kp is a poor proxy for storm-time wave power.
Bjerregaard, Peter; Becker, Ulrik
2013-01-01
Questionnaires are widely used to obtain information on health-related behaviour, and they are more often than not the only method that can be used to assess the distribution of behaviour in subgroups of the population. No validation studies of reported consumption of tobacco or alcohol have been published from circumpolar indigenous communities. The purpose of the study is to compare information on the consumption of tobacco and alcohol obtained from 3 population surveys in Greenland with import statistics. Estimates of consumption of cigarettes and alcohol using several different survey instruments in cross-sectional population studies from 1993-1994, 1999-2001 and 2005-2010 were compared with import statistics from the same years. For cigarettes, survey results accounted for virtually the total import. Alcohol consumption was significantly under-reported with reporting completeness ranging from 40% to 51% for different estimates of habitual weekly consumption in the 3 study periods. Including an estimate of binge drinking increased the estimated total consumption to 78% of the import. Compared with import statistics, questionnaire-based population surveys capture the consumption of cigarettes well in Greenland. Consumption of alcohol is under-reported, but asking about binge episodes in addition to the usual intake considerably increased the reported intake in this population and made it more in agreement with import statistics. It is unknown to what extent these findings at the population level can be inferred to population subgroups.
Johnson, Earl E; Mueller, H Gustav; Ricketts, Todd A
2009-01-01
To determine the amount of importance audiologists place on various items related to their selection of a preferred hearing aid brand manufacturer. Three hundred forty-three hearing aid-dispensing audiologists rated a total of 32 randomized items by survey methodology. Principle component analysis identified seven orthogonal statistical factors of importance. In rank order, these factors were Aptitude of the Brand, Image, Cost, Sales and Speed of Delivery, Exposure, Colleague Recommendations, and Contracts and Incentives. While it was hypothesized that differences among audiologists in the importance ratings of these factors would dictate their preference for a given brand, that was not our finding. Specifically, mean ratings for the six most important factors did not differ among audiologists preferring different brands. A statistically significant difference among audiologists preferring different brands was present, however, for one factor: Contracts and Incentives. Its assigned importance, though, was always lower than that for the other six factors. Although most audiologists have a preferred hearing aid brand, differences in the perceived importance of common factors attributed to brands do not largely determine preference for a particular brand.
NASA Astrophysics Data System (ADS)
Karl, Thomas R.; Wang, Wei-Chyung; Schlesinger, Michael E.; Knight, Richard W.; Portman, David
1990-10-01
Important surface observations such as the daily maximum and minimum temperature, daily precipitation, and cloud ceilings often have localized characteristics that are difficult to reproduce with the current resolution and the physical parameterizations in state-of-the-art General Circulation climate Models (GCMs). Many of the difficulties can be partially attributed to mismatches in scale, local topography. regional geography and boundary conditions between models and surface-based observations. Here, we present a method, called climatological projection by model statistics (CPMS), to relate GCM grid-point flee-atmosphere statistics, the predictors, to these important local surface observations. The method can be viewed as a generalization of the model output statistics (MOS) and perfect prog (PP) procedures used in numerical weather prediction (NWP) models. It consists of the application of three statistical methods: 1) principle component analysis (FICA), 2) canonical correlation, and 3) inflated regression analysis. The PCA reduces the redundancy of the predictors The canonical correlation is used to develop simultaneous relationships between linear combinations of the predictors, the canonical variables, and the surface-based observations. Finally, inflated regression is used to relate the important canonical variables to each of the surface-based observed variables.We demonstrate that even an early version of the Oregon State University two-level atmospheric GCM (with prescribed sea surface temperature) produces free-atmosphere statistics than can, when standardized using the model's internal means and variances (the MOS-like version of CPMS), closely approximate the observed local climate. When the model data are standardized by the observed free-atmosphere means and variances (the PP version of CPMS), however, the model does not reproduce the observed surface climate as well. Our results indicate that in the MOS-like version of CPMS the differences between the output of a ten-year GCM control run and the surface-based observations are often smaller than the differences between the observations of two ten-year periods. Such positive results suggest that GCMs may already contain important climatological information that can be used to infer the local climate.
Fine-Grained Sensitivity to Statistical Information in Adult Word Learning
ERIC Educational Resources Information Center
Vouloumanos, Athena
2008-01-01
A language learner trying to acquire a new word must often sift through many potential relations between particular words and their possible meanings. In principle, statistical information about the distribution of those mappings could serve as one important source of data, but little is known about whether learners can in fact track multiple…
Determinant Factors of Attitude towards Quantitative Subjects: Differences between Sexes
ERIC Educational Resources Information Center
Mondejar-Jimenez, Jose; Vargas-Vargas, Manuel
2010-01-01
Nowadays, almost all curricula in the social sciences contain at least one course in statistics, given the importance of this discipline as an analytical tool. This work identifies the latent factors relating to students' motivation and attitude towards statistics, tests their covariance structure for samples of both sexes, and identifies the…
Game Related Statistics Which Discriminate Between Winning and Losing Under-16 Male Basketball Games
Lorenzo, Alberto; Gómez, Miguel Ángel; Ortega, Enrique; Ibáñez, Sergio José; Sampaio, Jaime
2010-01-01
The aim of the present study was to identify the game-related statistics which discriminate between winning and losing teams in under-16 years old male basketball games. The sample gathered all 122 games in the 2004 and 2005 Under-16 European Championships. The game-related statistics analysed were the free-throws (both successful and unsuccessful), 2- and 3-points field-goals (both successful and unsuccessful) offensive and defensive rebounds, blocks, assists, fouls, turnovers and steals. The winning teams exhibited lower ball possessions per game and better offensive and defensive efficacy coefficients than the losing teams. Results from discriminant analysis were statistically significant and allowed to emphasize several structure coefficients (SC). In close games (final score differences below 9 points), the discriminant variables were the turnovers (SC = -0.47) and the assists (SC = 0.33). In balanced games (final score differences between 10 and 29 points), the variables that discriminated between the groups were the successful 2-point field-goals (SC = -0.34) and defensive rebounds (SC = -0. 36); and in unbalanced games (final score differences above 30 points) the variables that best discriminated both groups were the successful 2-point field-goals (SC = 0.37). These results allowed understanding that these players' specific characteristics result in a different game-related statistical profile and helped to point out the importance of the perceptive and decision making process in practice and in competition. Key points The players' game-related statistical profile varied according to game type, game outcome and in formative categories in basketball. The results of this work help to point out the different player's performance described in U-16 men's basketball teams compared with senior and professional men's basketball teams. The results obtained enhance the importance of the perceptive and decision making process in practice and in competition. PMID:24149794
Lisa A. Schulte; David J. Mladenoff; Erik V. Nordheim
2002-01-01
We developed a quantitative and replicable classification system to improve understanding of historical composition and structure within northern Wisconsin's forests. The classification system was based on statistical cluster analysis and two forest metrics, relative dominance (% basal area) and relative importance (mean of relative dominance and relative density...
2014-12-01
moving relative to the water in which they are immersed, reflecting the true school movement dynamics . There has also been work to implement this...Engineering Department Woods Hole Oceanographic Institution 98 Water Street, MS #11 Woods Hole, MA 02543 9. SPONSORING/MONITORING AGENCY NAME(S) AND...were measured with multi-beam sonars and quantified in terms of important aspects offish dynamics ; and predictions were made of echo statistics of a
A smoothed residual based goodness-of-fit statistic for nest-survival models
Rodney X. Sturdivant; Jay J. Rotella; Robin E. Russell
2008-01-01
Estimating nest success and identifying important factors related to nest-survival rates is an essential goal for many wildlife researchers interested in understanding avian population dynamics. Advances in statistical methods have led to a number of estimation methods and approaches to modeling this problem. Recently developed models allow researchers to include a...
ERIC Educational Resources Information Center
Sharma, Sashi
2007-01-01
Concerns about the importance of variation in statistics education and a lack of research in this topic led to a preliminary study which explored pre-service teachers' ideas in this area. The teachers completed a written questionnaire about variation in sampling and distribution contexts. Responses were categorised in relation to a framework that…
Statistical sensitivity analysis of a simple nuclear waste repository model
NASA Astrophysics Data System (ADS)
Ronen, Y.; Lucius, J. L.; Blow, E. M.
1980-06-01
A preliminary step in a comprehensive sensitivity analysis of the modeling of a nuclear waste repository. The purpose of the complete analysis is to determine which modeling parameters and physical data are most important in determining key design performance criteria and then to obtain the uncertainty in the design for safety considerations. The theory for a statistical screening design methodology is developed for later use in the overall program. The theory was applied to the test case of determining the relative importance of the sensitivity of near field temperature distribution in a single level salt repository to modeling parameters. The exact values of the sensitivities to these physical and modeling parameters were then obtained using direct methods of recalculation. The sensitivity coefficients found to be important for the sample problem were thermal loading, distance between the spent fuel canisters and their radius. Other important parameters were those related to salt properties at a point of interest in the repository.
Fixations on objects in natural scenes: dissociating importance from salience
't Hart, Bernard M.; Schmidt, Hannah C. E. F.; Roth, Christine; Einhäuser, Wolfgang
2013-01-01
The relation of selective attention to understanding of natural scenes has been subject to intense behavioral research and computational modeling, and gaze is often used as a proxy for such attention. The probability of an image region to be fixated typically correlates with its contrast. However, this relation does not imply a causal role of contrast. Rather, contrast may relate to an object's “importance” for a scene, which in turn drives attention. Here we operationalize importance by the probability that an observer names the object as characteristic for a scene. We modify luminance contrast of either a frequently named (“common”/“important”) or a rarely named (“rare”/“unimportant”) object, track the observers' eye movements during scene viewing and ask them to provide keywords describing the scene immediately after. When no object is modified relative to the background, important objects draw more fixations than unimportant ones. Increases of contrast make an object more likely to be fixated, irrespective of whether it was important for the original scene, while decreases in contrast have little effect on fixations. Any contrast modification makes originally unimportant objects more important for the scene. Finally, important objects are fixated more centrally than unimportant objects, irrespective of contrast. Our data suggest a dissociation between object importance (relevance for the scene) and salience (relevance for attention). If an object obeys natural scene statistics, important objects are also salient. However, when natural scene statistics are violated, importance and salience are differentially affected. Object salience is modulated by the expectation about object properties (e.g., formed by context or gist), and importance by the violation of such expectations. In addition, the dependence of fixated locations within an object on the object's importance suggests an analogy to the effects of word frequency on landing positions in reading. PMID:23882251
Extending Working Life: Which Competencies are Crucial in Near-Retirement Age?
Wiktorowicz, Justyna
2018-01-01
Nowadays, one of the most important economic and social phenomena is population ageing. Due to the low activity rate of older people, one of the most important challenges is to take various actions involving active ageing, which is supposed to extending working life, and along with it-improve the competencies of older people. The aim of this paper is to evaluate the relevance of different competencies for extending working life, with limiting the analysis for Poland. The paper also assesses the competencies of mature Polish people (aged 50+, but still in working age). In the statistical analysis, I used logistic regression, as well as descriptive statistics and appropriate statistical tests. The results show that among the actions aimed at extending working life, the most important are those related to lifelong learning, targeted at improving the competencies of the older generation. The competencies (both soft and hard) of people aged 50+ are more important than their formal education.
Respectful and Responsive Pedagogies for Mathematics and Statistics
ERIC Educational Resources Information Center
Averill, Robin; Clarke, Megan
2013-01-01
Teacher respect, important within culturally responsive practice, has seldom been explored in relation to mathematical pedagogy. Our study involving interviews, surveys, and lesson videos with Year 12 and 13 New Zealand mathematics students and teachers indicated specific pedagogical behaviours are important for demonstrating respect for students…
Factors Influencing Young People's Conceptions of Environment.
ERIC Educational Resources Information Center
Loughland, Tony; Reid, Anna; Walker, Kim; Petocz, Peter
2003-01-01
Explains the importance of environmental education in schools for achieving environmental protection and improvement. Statistically examines factors that incline students to a 'relation' rather than an 'object' conception of the environment. Concludes that development of the former would seem to be an important aim of environmental education and…
Statistics Anxiety Update: Refining the Construct and Recommendations for a New Research Agenda.
Chew, Peter K H; Dillon, Denise B
2014-03-01
Appreciation of the importance of statistics literacy for citizens of a democracy has resulted in an increasing number of degree programs making statistics courses mandatory for university students. Unfortunately, empirical evidence suggests that students in nonmathematical disciplines (e.g., social sciences) regard statistics courses as the most anxiety-inducing course in their degree programs. Although a literature review exists for statistics anxiety, it was done more than a decade ago, and newer studies have since added findings for consideration. In this article, we provide a current review of the statistics anxiety literature. Specifically, related variables, definitions, and measures of statistics anxiety are reviewed with the goal of refining the statistics anxiety construct. Antecedents, effects, and interventions of statistics anxiety are also reviewed to provide recommendations for statistics instructors and for a new research agenda. © The Author(s) 2014.
[The informational support of statistical observation related to children disability].
Son, I M; Polikarpov, A V; Ogrizko, E V; Golubeva, T Yu
2016-01-01
Within the framework of the Convention on rights of the disabled the revision is specified concerning criteria of identification of disability of children and reformation of system of medical social expertise according international standards of indices of health and indices related to health. In connection with it, it is important to consider the relationship between alterations in forms of the Federal statistical monitoring in the part of registration of disabled children in the Russian Federation and classification of health indices and indices related to health applied at identification of disability. The article presents analysis of relationship between alterations in forms of the Federal statistical monitoring in the part of registration of disabled children in the Russian Federation and applied classifications used at identification of disability (International classification of impairments, disabilities and handicap (ICDH), international classification of functioning, disability and health (ICF), international classification of functioning, disability and health, version for children and youth (ICF-CY). The intersectorial interaction is considered within the framework of statistics of children disability.
NASA Astrophysics Data System (ADS)
Eliazar, Iddo
2017-05-01
The exponential, the normal, and the Poisson statistical laws are of major importance due to their universality. Harmonic statistics are as universal as the three aforementioned laws, but yet they fall short in their 'public relations' for the following reason: the full scope of harmonic statistics cannot be described in terms of a statistical law. In this paper we describe harmonic statistics, in their full scope, via an object termed harmonic Poisson process: a Poisson process, over the positive half-line, with a harmonic intensity. The paper reviews the harmonic Poisson process, investigates its properties, and presents the connections of this object to an assortment of topics: uniform statistics, scale invariance, random multiplicative perturbations, Pareto and inverse-Pareto statistics, exponential growth and exponential decay, power-law renormalization, convergence and domains of attraction, the Langevin equation, diffusions, Benford's law, and 1/f noise.
Petersson, K M; Nichols, T E; Poline, J B; Holmes, A P
1999-01-01
Functional neuroimaging (FNI) provides experimental access to the intact living brain making it possible to study higher cognitive functions in humans. In this review and in a companion paper in this issue, we discuss some common methods used to analyse FNI data. The emphasis in both papers is on assumptions and limitations of the methods reviewed. There are several methods available to analyse FNI data indicating that none is optimal for all purposes. In order to make optimal use of the methods available it is important to know the limits of applicability. For the interpretation of FNI results it is also important to take into account the assumptions, approximations and inherent limitations of the methods used. This paper gives a brief overview over some non-inferential descriptive methods and common statistical models used in FNI. Issues relating to the complex problem of model selection are discussed. In general, proper model selection is a necessary prerequisite for the validity of the subsequent statistical inference. The non-inferential section describes methods that, combined with inspection of parameter estimates and other simple measures, can aid in the process of model selection and verification of assumptions. The section on statistical models covers approaches to global normalization and some aspects of univariate, multivariate, and Bayesian models. Finally, approaches to functional connectivity and effective connectivity are discussed. In the companion paper we review issues related to signal detection and statistical inference. PMID:10466149
Some challenges with statistical inference in adaptive designs.
Hung, H M James; Wang, Sue-Jane; Yang, Peiling
2014-01-01
Adaptive designs have generated a great deal of attention to clinical trial communities. The literature contains many statistical methods to deal with added statistical uncertainties concerning the adaptations. Increasingly encountered in regulatory applications are adaptive statistical information designs that allow modification of sample size or related statistical information and adaptive selection designs that allow selection of doses or patient populations during the course of a clinical trial. For adaptive statistical information designs, a few statistical testing methods are mathematically equivalent, as a number of articles have stipulated, but arguably there are large differences in their practical ramifications. We pinpoint some undesirable features of these methods in this work. For adaptive selection designs, the selection based on biomarker data for testing the correlated clinical endpoints may increase statistical uncertainty in terms of type I error probability, and most importantly the increased statistical uncertainty may be impossible to assess.
Background Knowledge in Learning-Based Relation Extraction
ERIC Educational Resources Information Center
Do, Quang Xuan
2012-01-01
In this thesis, we study the importance of background knowledge in relation extraction systems. We not only demonstrate the benefits of leveraging background knowledge to improve the systems' performance but also propose a principled framework that allows one to effectively incorporate knowledge into statistical machine learning models for…
Hazardous substances, the environment and public health: a statistical overview.
Hunter, W G; Crowley, J J
1979-01-01
The purpose of this paper is to provide an overview of the statistical problems that exist and procedures that are available when attempts are made to assess the possible harm which has been or might be caused by substances in the environment. These issues bear directly on important decisions of public policy such as those related to the establishment and enforcement of regulations. PMID:540596
Stefanov, Valeri T
2002-01-01
Background Pairs of related individuals are widely used in linkage analysis. Most of the tests for linkage analysis are based on statistics associated with identity by descent (IBD) data. The current biotechnology provides data on very densely packed loci, and therefore, it may provide almost continuous IBD data for pairs of closely related individuals. Therefore, the distribution theory for statistics on continuous IBD data is of interest. In particular, distributional results which allow the evaluation of p-values for relevant tests are of importance. Results A technology is provided for numerical evaluation, with any given accuracy, of the cumulative probabilities of some statistics on continuous genome data for pairs of closely related individuals. In the case of a pair of full-sibs, the following statistics are considered: (i) the proportion of genome with 2 (at least 1) haplotypes shared identical-by-descent (IBD) on a chromosomal segment, (ii) the number of distinct pieces (subsegments) of a chromosomal segment, on each of which exactly 2 (at least 1) haplotypes are shared IBD. The natural counterparts of these statistics for the other relationships are also considered. Relevant Maple codes are provided for a rapid evaluation of the cumulative probabilities of such statistics. The genomic continuum model, with Haldane's model for the crossover process, is assumed. Conclusions A technology, together with relevant software codes for its automated implementation, are provided for exact evaluation of the distributions of relevant statistics associated with continuous genome data on closely related individuals. PMID:11996673
Ng'andu, N H
1997-03-30
In the analysis of survival data using the Cox proportional hazard (PH) model, it is important to verify that the explanatory variables analysed satisfy the proportional hazard assumption of the model. This paper presents results of a simulation study that compares five test statistics to check the proportional hazard assumption of Cox's model. The test statistics were evaluated under proportional hazards and the following types of departures from the proportional hazard assumption: increasing relative hazards; decreasing relative hazards; crossing hazards; diverging hazards, and non-monotonic hazards. The test statistics compared include those based on partitioning of failure time and those that do not require partitioning of failure time. The simulation results demonstrate that the time-dependent covariate test, the weighted residuals score test and the linear correlation test have equally good power for detection of non-proportionality in the varieties of non-proportional hazards studied. Using illustrative data from the literature, these test statistics performed similarly.
Wicherts, Jelte M.; Bakker, Marjan; Molenaar, Dylan
2011-01-01
Background The widespread reluctance to share published research data is often hypothesized to be due to the authors' fear that reanalysis may expose errors in their work or may produce conclusions that contradict their own. However, these hypotheses have not previously been studied systematically. Methods and Findings We related the reluctance to share research data for reanalysis to 1148 statistically significant results reported in 49 papers published in two major psychology journals. We found the reluctance to share data to be associated with weaker evidence (against the null hypothesis of no effect) and a higher prevalence of apparent errors in the reporting of statistical results. The unwillingness to share data was particularly clear when reporting errors had a bearing on statistical significance. Conclusions Our findings on the basis of psychological papers suggest that statistical results are particularly hard to verify when reanalysis is more likely to lead to contrasting conclusions. This highlights the importance of establishing mandatory data archiving policies. PMID:22073203
Rasch fit statistics and sample size considerations for polytomous data.
Smith, Adam B; Rush, Robert; Fallowfield, Lesley J; Velikova, Galina; Sharpe, Michael
2008-05-29
Previous research on educational data has demonstrated that Rasch fit statistics (mean squares and t-statistics) are highly susceptible to sample size variation for dichotomously scored rating data, although little is known about this relationship for polytomous data. These statistics help inform researchers about how well items fit to a unidimensional latent trait, and are an important adjunct to modern psychometrics. Given the increasing use of Rasch models in health research the purpose of this study was therefore to explore the relationship between fit statistics and sample size for polytomous data. Data were collated from a heterogeneous sample of cancer patients (n = 4072) who had completed both the Patient Health Questionnaire - 9 and the Hospital Anxiety and Depression Scale. Ten samples were drawn with replacement for each of eight sample sizes (n = 25 to n = 3200). The Rating and Partial Credit Models were applied and the mean square and t-fit statistics (infit/outfit) derived for each model. The results demonstrated that t-statistics were highly sensitive to sample size, whereas mean square statistics remained relatively stable for polytomous data. It was concluded that mean square statistics were relatively independent of sample size for polytomous data and that misfit to the model could be identified using published recommended ranges.
Rasch fit statistics and sample size considerations for polytomous data
Smith, Adam B; Rush, Robert; Fallowfield, Lesley J; Velikova, Galina; Sharpe, Michael
2008-01-01
Background Previous research on educational data has demonstrated that Rasch fit statistics (mean squares and t-statistics) are highly susceptible to sample size variation for dichotomously scored rating data, although little is known about this relationship for polytomous data. These statistics help inform researchers about how well items fit to a unidimensional latent trait, and are an important adjunct to modern psychometrics. Given the increasing use of Rasch models in health research the purpose of this study was therefore to explore the relationship between fit statistics and sample size for polytomous data. Methods Data were collated from a heterogeneous sample of cancer patients (n = 4072) who had completed both the Patient Health Questionnaire – 9 and the Hospital Anxiety and Depression Scale. Ten samples were drawn with replacement for each of eight sample sizes (n = 25 to n = 3200). The Rating and Partial Credit Models were applied and the mean square and t-fit statistics (infit/outfit) derived for each model. Results The results demonstrated that t-statistics were highly sensitive to sample size, whereas mean square statistics remained relatively stable for polytomous data. Conclusion It was concluded that mean square statistics were relatively independent of sample size for polytomous data and that misfit to the model could be identified using published recommended ranges. PMID:18510722
Quantification of historical livestock importation into New Zealand 1860-1979.
Binney, B M; Biggs, P J; Carter, P E; Holland, B R; Holland, B M; French, N P
2014-11-01
To quantify the numbers of live cattle, sheep and poultry imported into New Zealand and, where possible, their country of origin from 1860 to 1979. Information on the origin and number of live animal importations into New Zealand was collected for cattle, sheep and poultry for the period 1868-1979 from the annual reports compiled by the New Zealand Registrar General's Office, Government Statistician's Office, Census and Statistics Office, Census and Statistics Department, Customs Department and Department of Statistics. Census data from 1851 to 1871 were also used to estimate the livestock population during this period. The number of animals imported and the mean population for each species in a decade were determined, and the major countries of origin were identified. A large number of cattle (53,384) and sheep (604,525) were imported in the 1860s, and then there was a marked reduction in importations. Live poultry were imported in relatively small numbers (20,701) from 1880 to 1939, then 1,564,330 live poultry were imported between 1960 and 1979. Australia was the predominant country of origin for sheep between 1868 and 1959 (51,347/60,918; 84.3%) and of cattle between 1868 and 1979 (10,080/15,157; 66.5%). Only 6,712 (11.0%) sheep and 3,909 (25.8%) cattle were imported from the United Kingdom over the same periods, and even fewer from other countries. The collated data and historical reports show that from 1860 to 1979 Australia has been the main source of livestock introduced into New Zealand. The pattern of importation showed that large numbers of cattle and sheep were initially imported in the 1860s, probably in response to rapid agricultural expansion. Thereafter importations continued at much reduced numbers. In contrast, relatively small numbers of poultry were introduced until the 1960s when large numbers were imported as part of the development of a modern high-production industry. The overall pattern for both cattle and sheep was of a bottleneck event, as initially a relatively limited number of animals arrived from outside populations, followed by population expansion with ongoing but limited immigration (admixture). Investigation into the genetic population structure of New Zealand's cattle and sheep, as well as their host-associated microorganisms, could reflect the impact of these early historical events.
Minimum Uncertainty Coherent States Attached to Nondegenerate Parametric Amplifiers
NASA Astrophysics Data System (ADS)
Dehghani, A.; Mojaveri, B.
2015-06-01
Exact analytical solutions for the two-mode nondegenerate parametric amplifier have been obtained by using the transformation from the two-dimensional harmonic oscillator Hamiltonian. Some important physical properties such as quantum statistics and quadrature squeezing of the corresponding states are investigated. In addition, these states carry classical features such as Poissonian statistics and minimize the Heisenberg uncertainty relation of a pair of the coordinate and the momentum operators.
Experience and Sentence Processing: Statistical Learning and Relative Clause Comprehension
Wells, Justine B.; Christiansen, Morten H.; Race, David S.; Acheson, Daniel J.; MacDonald, Maryellen C.
2009-01-01
Many explanations of the difficulties associated with interpreting object relative clauses appeal to the demands that object relatives make on working memory. MacDonald and Christiansen (2002) pointed to variations in reading experience as a source of differences, arguing that the unique word order of object relatives makes their processing more difficult and more sensitive to the effects of previous experience than the processing of subject relatives. This hypothesis was tested in a large-scale study manipulating reading experiences of adults over several weeks. The group receiving relative clause experience increased reading speeds for object relatives more than for subject relatives, whereas a control experience group did not. The reading time data were compared to performance of a computational model given different amounts of experience. The results support claims for experience-based individual differences and an important role for statistical learning in sentence comprehension processes. PMID:18922516
Novick, Steven; Shen, Yan; Yang, Harry; Peterson, John; LeBlond, Dave; Altan, Stan
2015-01-01
Dissolution (or in vitro release) studies constitute an important aspect of pharmaceutical drug development. One important use of such studies is for justifying a biowaiver for post-approval changes which requires establishing equivalence between the new and old product. We propose a statistically rigorous modeling approach for this purpose based on the estimation of what we refer to as the F2 parameter, an extension of the commonly used f2 statistic. A Bayesian test procedure is proposed in relation to a set of composite hypotheses that capture the similarity requirement on the absolute mean differences between test and reference dissolution profiles. Several examples are provided to illustrate the application. Results of our simulation study comparing the performance of f2 and the proposed method show that our Bayesian approach is comparable to or in many cases superior to the f2 statistic as a decision rule. Further useful extensions of the method, such as the use of continuous-time dissolution modeling, are considered.
Inverse statistical physics of protein sequences: a key issues review.
Cocco, Simona; Feinauer, Christoph; Figliuzzi, Matteo; Monasson, Rémi; Weigt, Martin
2018-03-01
In the course of evolution, proteins undergo important changes in their amino acid sequences, while their three-dimensional folded structure and their biological function remain remarkably conserved. Thanks to modern sequencing techniques, sequence data accumulate at unprecedented pace. This provides large sets of so-called homologous, i.e. evolutionarily related protein sequences, to which methods of inverse statistical physics can be applied. Using sequence data as the basis for the inference of Boltzmann distributions from samples of microscopic configurations or observables, it is possible to extract information about evolutionary constraints and thus protein function and structure. Here we give an overview over some biologically important questions, and how statistical-mechanics inspired modeling approaches can help to answer them. Finally, we discuss some open questions, which we expect to be addressed over the next years.
Inverse statistical physics of protein sequences: a key issues review
NASA Astrophysics Data System (ADS)
Cocco, Simona; Feinauer, Christoph; Figliuzzi, Matteo; Monasson, Rémi; Weigt, Martin
2018-03-01
In the course of evolution, proteins undergo important changes in their amino acid sequences, while their three-dimensional folded structure and their biological function remain remarkably conserved. Thanks to modern sequencing techniques, sequence data accumulate at unprecedented pace. This provides large sets of so-called homologous, i.e. evolutionarily related protein sequences, to which methods of inverse statistical physics can be applied. Using sequence data as the basis for the inference of Boltzmann distributions from samples of microscopic configurations or observables, it is possible to extract information about evolutionary constraints and thus protein function and structure. Here we give an overview over some biologically important questions, and how statistical-mechanics inspired modeling approaches can help to answer them. Finally, we discuss some open questions, which we expect to be addressed over the next years.
Statistical Analyses of Scatterplots to Identify Important Factors in Large-Scale Simulations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kleijnen, J.P.C.; Helton, J.C.
1999-04-01
The robustness of procedures for identifying patterns in scatterplots generated in Monte Carlo sensitivity analyses is investigated. These procedures are based on attempts to detect increasingly complex patterns in the scatterplots under consideration and involve the identification of (1) linear relationships with correlation coefficients, (2) monotonic relationships with rank correlation coefficients, (3) trends in central tendency as defined by means, medians and the Kruskal-Wallis statistic, (4) trends in variability as defined by variances and interquartile ranges, and (5) deviations from randomness as defined by the chi-square statistic. The following two topics related to the robustness of these procedures are consideredmore » for a sequence of example analyses with a large model for two-phase fluid flow: the presence of Type I and Type II errors, and the stability of results obtained with independent Latin hypercube samples. Observations from analysis include: (1) Type I errors are unavoidable, (2) Type II errors can occur when inappropriate analysis procedures are used, (3) physical explanations should always be sought for why statistical procedures identify variables as being important, and (4) the identification of important variables tends to be stable for independent Latin hypercube samples.« less
Perturbation Selection and Local Influence Analysis for Nonlinear Structural Equation Model
ERIC Educational Resources Information Center
Chen, Fei; Zhu, Hong-Tu; Lee, Sik-Yum
2009-01-01
Local influence analysis is an important statistical method for studying the sensitivity of a proposed model to model inputs. One of its important issues is related to the appropriate choice of a perturbation vector. In this paper, we develop a general method to select an appropriate perturbation vector and a second-order local influence measure…
2013-06-01
or indicators are used as long range memory measurements. Hurst and Holder exponents are the most important and popular parameters. Traditionally...the relation between two important parameters, the Hurst exponent (measurement of global long range memory) and the Entropy (measurement of...empirical results and future study. II. BACKGROUND We recall briey the mathematical and statistical definitions and properties of the Hurst exponents
Non-classical State via Superposition of Two Opposite Coherent States
NASA Astrophysics Data System (ADS)
Ren, Gang; Du, Jian-ming; Yu, Hai-jun
2018-04-01
We study the non-classical properties of the states generated by superpositions of two opposite coherent states with the arbitrary relative phase factors. We show that the relative phase factors plays an important role in these superpositions. We demonstrate this result by discussing their squeezing properties, quantum statistical properties and fidelity in principle.
Age and experience shape developmental changes in the neural basis of language-related learning.
McNealy, Kristin; Mazziotta, John C; Dapretto, Mirella
2011-11-01
Very little is known about the neural underpinnings of language learning across the lifespan and how these might be modified by maturational and experiential factors. Building on behavioral research highlighting the importance of early word segmentation (i.e. the detection of word boundaries in continuous speech) for subsequent language learning, here we characterize developmental changes in brain activity as this process occurs online, using data collected in a mixed cross-sectional and longitudinal design. One hundred and fifty-six participants, ranging from age 5 to adulthood, underwent functional magnetic resonance imaging (fMRI) while listening to three novel streams of continuous speech, which contained either strong statistical regularities, strong statistical regularities and speech cues, or weak statistical regularities providing minimal cues to word boundaries. All age groups displayed significant signal increases over time in temporal cortices for the streams with high statistical regularities; however, we observed a significant right-to-left shift in the laterality of these learning-related increases with age. Interestingly, only the 5- to 10-year-old children displayed significant signal increases for the stream with low statistical regularities, suggesting an age-related decrease in sensitivity to more subtle statistical cues. Further, in a sample of 78 10-year-olds, we examined the impact of proficiency in a second language and level of pubertal development on learning-related signal increases, showing that the brain regions involved in language learning are influenced by both experiential and maturational factors. 2011 Blackwell Publishing Ltd.
Nutritional status and CD4 cell counts in patients with HIV/AIDS receiving antiretroviral therapy.
Santos, Ana Célia Oliveira dos; Almeida, Ana Maria Rampeloti
2013-01-01
Even with current highly active antiretroviral therapy, individuals with AIDS continue to exhibit important nutritional deficits and reduced levels of albumin and hemoglobin, which may be directly related to their cluster of differentiation 4 (CD4) cell counts. The aim of this study was to characterize the nutritional status of individuals with human immunodeficiency virus/acquired immunodeficiency syndrome (HIV/AIDS) and relate the findings to the albumin level, hemoglobin level and CD4 cell count. Patients over 20 years of age with AIDS who were hospitalized in a university hospital and were receiving antiretroviral therapy were studied with regard to clinical, anthropometric, biochemical and sociodemographic characteristics. Body mass index, percentage of weight loss, arm circumference, triceps skinfold and arm muscle circumference were analyzed. Data on albumin, hemoglobin, hematocrit and CD4 cell count were obtained from patient charts. Statistical analysis was performed using Fisher's exact test, Student's t-test for independent variables and the Mann-Whitney U-test. The level of significance was set to 0.05 (α = 5%). Statistical analysis was performed using Statistical Package for the Social Sciences (SPSS) 17.0 software for Windows. Of the 50 patients evaluated, 70% were male. The prevalence of malnutrition was higher when the definition was based on arm circumference and triceps skinfold measurement. The concentrations of all biochemical variables were significantly lower among patients with a body mass index of less than 18.5kg/m2. The CD4 cell count, albumin, hemoglobin and hematocrit anthropometric measures were directly related to each other. These findings underscore the importance of nutritional follow-up for underweight patients with AIDS, as nutritional status proved to be related to important biochemical alterations.
Statistical Modeling of Natural Backgrounds in Hyperspectral LWIR Data
2016-09-06
extremely important for studying performance trades. First, we study the validity of this model using real hyperspectral data, and compare the relative...difficult to validate any statistical model created for a target of interest. However, since background measurements are plentiful, it is reasonable to...Golden, S., Less, D., Jin, X., and Rynes, P., “ Modeling and analysis of LWIR signature variability associated with 3d and BRDF effects,” 98400P (May 2016
A new universality class in corpus of texts; A statistical physics study
NASA Astrophysics Data System (ADS)
Najafi, Elham; Darooneh, Amir H.
2018-05-01
Text can be regarded as a complex system. There are some methods in statistical physics which can be used to study this system. In this work, by means of statistical physics methods, we reveal new universal behaviors of texts associating with the fractality values of words in a text. The fractality measure indicates the importance of words in a text by considering distribution pattern of words throughout the text. We observed a power law relation between fractality of text and vocabulary size for texts and corpora. We also observed this behavior in studying biological data.
Changing viewer perspectives reveals constraints to implicit visual statistical learning.
Jiang, Yuhong V; Swallow, Khena M
2014-10-07
Statistical learning-learning environmental regularities to guide behavior-likely plays an important role in natural human behavior. One potential use is in search for valuable items. Because visual statistical learning can be acquired quickly and without intention or awareness, it could optimize search and thereby conserve energy. For this to be true, however, visual statistical learning needs to be viewpoint invariant, facilitating search even when people walk around. To test whether implicit visual statistical learning of spatial information is viewpoint independent, we asked participants to perform a visual search task from variable locations around a monitor placed flat on a stand. Unbeknownst to participants, the target was more often in some locations than others. In contrast to previous research on stationary observers, visual statistical learning failed to produce a search advantage for targets in high-probable regions that were stable within the environment but variable relative to the viewer. This failure was observed even when conditions for spatial updating were optimized. However, learning was successful when the rich locations were referenced relative to the viewer. We conclude that changing viewer perspective disrupts implicit learning of the target's location probability. This form of learning shows limited integration with spatial updating or spatiotopic representations. © 2014 ARVO.
Hill, Mary C.
2010-01-01
Doherty and Hunt (2009) present important ideas for first-order-second moment sensitivity analysis, but five issues are discussed in this comment. First, considering the composite-scaled sensitivity (CSS) jointly with parameter correlation coefficients (PCC) in a CSS/PCC analysis addresses the difficulties with CSS mentioned in the introduction. Second, their new parameter identifiability statistic actually is likely to do a poor job of parameter identifiability in common situations. The statistic instead performs the very useful role of showing how model parameters are included in the estimated singular value decomposition (SVD) parameters. Its close relation to CSS is shown. Third, the idea from p. 125 that a suitable truncation point for SVD parameters can be identified using the prediction variance is challenged using results from Moore and Doherty (2005). Fourth, the relative error reduction statistic of Doherty and Hunt is shown to belong to an emerging set of statistics here named perturbed calculated variance statistics. Finally, the perturbed calculated variance statistics OPR and PPR mentioned on p. 121 are shown to explicitly include the parameter null-space component of uncertainty. Indeed, OPR and PPR results that account for null-space uncertainty have appeared in the literature since 2000.
Helping Students Assess the Relative Importance of Different Intermolecular Interactions
ERIC Educational Resources Information Center
Jasien, Paul G.
2008-01-01
A semi-quantitative model has been developed to estimate the relative effects of dispersion, dipole-dipole interactions, and H-bonding on the normal boiling points ("T[subscript b]") for a subset of simple organic systems. The model is based upon a statistical analysis using multiple linear regression on a series of straight-chain organic…
Seven ways to increase power without increasing N.
Hansen, W B; Collins, L M
1994-01-01
Many readers of this monograph may wonder why a chapter on statistical power was included. After all, by now the issue of statistical power is in many respects mundane. Everyone knows that statistical power is a central research consideration, and certainly most National Institute on Drug Abuse grantees or prospective grantees understand the importance of including a power analysis in research proposals. However, there is ample evidence that, in practice, prevention researchers are not paying sufficient attention to statistical power. If they were, the findings observed by Hansen (1992) in a recent review of the prevention literature would not have emerged. Hansen (1992) examined statistical power based on 46 cohorts followed longitudinally, using nonparametric assumptions given the subjects' age at posttest and the numbers of subjects. Results of this analysis indicated that, in order for a study to attain 80-percent power for detecting differences between treatment and control groups, the difference between groups at posttest would need to be at least 8 percent (in the best studies) and as much as 16 percent (in the weakest studies). In order for a study to attain 80-percent power for detecting group differences in pre-post change, 22 of the 46 cohorts would have needed relative pre-post reductions of greater than 100 percent. Thirty-three of the 46 cohorts had less than 50-percent power to detect a 50-percent relative reduction in substance use. These results are consistent with other review findings (e.g., Lipsey 1990) that have shown a similar lack of power in a broad range of research topics. Thus, it seems that, although researchers are aware of the importance of statistical power (particularly of the necessity for calculating it when proposing research), they somehow are failing to end up with adequate power in their completed studies. This chapter argues that the failure of many prevention studies to maintain adequate statistical power is due to an overemphasis on sample size (N) as the only, or even the best, way to increase statistical power. It is easy to see how this overemphasis has come about. Sample size is easy to manipulate, has the advantage of being related to power in a straight-forward way, and usually is under the direct control of the researcher, except for limitations imposed by finances or subject availability. Another option for increasing power is to increase the alpha used for hypothesis-testing but, as very few researchers seriously consider significance levels much larger than the traditional .05, this strategy seldom is used. Of course, sample size is important, and the authors of this chapter are not recommending that researchers cease choosing sample sizes carefully. Rather, they argue that researchers should not confine themselves to increasing N to enhance power. It is important to take additional measures to maintain and improve power over and above making sure the initial sample size is sufficient. The authors recommend two general strategies. One strategy involves attempting to maintain the effective initial sample size so that power is not lost needlessly. The other strategy is to take measures to maximize the third factor that determines statistical power: effect size.
Using ontology network structure in text mining.
Berndt, Donald J; McCart, James A; Luther, Stephen L
2010-11-13
Statistical text mining treats documents as bags of words, with a focus on term frequencies within documents and across document collections. Unlike natural language processing (NLP) techniques that rely on an engineered vocabulary or a full-featured ontology, statistical approaches do not make use of domain-specific knowledge. The freedom from biases can be an advantage, but at the cost of ignoring potentially valuable knowledge. The approach proposed here investigates a hybrid strategy based on computing graph measures of term importance over an entire ontology and injecting the measures into the statistical text mining process. As a starting point, we adapt existing search engine algorithms such as PageRank and HITS to determine term importance within an ontology graph. The graph-theoretic approach is evaluated using a smoking data set from the i2b2 National Center for Biomedical Computing, cast as a simple binary classification task for categorizing smoking-related documents, demonstrating consistent improvements in accuracy.
Mazor, Kathleen M; Baril, Joann; Dugan, Elizabeth; Spencer, Frederick; Burgwinkle, Pamela; Gurwitz, Jerry H
2007-12-01
To determine the relative impact of incorporating narrative evidence, statistical evidence or both into patient education about warfarin, a widely used oral anticoagulant medication. 600 patients receiving anticoagulant therapy were randomly assigned to view one of three versions of a video depicting a physician-patient encounter where anticoagulation treatment was discussed, or usual care (no video). The videos differed in whether the physician used narrative evidence (patient anecdotes), statistical evidence, or both to highlight key information. 317 patients completed both the baseline and post-test questionnaires. Questions assessed knowledge, beliefs and adherence to medication and laboratory monitoring regimens. All three approaches positively effected patients' warfarin-related knowledge, and beliefs in the importance of lab testing; there was also some indication that viewing a video strengthened belief in the benefits of warfarin. There was some indication that narrative evidence had a greater impact than statistical evidence on beliefs about the importance of lab testing and on knowledge. No other evidence of the differential effectiveness of either approach was found. No statistically significant effect was found on intent to adhere, or documented adherence to lab monitoring. Videos depicting a physician-patient dialogue about warfarin were effective in educating patients about anticoagulant medication, and had a positive impact on their beliefs. The use of narrative evidence in the form of patient anecdotes may be more effective than statistical evidence for some patient outcomes. Patients on oral anticoagulant therapy may benefit from periodic educational efforts reinforcing key medication safety information, even after initial education and ongoing monitoring. Incorporating patient anecdotes into physician-patient dialogues or educational materials may increase the effectiveness of the message.
A summary of selected nationwide school bus crash statistics in 1989
DOT National Transportation Integrated Search
1991-05-01
Author's abstract: Crashes involving vehicles used in the noncommercial transport of persons for either school related or private group transportation present a small, but important, portion of the nation's motor vehicle crash experience each year. H...
Comparing Visual and Statistical Analysis of Multiple Baseline Design Graphs.
Wolfe, Katie; Dickenson, Tammiee S; Miller, Bridget; McGrath, Kathleen V
2018-04-01
A growing number of statistical analyses are being developed for single-case research. One important factor in evaluating these methods is the extent to which each corresponds to visual analysis. Few studies have compared statistical and visual analysis, and information about more recently developed statistics is scarce. Therefore, our purpose was to evaluate the agreement between visual analysis and four statistical analyses: improvement rate difference (IRD); Tau-U; Hedges, Pustejovsky, Shadish (HPS) effect size; and between-case standardized mean difference (BC-SMD). Results indicate that IRD and BC-SMD had the strongest overall agreement with visual analysis. Although Tau-U had strong agreement with visual analysis on raw values, it had poorer agreement when those values were dichotomized to represent the presence or absence of a functional relation. Overall, visual analysis appeared to be more conservative than statistical analysis, but further research is needed to evaluate the nature of these disagreements.
Technical Note: The Initial Stages of Statistical Data Analysis
Tandy, Richard D.
1998-01-01
Objective: To provide an overview of several important data-related considerations in the design stage of a research project and to review the levels of measurement and their relationship to the statistical technique chosen for the data analysis. Background: When planning a study, the researcher must clearly define the research problem and narrow it down to specific, testable questions. The next steps are to identify the variables in the study, decide how to group and treat subjects, and determine how to measure, and the underlying level of measurement of, the dependent variables. Then the appropriate statistical technique can be selected for data analysis. Description: The four levels of measurement in increasing complexity are nominal, ordinal, interval, and ratio. Nominal data are categorical or “count” data, and the numbers are treated as labels. Ordinal data can be ranked in a meaningful order by magnitude. Interval data possess the characteristics of ordinal data and also have equal distances between levels. Ratio data have a natural zero point. Nominal and ordinal data are analyzed with nonparametric statistical techniques and interval and ratio data with parametric statistical techniques. Advantages: Understanding the four levels of measurement and when it is appropriate to use each is important in determining which statistical technique to use when analyzing data. PMID:16558489
Comparing geological and statistical approaches for element selection in sediment tracing research
NASA Astrophysics Data System (ADS)
Laceby, J. Patrick; McMahon, Joe; Evrard, Olivier; Olley, Jon
2015-04-01
Elevated suspended sediment loads reduce reservoir capacity and significantly increase the cost of operating water treatment infrastructure, making the management of sediment supply to reservoirs of increasingly importance. Sediment fingerprinting techniques can be used to determine the relative contributions of different sources of sediment accumulating in reservoirs. The objective of this research is to compare geological and statistical approaches to element selection for sediment fingerprinting modelling. Time-integrated samplers (n=45) were used to obtain source samples from four major subcatchments flowing into the Baroon Pocket Dam in South East Queensland, Australia. The geochemistry of potential sources were compared to the geochemistry of sediment cores (n=12) sampled in the reservoir. The geochemical approach selected elements for modelling that provided expected, observed and statistical discrimination between sediment sources. Two statistical approaches selected elements for modelling with the Kruskal-Wallis H-test and Discriminatory Function Analysis (DFA). In particular, two different significance levels (0.05 & 0.35) for the DFA were included to investigate the importance of element selection on modelling results. A distribution model determined the relative contributions of different sources to sediment sampled in the Baroon Pocket Dam. Elemental discrimination was expected between one subcatchment (Obi Obi Creek) and the remaining subcatchments (Lexys, Falls and Bridge Creek). Six major elements were expected to provide discrimination. Of these six, only Fe2O3 and SiO2 provided expected, observed and statistical discrimination. Modelling results with this geological approach indicated 36% (+/- 9%) of sediment sampled in the reservoir cores were from mafic-derived sources and 64% (+/- 9%) were from felsic-derived sources. The geological and the first statistical approach (DFA0.05) differed by only 1% (σ 5%) for 5 out of 6 model groupings with only the Lexys Creek modelling results differing significantly (35%). The statistical model with expanded elemental selection (DFA0.35) differed from the geological model by an average of 30% for all 6 models. Elemental selection for sediment fingerprinting therefore has the potential to impact modeling results. Accordingly is important to incorporate both robust geological and statistical approaches when selecting elements for sediment fingerprinting. For the Baroon Pocket Dam, management should focus on reducing the supply of sediments derived from felsic sources in each of the subcatchments.
The Ups and Downs of Repeated Cleavage and Internal Fragment Production in Top-Down Proteomics.
Lyon, Yana A; Riggs, Dylan; Fornelli, Luca; Compton, Philip D; Julian, Ryan R
2018-01-01
Analysis of whole proteins by mass spectrometry, or top-down proteomics, has several advantages over methods relying on proteolysis. For example, proteoforms can be unambiguously identified and examined. However, from a gas-phase ion-chemistry perspective, proteins are enormous molecules that present novel challenges relative to peptide analysis. Herein, the statistics of cleaving the peptide backbone multiple times are examined to evaluate the inherent propensity for generating internal versus terminal ions. The raw statistics reveal an inherent bias favoring production of terminal ions, which holds true regardless of protein size. Importantly, even if the full suite of internal ions is generated by statistical dissociation, terminal ions are predicted to account for at least 50% of the total ion current, regardless of protein size, if there are three backbone dissociations or fewer. Top-down analysis should therefore be a viable approach for examining proteins of significant size. Comparison of the purely statistical analysis with actual top-down data derived from ultraviolet photodissociation (UVPD) and higher-energy collisional dissociation (HCD) reveals that terminal ions account for much of the total ion current in both experiments. Terminal ion production is more favored in UVPD relative to HCD, which is likely due to differences in the mechanisms controlling fragmentation. Importantly, internal ions are not found to dominate from either the theoretical or experimental point of view. Graphical abstract ᅟ.
The Ups and Downs of Repeated Cleavage and Internal Fragment Production in Top-Down Proteomics
NASA Astrophysics Data System (ADS)
Lyon, Yana A.; Riggs, Dylan; Fornelli, Luca; Compton, Philip D.; Julian, Ryan R.
2018-01-01
Analysis of whole proteins by mass spectrometry, or top-down proteomics, has several advantages over methods relying on proteolysis. For example, proteoforms can be unambiguously identified and examined. However, from a gas-phase ion-chemistry perspective, proteins are enormous molecules that present novel challenges relative to peptide analysis. Herein, the statistics of cleaving the peptide backbone multiple times are examined to evaluate the inherent propensity for generating internal versus terminal ions. The raw statistics reveal an inherent bias favoring production of terminal ions, which holds true regardless of protein size. Importantly, even if the full suite of internal ions is generated by statistical dissociation, terminal ions are predicted to account for at least 50% of the total ion current, regardless of protein size, if there are three backbone dissociations or fewer. Top-down analysis should therefore be a viable approach for examining proteins of significant size. Comparison of the purely statistical analysis with actual top-down data derived from ultraviolet photodissociation (UVPD) and higher-energy collisional dissociation (HCD) reveals that terminal ions account for much of the total ion current in both experiments. Terminal ion production is more favored in UVPD relative to HCD, which is likely due to differences in the mechanisms controlling fragmentation. Importantly, internal ions are not found to dominate from either the theoretical or experimental point of view. [Figure not available: see fulltext.
Random forests for classification in ecology
Cutler, D.R.; Edwards, T.C.; Beard, K.H.; Cutler, A.; Hess, K.T.; Gibson, J.; Lawler, J.J.
2007-01-01
Classification procedures are some of the most widely used statistical methods in ecology. Random forests (RF) is a new and powerful statistical classifier that is well established in other disciplines but is relatively unknown in ecology. Advantages of RF compared to other statistical classifiers include (1) very high classification accuracy; (2) a novel method of determining variable importance; (3) ability to model complex interactions among predictor variables; (4) flexibility to perform several types of statistical data analysis, including regression, classification, survival analysis, and unsupervised learning; and (5) an algorithm for imputing missing values. We compared the accuracies of RF and four other commonly used statistical classifiers using data on invasive plant species presence in Lava Beds National Monument, California, USA, rare lichen species presence in the Pacific Northwest, USA, and nest sites for cavity nesting birds in the Uinta Mountains, Utah, USA. We observed high classification accuracy in all applications as measured by cross-validation and, in the case of the lichen data, by independent test data, when comparing RF to other common classification methods. We also observed that the variables that RF identified as most important for classifying invasive plant species coincided with expectations based on the literature. ?? 2007 by the Ecological Society of America.
NASA Astrophysics Data System (ADS)
Aouaini, F.; Knani, S.; Ben Yahia, M.; Ben Lamine, A.
2015-08-01
Water sorption isotherms of foodstuffs are very important in different areas of food science engineering such as for design, modeling and optimization of many processes. The equilibrium moisture content is an important parameter in models used to predict changes in the moisture content of a product during storage. A formulation of multilayer model with two energy levels was based on statistical physics and theoretical considerations. Thanks to the grand canonical ensemble in statistical physics. Some physicochemical parameters related to the adsorption process were introduced in the analytical model expression. The data tabulated in literature of water adsorption at different temperatures on: chickpea seeds, lentil seeds, potato and on green peppers were described applying the most popular models applied in food science. We also extend the study to the newest proposed model. It is concluded that among studied models the proposed model seems to be the best for description of data in the whole range of relative humidity. By using our model, we were able to determine the thermodynamic functions. The measurement of desorption isotherms, in particular a gas over a solid porous, allows access to the distribution of pore size PSD.
Schlichting, Margaret L; Guarino, Katharine F; Schapiro, Anna C; Turk-Browne, Nicholas B; Preston, Alison R
2017-01-01
Despite the importance of learning and remembering across the lifespan, little is known about how the episodic memory system develops to support the extraction of associative structure from the environment. Here, we relate individual differences in volumes along the hippocampal long axis to performance on statistical learning and associative inference tasks-both of which require encoding associations that span multiple episodes-in a developmental sample ranging from ages 6 to 30 years. Relating age to volume, we found dissociable patterns across the hippocampal long axis, with opposite nonlinear volume changes in the head and body. These structural differences were paralleled by performance gains across the age range on both tasks, suggesting improvements in the cross-episode binding ability from childhood to adulthood. Controlling for age, we also found that smaller hippocampal heads were associated with superior behavioral performance on both tasks, consistent with this region's hypothesized role in forming generalized codes spanning events. Collectively, these results highlight the importance of examining hippocampal development as a function of position along the hippocampal axis and suggest that the hippocampal head is particularly important in encoding associative structure across development.
Taylor, Kirsten I.; Devereux, Barry J.; Acres, Kadia; Randall, Billi; Tyler, Lorraine K.
2013-01-01
Conceptual representations are at the heart of our mental lives, involved in every aspect of cognitive functioning. Despite their centrality, a long-standing debate persists as to how the meanings of concepts are represented and processed. Many accounts agree that the meanings of concrete concepts are represented by their individual features, but disagree about the importance of different feature-based variables: some views stress the importance of the information carried by distinctive features in conceptual processing, others the features which are shared over many concepts, and still others the extent to which features co-occur. We suggest that previously disparate theoretical positions and experimental findings can be unified by an account which claims that task demands determine how concepts are processed in addition to the effects of feature distinctiveness and co-occurrence. We tested these predictions in a basic-level naming task which relies on distinctive feature information (Experiment 1) and a domain decision task which relies on shared feature information (Experiment 2). Both used large-scale regression designs with the same visual objects, and mixed-effects models incorporating participant, session, stimulus-related and feature statistic variables to model the performance. We found that concepts with relatively more distinctive and more highly correlated distinctive relative to shared features facilitated basic-level naming latencies, while concepts with relatively more shared and more highly correlated shared relative to distinctive features speeded domain decisions. These findings demonstrate that the feature statistics of distinctiveness (shared vs. distinctive) and correlational strength, as well as the task demands, determine how concept meaning is processed in the conceptual system. PMID:22137770
McAlinden, Colm; Khadka, Jyoti; Pesudovs, Konrad
2011-07-01
The ever-expanding choice of ocular metrology and imaging equipment has driven research into the validity of their measurements. Consequently, studies of the agreement between two instruments or clinical tests have proliferated in the ophthalmic literature. It is important that researchers apply the appropriate statistical tests in agreement studies. Correlation coefficients are hazardous and should be avoided. The 'limits of agreement' method originally proposed by Altman and Bland in 1983 is the statistical procedure of choice. Its step-by-step use and practical considerations in relation to optometry and ophthalmology are detailed in addition to sample size considerations and statistical approaches to precision (repeatability or reproducibility) estimates. Ophthalmic & Physiological Optics © 2011 The College of Optometrists.
Characteristics and contributory causes related to large truck crashes (phase II) : all crashes.
DOT National Transportation Integrated Search
2012-03-01
Statistics clearly demonstrate that large-truck crashes contribute to a significant percentage of high-severity crashes. It is : therefore important for the highway safety community to identify the characteristics and contributory causes of these typ...
Game Location and Team Quality Effects on Performance Profiles in Professional Soccer
Lago-Peñas, Carlos; Lago-Ballesteros, Joaquin
2011-01-01
Home advantage in team sports has an important role in determining the outcome of a game. The aim of the present study was to identify the soccer game- related statistics that best discriminate home and visiting teams according to the team quality. The sample included all 380 games of the Spanish professional men’s league. The independent variables were game location (home or away) and the team quality. Teams were classified into four groups according to their final ranking at the end of the league. The game-related statistics registered were divided into three groups: (i) variables related to goals scored; (ii) variables related to offense and (iii) variables related to defense. A univariate (t-test and Mann-Whitney U) and multivariate (discriminant analysis) analysis of data was done. Results showed that home teams have significantly higher means for goal scored, total shots, shots on goal, attacking moves, box moves, crosses, offsides committed, assists, passes made, successful passes, dribbles made, successful dribbles, ball possession, and gains of possession, while visiting teams presented higher means for losses of possession and yellow cards. In addition, the findings of the current study confirm that game location and team quality are important in determining technical and tactical performances in matches. Teams described as superior and those described as inferior did not experience the same home advantage. Future research should consider the influence of other confounding variables such as weather conditions, game status and team form. Key points Home teams have significantly higher figures for attack indicators probably due to facilities familiarity and crowd effects. The teams’ game-related statistics profile varied according to game location and team quality. Teams described as superior and those described as inferior did not experience the same home advantage. PMID:24150619
Reynolds number dependence of relative dispersion statistics in isotropic turbulence
NASA Astrophysics Data System (ADS)
Sawford, Brian L.; Yeung, P. K.; Hackl, Jason F.
2008-06-01
Direct numerical simulation results for a range of relative dispersion statistics over Taylor-scale Reynolds numbers up to 650 are presented in an attempt to observe and quantify inertial subrange scaling and, in particular, Richardson's t3 law. The analysis includes the mean-square separation and a range of important but less-studied differential statistics for which the motion is defined relative to that at time t =0. It seeks to unambiguously identify and quantify the Richardson scaling by demonstrating convergence with both the Reynolds number and initial separation. According to these criteria, the standard compensated plots for these statistics in inertial subrange scaling show clear evidence of a Richardson range but with an imprecise estimate for the Richardson constant. A modified version of the cube-root plots introduced by Ott and Mann [J. Fluid Mech. 422, 207 (2000)] confirms such convergence. It has been used to yield more precise estimates for Richardson's constant g which decrease with Taylor-scale Reynolds numbers over the range of 140-650. Extrapolation to the large Reynolds number limit gives an asymptotic value for Richardson's constant in the range g =0.55-0.57, depending on the functional form used to make the extrapolation.
ERIC Educational Resources Information Center
Hudson, Lisa; Bhandari, Rajika; Peter, Katharin; Bills, David B.
2005-01-01
Of the many purposes education serves in society, one of the most important is to prepare people for work. In today's economy, education is important not just to help adults enter the labor market, but also to ensure that adults remain marketable throughout their working lives. This report examines how adults in the labor force use formal…
Local multiplicity adjustment for the spatial scan statistic using the Gumbel distribution.
Gangnon, Ronald E
2012-03-01
The spatial scan statistic is an important and widely used tool for cluster detection. It is based on the simultaneous evaluation of the statistical significance of the maximum likelihood ratio test statistic over a large collection of potential clusters. In most cluster detection problems, there is variation in the extent of local multiplicity across the study region. For example, using a fixed maximum geographic radius for clusters, urban areas typically have many overlapping potential clusters, whereas rural areas have relatively few. The spatial scan statistic does not account for local multiplicity variation. We describe a previously proposed local multiplicity adjustment based on a nested Bonferroni correction and propose a novel adjustment based on a Gumbel distribution approximation to the distribution of a local scan statistic. We compare the performance of all three statistics in terms of power and a novel unbiased cluster detection criterion. These methods are then applied to the well-known New York leukemia dataset and a Wisconsin breast cancer incidence dataset. © 2011, The International Biometric Society.
Local multiplicity adjustment for the spatial scan statistic using the Gumbel distribution
Gangnon, Ronald E.
2011-01-01
Summary The spatial scan statistic is an important and widely used tool for cluster detection. It is based on the simultaneous evaluation of the statistical significance of the maximum likelihood ratio test statistic over a large collection of potential clusters. In most cluster detection problems, there is variation in the extent of local multiplicity across the study region. For example, using a fixed maximum geographic radius for clusters, urban areas typically have many overlapping potential clusters, while rural areas have relatively few. The spatial scan statistic does not account for local multiplicity variation. We describe a previously proposed local multiplicity adjustment based on a nested Bonferroni correction and propose a novel adjustment based on a Gumbel distribution approximation to the distribution of a local scan statistic. We compare the performance of all three statistics in terms of power and a novel unbiased cluster detection criterion. These methods are then applied to the well-known New York leukemia dataset and a Wisconsin breast cancer incidence dataset. PMID:21762118
Spence, Jeffrey S; Brier, Matthew R; Hart, John; Ferree, Thomas C
2013-03-01
Linear statistical models are used very effectively to assess task-related differences in EEG power spectral analyses. Mixed models, in particular, accommodate more than one variance component in a multisubject study, where many trials of each condition of interest are measured on each subject. Generally, intra- and intersubject variances are both important to determine correct standard errors for inference on functions of model parameters, but it is often assumed that intersubject variance is the most important consideration in a group study. In this article, we show that, under common assumptions, estimates of some functions of model parameters, including estimates of task-related differences, are properly tested relative to the intrasubject variance component only. A substantial gain in statistical power can arise from the proper separation of variance components when there is more than one source of variability. We first develop this result analytically, then show how it benefits a multiway factoring of spectral, spatial, and temporal components from EEG data acquired in a group of healthy subjects performing a well-studied response inhibition task. Copyright © 2011 Wiley Periodicals, Inc.
Detailed Analysis of the Interoccurrence Time Statistics in Seismic Activity
NASA Astrophysics Data System (ADS)
Tanaka, Hiroki; Aizawa, Yoji
2017-02-01
The interoccurrence time statistics of seismiciry is studied theoretically as well as numerically by taking into account the conditional probability and the correlations among many earthquakes in different magnitude levels. It is known so far that the interoccurrence time statistics is well approximated by the Weibull distribution, but the more detailed information about the interoccurrence times can be obtained from the analysis of the conditional probability. Firstly, we propose the Embedding Equation Theory (EET), where the conditional probability is described by two kinds of correlation coefficients; one is the magnitude correlation and the other is the inter-event time correlation. Furthermore, the scaling law of each correlation coefficient is clearly determined from the numerical data-analysis carrying out with the Preliminary Determination of Epicenter (PDE) Catalog and the Japan Meteorological Agency (JMA) Catalog. Secondly, the EET is examined to derive the magnitude dependence of the interoccurrence time statistics and the multi-fractal relation is successfully formulated. Theoretically we cannot prove the universality of the multi-fractal relation in seismic activity; nevertheless, the theoretical results well reproduce all numerical data in our analysis, where several common features or the invariant aspects are clearly observed. Especially in the case of stationary ensembles the multi-fractal relation seems to obey an invariant curve, furthermore in the case of non-stationary (moving time) ensembles for the aftershock regime the multi-fractal relation seems to satisfy a certain invariant curve at any moving times. It is emphasized that the multi-fractal relation plays an important role to unify the statistical laws of seismicity: actually the Gutenberg-Richter law and the Weibull distribution are unified in the multi-fractal relation, and some universality conjectures regarding the seismicity are briefly discussed.
Investigation of Polarization Phase Difference Related to Forest Fields Characterizations
NASA Astrophysics Data System (ADS)
Majidi, M.; Maghsoudi, Y.
2013-09-01
The information content of Synthetic Aperture Radar (SAR) data significantly included in the radiometric polarization channels, hence polarimetric SAR data should be analyzed in relation with target structure. The importance of the phase difference between two co-polarized scattered signals due to the possible association between the biophysical parameters and the measured Polarization Phase Difference (PPD) statistics of the backscattered signal recorded components has been recognized in geophysical remote sensing. This paper examines two Radarsat-2 images statistics of the phase difference to describe the feasibility of relationship with the physical properties of scattering targets and tries to understand relevance of PPD statistics with various types of forest fields. As well as variation of incidence angle due to affecting on PPD statistics is investigated. The experimental forest pieces that are used in this research are characterized white pine (Pinus strobus L.), red pine (Pinus resinosa Ait.), jack pine (Pinus banksiana Lamb.), white spruce (Picea glauca (Moench Voss), black spruce (Picea mariana (Mill) B.S.P.), poplar (Populus L.), red oak (Quercus rubra L.) , aspen and ground vegetation. The experimental results show that despite of biophysical parameters have a wide diversity, PPD statistics are almost the same. Forest fields distributions as distributed targets have close to zero means regardless of the incidence angle. Also, The PPD distribution are function of both target and sensor parameters, but for more appropriate examination related to PPD statistics the observations should made in the leaf-off season or in bands with lower frequencies.
A phylogenetic transform enhances analysis of compositional microbiota data.
Silverman, Justin D; Washburne, Alex D; Mukherjee, Sayan; David, Lawrence A
2017-02-15
Surveys of microbial communities (microbiota), typically measured as relative abundance of species, have illustrated the importance of these communities in human health and disease. Yet, statistical artifacts commonly plague the analysis of relative abundance data. Here, we introduce the PhILR transform, which incorporates microbial evolutionary models with the isometric log-ratio transform to allow off-the-shelf statistical tools to be safely applied to microbiota surveys. We demonstrate that analyses of community-level structure can be applied to PhILR transformed data with performance on benchmarks rivaling or surpassing standard tools. Additionally, by decomposing distance in the PhILR transformed space, we identified neighboring clades that may have adapted to distinct human body sites. Decomposing variance revealed that covariation of bacterial clades within human body sites increases with phylogenetic relatedness. Together, these findings illustrate how the PhILR transform combines statistical and phylogenetic models to overcome compositional data challenges and enable evolutionary insights relevant to microbial communities.
The psychophysiology of real-time financial risk processing.
Lo, Andrew W; Repin, Dmitry V
2002-04-01
A longstanding controversy in economics and finance is whether financial markets are governed by rational forces or by emotional responses. We study the importance of emotion in the decision-making process of professional securities traders by measuring their physiological characteristics (e.g., skin conductance, blood volume pulse, etc.) during live trading sessions while simultaneously capturing real-time prices from which market events can be detected. In a sample of 10 traders, we find statistically significant differences in mean electrodermal responses during transient market events relative to no-event control periods, and statistically significant mean changes in cardiovascular variables during periods of heightened market volatility relative to normal-volatility control periods. We also observe significant differences in these physiological responses across the 10 traders that may be systematically related to the traders' levels of experience.
Probability Issues in without Replacement Sampling
ERIC Educational Resources Information Center
Joarder, A. H.; Al-Sabah, W. S.
2007-01-01
Sampling without replacement is an important aspect in teaching conditional probabilities in elementary statistics courses. Different methods proposed in different texts for calculating probabilities of events in this context are reviewed and their relative merits and limitations in applications are pinpointed. An alternative representation of…
Random elements on lattices: Review and statistical applications
NASA Astrophysics Data System (ADS)
Potocký, Rastislav; Villarroel, Claudia Navarro; Sepúlveda, Maritza; Luna, Guillermo; Stehlík, Milan
2017-07-01
We discuss important contributions to random elements on lattices. We relate to both algebraic and probabilistic properties. Several applications and concepts are discussed, e.g. positive dependence, Random walks and distributions on lattices, Super-lattices, learning. The application to Chilean Ecology is given.
Laplace approximation for Bessel functions of matrix argument
NASA Astrophysics Data System (ADS)
Butler, Ronald W.; Wood, Andrew T. A.
2003-06-01
We derive Laplace approximations to three functions of matrix argument which arise in statistics and elsewhere: matrix Bessel A[nu]; matrix Bessel B[nu]; and the type II confluent hypergeometric function of matrix argument, [Psi]. We examine the theoretical and numerical properties of the approximations. On the theoretical side, it is shown that the Laplace approximations to A[nu], B[nu] and [Psi] given here, together with the Laplace approximations to the matrix argument functions 1F1 and 2F1 presented in Butler and Wood (Laplace approximations to hyper-geometric functions with matrix argument, Ann. Statist. (2002)), satisfy all the important confluence relations and symmetry relations enjoyed by the original functions.
NASA Astrophysics Data System (ADS)
Poulos, M. J.; Pierce, J. L.; McNamara, J. P.; Flores, A. N.; Benner, S. G.
2015-12-01
Terrain aspect alters the spatial distribution of insolation across topography, driving eco-pedo-hydro-geomorphic feedbacks that can alter landform evolution and result in valley asymmetries for a suite of land surface characteristics (e.g. slope length and steepness, vegetation, soil properties, and drainage development). Asymmetric valleys serve as natural laboratories for studying how landscapes respond to climate perturbation. In the semi-arid montane granodioritic terrain of the Idaho batholith, Northern Rocky Mountains, USA, prior works indicate that reduced insolation on northern (pole-facing) aspects prolongs snow pack persistence, and is associated with thicker, finer-grained soils, that retain more water, prolong the growing season, support coniferous forest rather than sagebrush steppe ecosystems, stabilize slopes at steeper angles, and produce sparser drainage networks. We hypothesize that the primary drivers of valley asymmetry development are changes in the pedon-scale water-balance that coalesce to alter catchment-scale runoff and drainage development, and ultimately cause the divide between north and south-facing land surfaces to migrate northward. We explore this conceptual framework by coupling land surface analyses with statistical modeling to assess relationships and the relative importance of land surface characteristics. Throughout the Idaho batholith, we systematically mapped and tabulated various statistical measures of landforms, land cover, and hydroclimate within discrete valley segments (n=~10,000). We developed a random forest based statistical model to predict valley slope asymmetry based upon numerous measures (n>300) of landscape asymmetries. Preliminary results suggest that drainages are tightly coupled with hillslopes throughout the region, with drainage-network slope being one of the strongest predictors of land-surface-averaged slope asymmetry. When slope-related statistics are excluded, due to possible autocorrelation, valley slope asymmetry is most strongly predicted by asymmetries of insolation and drainage density, which generally supports a water-balance based conceptual model of valley asymmetry development. Surprisingly, vegetation asymmetries had relatively low predictive importance.
Samadzadeh, Gholam Reza; Rigi, Tahereh; Ganjali, Ali Reza
2013-01-01
Surveying valuable and most recent information from internet, has become vital for researchers and scholars, because every day, thousands and perhaps millions of scientific works are brought out as digital resources which represented by internet and researchers can't ignore this great resource to find related documents for their literature search, which may not be found in any library. With regard to variety of documents presented on the internet, search engines are one of the most effective search tools for finding information. The aim of this study is to evaluate the three criteria, recall, preciseness and importance of the four search engines which are PubMed, Science Direct, Google Scholar and federated search of Iranian National Medical Digital Library in addiction (prevention and treatment) to select the most effective search engine for offering the best literature research. This research was a cross-sectional study by which four popular search engines in medical sciences were evaluated. To select keywords, medical subject heading (Mesh) was used. We entered given keywords in the search engines and after searching, 10 first entries were evaluated. Direct observation was used as a mean for data collection and they were analyzed by descriptive statistics (number, percent number and mean) and inferential statistics, One way analysis of variance (ANOVA) and post hoc Tukey in Spss. 15 statistical software. P Value < 0.05 was considered statistically significant. Results have shown that the search engines had different operations with regard to the evaluated criteria. Since P Value was 0.004 < 0.05 for preciseness and was 0.002 < 0.05 for importance, it shows significant difference among search engines. PubMed, Science Direct and Google Scholar were the best in recall, preciseness and importance respectively. As literature research is one of the most important stages of research, it's better for researchers, especially Substance-Related Disorders scholars to use different search engines with the best recall, preciseness and importance in that subject field to reach desirable results while searching and they don't depend on just one search engine.
Samadzadeh, Gholam Reza; Rigi, Tahereh; Ganjali, Ali Reza
2013-01-01
Background Surveying valuable and most recent information from internet, has become vital for researchers and scholars, because every day, thousands and perhaps millions of scientific works are brought out as digital resources which represented by internet and researchers can’t ignore this great resource to find related documents for their literature search, which may not be found in any library. With regard to variety of documents presented on the internet, search engines are one of the most effective search tools for finding information. Objectives The aim of this study is to evaluate the three criteria, recall, preciseness and importance of the four search engines which are PubMed, Science Direct, Google Scholar and federated search of Iranian National Medical Digital Library in addiction (prevention and treatment) to select the most effective search engine for offering the best literature research. Materials and Methods This research was a cross-sectional study by which four popular search engines in medical sciences were evaluated. To select keywords, medical subject heading (Mesh) was used. We entered given keywords in the search engines and after searching, 10 first entries were evaluated. Direct observation was used as a mean for data collection and they were analyzed by descriptive statistics (number, percent number and mean) and inferential statistics, One way analysis of variance (ANOVA) and post hoc Tukey in Spss. 15 statistical software. P Value < 0.05 was considered statistically significant. Results Results have shown that the search engines had different operations with regard to the evaluated criteria. Since P Value was 0.004 < 0.05 for preciseness and was 0.002 < 0.05 for importance, it shows significant difference among search engines. PubMed, Science Direct and Google Scholar were the best in recall, preciseness and importance respectively. Conclusions As literature research is one of the most important stages of research, it's better for researchers, especially Substance-Related Disorders scholars to use different search engines with the best recall, preciseness and importance in that subject field to reach desirable results while searching and they don’t depend on just one search engine. PMID:24971257
Low-level contrast statistics are diagnostic of invariance of natural textures
Groen, Iris I. A.; Ghebreab, Sennay; Lamme, Victor A. F.; Scholte, H. Steven
2012-01-01
Texture may provide important clues for real world object and scene perception. To be reliable, these clues should ideally be invariant to common viewing variations such as changes in illumination and orientation. In a large image database of natural materials, we found textures with low-level contrast statistics that varied substantially under viewing variations, as well as textures that remained relatively constant. This led us to ask whether textures with constant contrast statistics give rise to more invariant representations compared to other textures. To test this, we selected natural texture images with either high (HV) or low (LV) variance in contrast statistics and presented these to human observers. In two distinct behavioral categorization paradigms, participants more often judged HV textures as “different” compared to LV textures, showing that textures with constant contrast statistics are perceived as being more invariant. In a separate electroencephalogram (EEG) experiment, evoked responses to single texture images (single-image ERPs) were collected. The results show that differences in contrast statistics correlated with both early and late differences in occipital ERP amplitude between individual images. Importantly, ERP differences between images of HV textures were mainly driven by illumination angle, which was not the case for LV images: there, differences were completely driven by texture membership. These converging neural and behavioral results imply that some natural textures are surprisingly invariant to illumination changes and that low-level contrast statistics are diagnostic of the extent of this invariance. PMID:22701419
National policies for technical change: Where are the increasing returns to economic research?
Pavitt, Keith
1996-01-01
Improvements over the past 30 years in statistical data, analysis, and related theory have strengthened the basis for science and technology policy by confirming the importance of technical change in national economic performance. But two important features of scientific and technological activities in the Organization for Economic Cooperation and Development countries are still not addressed adequately in mainstream economics: (i) the justification of public funding for basic research and (ii) persistent international differences in investment in research and development and related activities. In addition, one major gap is now emerging in our systems of empirical measurement—the development of software technology, especially in the service sector. There are therefore dangers of diminishing returns to the usefulness of economic research, which continues to rely completely on established theory and established statistical sources. Alternative propositions that deserve serious consideration are: (i) the economic usefulness of basic research is in the provision of (mainly tacit) skills rather than codified and applicable information; (ii) in developing and exploiting technological opportunities, institutional competencies are just as important as the incentive structures that they face; and (iii) software technology developed in traditional service sectors may now be a more important locus of technical change than software technology developed in “high-tech” manufacturing. PMID:8917481
MODELING FISH AND SHELLFISH DISTRIBUTIONS IN THE MOBILE BAY ESTUARY, USA
Estuaries in the Gulf of Mexico provide rich habitat for many fish and shellfish, including those that have been identified as economically and ecologically important. For the Mobile Bay estuary, we developed statistical models to relate distributions of individual species and sp...
Taylor, Kirsten I; Devereux, Barry J; Acres, Kadia; Randall, Billi; Tyler, Lorraine K
2012-03-01
Conceptual representations are at the heart of our mental lives, involved in every aspect of cognitive functioning. Despite their centrality, a long-standing debate persists as to how the meanings of concepts are represented and processed. Many accounts agree that the meanings of concrete concepts are represented by their individual features, but disagree about the importance of different feature-based variables: some views stress the importance of the information carried by distinctive features in conceptual processing, others the features which are shared over many concepts, and still others the extent to which features co-occur. We suggest that previously disparate theoretical positions and experimental findings can be unified by an account which claims that task demands determine how concepts are processed in addition to the effects of feature distinctiveness and co-occurrence. We tested these predictions in a basic-level naming task which relies on distinctive feature information (Experiment 1) and a domain decision task which relies on shared feature information (Experiment 2). Both used large-scale regression designs with the same visual objects, and mixed-effects models incorporating participant, session, stimulus-related and feature statistic variables to model the performance. We found that concepts with relatively more distinctive and more highly correlated distinctive relative to shared features facilitated basic-level naming latencies, while concepts with relatively more shared and more highly correlated shared relative to distinctive features speeded domain decisions. These findings demonstrate that the feature statistics of distinctiveness (shared vs. distinctive) and correlational strength, as well as the task demands, determine how concept meaning is processed in the conceptual system. Copyright © 2011 Elsevier B.V. All rights reserved.
Factors that influence the use and perceptions of employee assistance programs at six worksites.
French, M T; Dunlap, L J; Roman, P M; Steele, P D
1997-10-01
Employee assistance programs (EAPs) have gained significant importance in contemporary worksites. This article uses data from 6 case studies to examine several research questions regarding the relationship between worker demographic (e.g., gender, job tenure, and marital status), substance use, and workplace policies and the actual and potential use of the company EAP. Unlike in most of the existing literature, the authors did not find that gender, marital status, or job dissatisfaction are statistically related to actual or potential EAP use at most worksites. However, job tenure and some substance use behaviors were related to actual EAP use in a positive and statistically significant way. Another important finding, underlying the credible integration of EAPs into worksite culture, is the positive and robust relationship between employee trust and confidence in the EAP and actual use. The results of our study both reinforce some long-established principles in the EAP field and encourage further consideration of other beliefs.
Cleary, Timothy J; Velardi, Brittany; Schnaidman, Bracha
2017-10-01
The current study examined the effectiveness of an applied self-regulated learning intervention (Self-Regulation Empowerment Program (SREP)) relative to an existing, school-based remedial mathematics intervention for improving the motivation, strategic skills, and mathematics achievement of academically at-risk middle school students. Although significant group differences in student self-regulated learning (SRL) were not observed when using self-report questionnaires, medium to large and statistically significant group differences were observed across several contextualized, situation-specific measures of strategic and regulatory thinking. The SREP group also exhibited a statistically significant and more positive trend in achievement scores over two years in middle school relative to the comparison condition. Finally, SREP students and coaches reported SREP to be a socially-valid intervention, in terms of acceptability and importance. The importance of this study and critical areas for future research are highlighted and discussed. Copyright © 2017 Society for the Study of School Psychology. Published by Elsevier Ltd. All rights reserved.
Rushton, Paul R P; Grevitt, Michael P
2013-04-20
Review and statistical analysis of studies evaluating health-related quality of life (HRQOL) in adolescents with untreated adolescent idiopathic scoliosis (AIS) using Scoliosis Research Society (SRS) outcomes. To apply normative values and minimum clinical important differences for the SRS-22r to the literature. Identify whether the HRQOL of adolescents with untreated AIS differs from unaffected peers and whether any differences are clinically relevant. The effect of untreated AIS on adolescent HRQOL is uncertain. The lack of published normative values and minimum clinical important difference for the SRS-22r has so far hindered our interpretation of previous studies. The publication of this background data allows these studies to be re-examined. Using suitable inclusion criteria, a literature search identified studies examining HRQOL in untreated adolescents with AIS. Each cohort was analyzed individually. Statistically significant differences were identified by using 95% confidence intervals for the difference in SRS-22r domain mean scores between the cohorts with AIS and the published data for unaffected adolescents. If the lower bound of the confidence interval was greater than the minimum clinical important difference, the difference was considered clinically significant. Of the 21 included patient cohorts, 81% reported statistically worse pain than those unaffected. Yet in only 5% of cohorts was this difference clinically important. Of the 11 cohorts included examining patient self-image, 91% reported statistically worse scores than those unaffected. In 73% of cohorts this difference was clinically significant. Affected cohorts tended to score well in function/activity and mental health domains and differences from those unaffected rarely reached clinically significant values. Pain and self-image tend to be statistically lower among cohorts with AIS than those unaffected. The literature to date suggests that it is only self-image which consistently differs clinically. This should be considered when assessing the possible benefits of surgery.
Health tourism on the rise? Evidence from the Balance of Payments Statistics.
Loh, Chung-Ping A
2014-09-01
The study assesses the presence and magnitude of global trends in health tourism using health-related travel (HRT) spending reported in the International Monetary Fund's Balance of Payments Statistics database. Linear regression and quantile regression are applied to estimate secular trends of the import and export of HRT based on a sample of countries from 2003 to 2009. The results show that from 2003 to 2009 the import and export of health tourism rose among countries with a high volume of such activities (accounting for the upper 40% of the countries), but not among those with a low volume. The uneven growth in health tourism has generated greater contrast between countries with high and low volumes of health tourism activities. However, the growth in the total import of health tourism did not outpace the population growth, implying that in general the population's tendency to engage in health tourism remained static.
Rank and independence in contingency table
NASA Astrophysics Data System (ADS)
Tsumoto, Shusaku
2004-04-01
A contingency table summarizes the conditional frequencies of two attributes and shows how these two attributes are dependent on each other. Thus, this table is a fundamental tool for pattern discovery with conditional probabilities, such as rule discovery. In this paper, a contingency table is interpreted from the viewpoint of statistical independence and granular computing. The first important observation is that a contingency table compares two attributes with respect to the number of equivalence classes. For example, a n x n table compares two attributes with the same granularity, while a m x n(m >= n) table compares two attributes with different granularities. The second important observation is that matrix algebra is a key point of analysis of this table. Especially, the degree of independence, rank plays a very important role in evaluating the degree of statistical independence. Relations between rank and the degree of dependence are also investigated.
Identification of curriculum content for a renewable energy graduate degree program
NASA Astrophysics Data System (ADS)
Haughery, John R.
There currently exists a disconnect between renewable energy industry workforce needs and academic program proficiencies. This is evidenced by an absence of clear curriculum content on renewable energy graduate program websites. The purpose of this study was to identify a set of curriculum content for graduate degrees in renewable energy. At the conclusion, a clear list of 42 content items was identified and statistically ranked. The content items identified were based on a review of literature from government initiatives, professional society's body of knowledge, and related research studies. Leaders and experts in the field of renewable energy and sustainability were surveyed, using a five-point Likert-Scale model. This allowed each item's importance level to be analyzed and prioritized based on non-parametric statistical analysis methods. The study found seven competency items to be very important , 30 to be important, and five to be somewhat important. The results were also appropriate for use as a framework in developing or improving renewable energy graduate programs.
Evidential Value That Exercise Improves BMI z-Score in Overweight and Obese Children and Adolescents
Kelley, George A.; Kelley, Kristi S.
2015-01-01
Background. Given the cardiovascular disease (CVD) related importance of understanding the true effects of exercise on adiposity in overweight and obese children and adolescents, this study examined whether there is evidential value to rule out excessive and inappropriate reporting of statistically significant results, a major problem in the published literature, with respect to exercise-induced improvements in BMI z-score among overweight and obese children and adolescents. Methods. Using data from a previous meta-analysis of 10 published studies that included 835 overweight and obese children and adolescents, a novel, recently developed approach (p-curve) was used to test for evidential value and rule out selective reporting of findings. Chi-squared tests (χ 2) were used to test for statistical significance with alpha (p) values <0.05 considered statistically significant. Results. Six of 10 findings (60%) were statistically significant. Statistically significant right-skew to rule out selective reporting was found (χ 2 = 38.8, p = 0.0001). Conversely, studies neither lacked evidential value (χ 2 = 6.8, p = 0.87) nor lacked evidential value and were intensely p-hacked (χ 2 = 4.3, p = 0.98). Conclusion. Evidential value results confirm that exercise reduces BMI z-score in overweight and obese children and adolescents, an important therapeutic strategy for treating and preventing CVD. PMID:26509145
Kelley, George A; Kelley, Kristi S
2015-01-01
Background. Given the cardiovascular disease (CVD) related importance of understanding the true effects of exercise on adiposity in overweight and obese children and adolescents, this study examined whether there is evidential value to rule out excessive and inappropriate reporting of statistically significant results, a major problem in the published literature, with respect to exercise-induced improvements in BMI z-score among overweight and obese children and adolescents. Methods. Using data from a previous meta-analysis of 10 published studies that included 835 overweight and obese children and adolescents, a novel, recently developed approach (p-curve) was used to test for evidential value and rule out selective reporting of findings. Chi-squared tests (χ (2)) were used to test for statistical significance with alpha (p) values <0.05 considered statistically significant. Results. Six of 10 findings (60%) were statistically significant. Statistically significant right-skew to rule out selective reporting was found (χ (2) = 38.8, p = 0.0001). Conversely, studies neither lacked evidential value (χ (2) = 6.8, p = 0.87) nor lacked evidential value and were intensely p-hacked (χ (2) = 4.3, p = 0.98). Conclusion. Evidential value results confirm that exercise reduces BMI z-score in overweight and obese children and adolescents, an important therapeutic strategy for treating and preventing CVD.
Effect Size as the Essential Statistic in Developing Methods for mTBI Diagnosis.
Gibson, Douglas Brandt
2015-01-01
The descriptive statistic known as "effect size" measures the distinguishability of two sets of data. Distingishability is at the core of diagnosis. This article is intended to point out the importance of effect size in the development of effective diagnostics for mild traumatic brain injury and to point out the applicability of the effect size statistic in comparing diagnostic efficiency across the main proposed TBI diagnostic methods: psychological, physiological, biochemical, and radiologic. Comparing diagnostic approaches is difficult because different researcher in different fields have different approaches to measuring efficacy. Converting diverse measures to effect sizes, as is done in meta-analysis, is a relatively easy way to make studies comparable.
Comparison of direct numerical simulation databases of turbulent channel flow at Reτ = 180
NASA Astrophysics Data System (ADS)
Vreman, A. W.; Kuerten, J. G. M.
2014-01-01
Direct numerical simulation (DNS) databases are compared to assess the accuracy and reproducibility of standard and non-standard turbulence statistics of incompressible plane channel flow at Reτ = 180. Two fundamentally different DNS codes are shown to produce maximum relative deviations below 0.2% for the mean flow, below 1% for the root-mean-square velocity and pressure fluctuations, and below 2% for the three components of the turbulent dissipation. Relatively fine grids and long statistical averaging times are required. An analysis of dissipation spectra demonstrates that the enhanced resolution is necessary for an accurate representation of the smallest physical scales in the turbulent dissipation. The results are related to the physics of turbulent channel flow in several ways. First, the reproducibility supports the hitherto unproven theoretical hypothesis that the statistically stationary state of turbulent channel flow is unique. Second, the peaks of dissipation spectra provide information on length scales of the small-scale turbulence. Third, the computed means and fluctuations of the convective, pressure, and viscous terms in the momentum equation show the importance of the different forces in the momentum equation relative to each other. The Galilean transformation that leads to minimum peak fluctuation of the convective term is determined. Fourth, an analysis of higher-order statistics is performed. The skewness of the longitudinal derivative of the streamwise velocity is stronger than expected (-1.5 at y+ = 30). This skewness and also the strong near-wall intermittency of the normal velocity are related to coherent structures.
A statistical investigation of the mass discrepancy-acceleration relation
NASA Astrophysics Data System (ADS)
Desmond, Harry
2017-02-01
We use the mass discrepancy-acceleration relation (the correlation between the ratio of total-to-visible mass and acceleration in galaxies; MDAR) to test the galaxy-halo connection. We analyse the MDAR using a set of 16 statistics that quantify its four most important features: shape, scatter, the presence of a `characteristic acceleration scale', and the correlation of its residuals with other galaxy properties. We construct an empirical framework for the galaxy-halo connection in LCDM to generate predictions for these statistics, starting with conventional correlations (halo abundance matching; AM) and introducing more where required. Comparing to the SPARC data, we find that: (1) the approximate shape of the MDAR is readily reproduced by AM, and there is no evidence that the acceleration at which dark matter becomes negligible has less spread in the data than in AM mocks; (2) even under conservative assumptions, AM significantly overpredicts the scatter in the relation and its normalization at low acceleration, and furthermore positions dark matter too close to galaxies' centres on average; (3) the MDAR affords 2σ evidence for an anticorrelation of galaxy size and Hubble type with halo mass or concentration at fixed stellar mass. Our analysis lays the groundwork for a bottom-up determination of the galaxy-halo connection from relations such as the MDAR, provides concrete statistical tests for specific galaxy formation models, and brings into sharper focus the relative evidence accorded by galaxy kinematics to LCDM and modified gravity alternatives.
A statistical investigation of the mass discrepancy–acceleration relation
Desmond, Harry
2016-10-08
We use the mass discrepancy–acceleration relation (the correlation between the ratio of total-to-visible mass and acceleration in galaxies; MDAR) to test the galaxy–halo connection. Here, we analyse the MDAR using a set of 16 statistics that quantify its four most important features: shape, scatter, the presence of a ‘characteristic acceleration scale’, and the correlation of its residuals with other galaxy properties. We construct an empirical framework for the galaxy–halo connection in LCDM to generate predictions for these statistics, starting with conventional correlations (halo abundance matching; AM) and introducing more where required. Comparing to the SPARC data, we find that: (1)more » the approximate shape of the MDAR is readily reproduced by AM, and there is no evidence that the acceleration at which dark matter becomes negligible has less spread in the data than in AM mocks; (2) even under conservative assumptions, AM significantly overpredicts the scatter in the relation and its normalization at low acceleration, and furthermore positions dark matter too close to galaxies’ centres on average; (3) the MDAR affords 2σ evidence for an anticorrelation of galaxy size and Hubble type with halo mass or concentration at fixed stellar mass. Lastly, our analysis lays the groundwork for a bottom-up determination of the galaxy–halo connection from relations such as the MDAR, provides concrete statistical tests for specific galaxy formation models, and brings into sharper focus the relative evidence accorded by galaxy kinematics to LCDM and modified gravity alternatives.« less
Moral foundations in an interacting neural networks society: A statistical mechanics analysis
NASA Astrophysics Data System (ADS)
Vicente, R.; Susemihl, A.; Jericó, J. P.; Caticha, N.
2014-04-01
The moral foundations theory supports that people, across cultures, tend to consider a small number of dimensions when classifying issues on a moral basis. The data also show that the statistics of weights attributed to each moral dimension is related to self-declared political affiliation, which in turn has been connected to cognitive learning styles by the recent literature in neuroscience and psychology. Inspired by these data, we propose a simple statistical mechanics model with interacting neural networks classifying vectors and learning from members of their social neighbourhood about their average opinion on a large set of issues. The purpose of learning is to reduce dissension among agents when disagreeing. We consider a family of learning algorithms parametrized by δ, that represents the importance given to corroborating (same sign) opinions. We define an order parameter that quantifies the diversity of opinions in a group with homogeneous learning style. Using Monte Carlo simulations and a mean field approximation we find the relation between the order parameter and the learning parameter δ at a temperature we associate with the importance of social influence in a given group. In concordance with data, groups that rely more strongly on corroborating evidence sustain less opinion diversity. We discuss predictions of the model and propose possible experimental tests.
Schlichting, Margaret L.; Guarino, Katharine F.; Schapiro, Anna C.; Turk-Browne, Nicholas B.; Preston, Alison R.
2016-01-01
Despite the importance of learning and remembering across the lifespan, little is known about how the episodic memory system develops to support the extraction of associative structure from the environment. Here, we relate individual differences in volumes along the hippocampal long axis to performance on statistical learning and associative inference tasks—both of which require encoding associations that span multiple episodes—in a developmental sample ranging from ages 6–30 years. Relating age to volume, we found dissociable patterns across the hippocampal long axis, with opposite nonlinear volume changes in the head and body. These structural differences were paralleled by performance gains across the age range on both tasks, suggesting improvements in the cross-episode binding ability from childhood to adulthood. Controlling for age, we also found that smaller hippocampal heads were associated with superior behavioral performance on both tasks, consistent with this region’s hypothesized role in forming generalized codes spanning events. Collectively, these results highlight the importance of examining hippocampal development as a function of position along the hippocampal axis and suggest that the hippocampal head is particularly important in encoding associative structure across development. PMID:27575916
Confounding in statistical mediation analysis: What it is and how to address it.
Valente, Matthew J; Pelham, William E; Smyth, Heather; MacKinnon, David P
2017-11-01
Psychology researchers are often interested in mechanisms underlying how randomized interventions affect outcomes such as substance use and mental health. Mediation analysis is a common statistical method for investigating psychological mechanisms that has benefited from exciting new methodological improvements over the last 2 decades. One of the most important new developments is methodology for estimating causal mediated effects using the potential outcomes framework for causal inference. Potential outcomes-based methods developed in epidemiology and statistics have important implications for understanding psychological mechanisms. We aim to provide a concise introduction to and illustration of these new methods and emphasize the importance of confounder adjustment. First, we review the traditional regression approach for estimating mediated effects. Second, we describe the potential outcomes framework. Third, we define what a confounder is and how the presence of a confounder can provide misleading evidence regarding mechanisms of interventions. Fourth, we describe experimental designs that can help rule out confounder bias. Fifth, we describe new statistical approaches to adjust for measured confounders of the mediator-outcome relation and sensitivity analyses to probe effects of unmeasured confounders on the mediated effect. All approaches are illustrated with application to a real counseling intervention dataset. Counseling psychologists interested in understanding the causal mechanisms of their interventions can benefit from incorporating the most up-to-date techniques into their mediation analyses. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Dynamic and thermodynamic processes driving the January 2014 precipitation record in southern UK
NASA Astrophysics Data System (ADS)
Oueslati, B.; Yiou, P.; Jezequel, A.
2017-12-01
Regional extreme precipitation are projected to intensify as a response to planetary climate change, with important impacts on societies. Understanding and anticipating those events remain a major challenge. In this study, we revisit the mechanisms of winter precipitation record that occurred in southern United Kingdom in January 2014. The physical drivers of this event are analyzed using the water vapor budget. Precipitation changes are decomposed into dynamic contributions, related to changes in atmospheric circulation, and thermodynamic contributions, related to changes in water vapor. We attempt to quantify the relative importance of the two contributions during this event and examine the applicability of Clausius-Clapeyron scaling. This work provides a physical interpretation of the mechanisms associated with Southern UK's wettest event, which is complementary to other studies based on statistical approaches (Schaller et al., 2016, Yiou et al., 2017). The analysis is carried out using the ERA-Interim reanalysis. This is motivated by the horizontal resolution of this dataset. It is then applied to present-day simulations and future projections of CMIP5 models on selected extreme precipitation events in southern UK that are comparable to January 2014 in terms of atmospheric circulation.References:Schaller, N. et al. Human influence on climate in the 2014 southern England winter floods and their impacts, Nature Clim. Change, 2016, 6, 627-634 Yiou, P., et al. A statistical framework for conditional extreme event attribution Advances in Statistical Climatology, Meteorology and Oceanography, 2017, 3, 17-31
Peer Review Documents Related to the Evaluation of ...
BMDS is one of the Agency's premier tools for estimating risk assessments, therefore the validity and reliability of its statistical models are of paramount importance. This page provides links to peer review and expert summaries of the BMDS application and its models as they were developed and eventually released documenting the rigorous review process taken to provide the best science tools available for statistical modeling. This page provides links to peer reviews and expert summaries of the BMDS applications and its models as they were developed and eventually released.
Effects of quantum coherence on work statistics
NASA Astrophysics Data System (ADS)
Xu, Bao-Ming; Zou, Jian; Guo, Li-Sha; Kong, Xiang-Mu
2018-05-01
In the conventional two-point measurement scheme of quantum thermodynamics, quantum coherence is destroyed by the first measurement. But as we know the coherence really plays an important role in the quantum thermodynamics process, and how to describe the work statistics for a quantum coherent process is still an open question. In this paper, we use the full counting statistics method to investigate the effects of quantum coherence on work statistics. First, we give a general discussion and show that for a quantum coherent process, work statistics is very different from that of the two-point measurement scheme, specifically the average work is increased or decreased and the work fluctuation can be decreased by quantum coherence, which strongly depends on the relative phase, the energy level structure, and the external protocol. Then, we concretely consider a quenched one-dimensional transverse Ising model and show that quantum coherence has a more significant influence on work statistics in the ferromagnetism regime compared with that in the paramagnetism regime, so that due to the presence of quantum coherence the work statistics can exhibit the critical phenomenon even at high temperature.
Common pitfalls in statistical analysis: Clinical versus statistical significance
Ranganathan, Priya; Pramesh, C. S.; Buyse, Marc
2015-01-01
In clinical research, study results, which are statistically significant are often interpreted as being clinically important. While statistical significance indicates the reliability of the study results, clinical significance reflects its impact on clinical practice. The third article in this series exploring pitfalls in statistical analysis clarifies the importance of differentiating between statistical significance and clinical significance. PMID:26229754
Caregivers' health literacy and their young children's oral-health-related expenditures.
Vann, W F; Divaris, K; Gizlice, Z; Baker, A D; Lee, J Y
2013-07-01
Caregivers' health literacy has emerged as an important determinant of young children's health care and outcomes. We examined the hypothesis that caregivers' health literacy influences children's oral-health-care-related expenditures. This was a prospective cohort study of 1,132 child/caregiver dyads (children's mean age = 19 months), participating in the Carolina Oral Health Literacy Project. Health literacy was measured by the REALD-30 (word recognition based) and NVS (comprehension based) instruments. Follow-up data included child Medicaid claims for CY2008-10. We quantified expenditures using annualized 2010 fee-adjusted Medicaid-paid dollars for oral-health-related visits involving preventive, restorative, and emergency care. We used descriptive, bivariate, and multivariate statistical methods based on generalized gamma models. Mean oral-health-related annual expenditures totaled $203: preventive--$81, restorative--$99, and emergency care--$22. Among children who received services, mean expenditures were: emergency hospital-based--$1282, preventive--$106, and restorative care--$343. Caregivers' low literacy in the oral health context was associated with a statistically non-significant increase in total expenditures (average annual difference = $40; 95% confidence interval, -32, 111). Nevertheless, with both instruments, emergency dental care expenditures were consistently elevated among children of low-literacy caregivers. These findings provide initial support for health literacy as an important determinant of the meaningful use and cost of oral health care.
Spatial variability effects on precision and power of forage yield estimation
USDA-ARS?s Scientific Manuscript database
Spatial analyses of yield trials are important, as they adjust cultivar means for spatial variation and improve the statistical precision of yield estimation. While the relative efficiency of spatial analysis has been frequently reported in several yield trials, its application on long-term forage y...
The Netherlands in Fifty Maps: An Annotated Atlas.
ERIC Educational Resources Information Center
Tamsma, R.
1988-01-01
Statistical data relating to The Netherlands is presented in 50 maps that were originally published in the "Journal of Economic and Social Geography" between 1977 and 1986. Lengthy annotations accompanying each map provide important background information on the issues depicted. These elucidations were written with the foreign reader in…
Real lasers and other deformed objects
NASA Technical Reports Server (NTRS)
Solomon, Allan I.
1995-01-01
In this talk we re-examine three important properties of quantum laser systems: (1) photon counting statistics; (2) squeezing; and (3) signal-to-quantum noise ratio. None of these phenomena depends on the choice of hamiltonian; indeed, we analyze them initially without restriction to any specific form of the commutation relations.
Indicators are commonly used for evaluating relative sustainability for competing products and processes. When a set of indicators is chosen for a particular system of study, it is important to ensure that they are variable independently of each other. Often the number of indicat...
Do climate extreme events foster violent civil conflicts? A coincidence analysis
NASA Astrophysics Data System (ADS)
Schleussner, Carl-Friedrich; Donges, Jonathan F.; Donner, Reik V.
2014-05-01
Civil conflicts promoted by adverse environmental conditions represent one of the most important potential feedbacks in the global socio-environmental nexus. While the role of climate extremes as a triggering factor is often discussed, no consensus is yet reached about the cause-and-effect relation in the observed data record. Here we present results of a rigorous statistical coincidence analysis based on the Munich Re Inc. extreme events database and the Uppsala conflict data program. We report evidence for statistically significant synchronicity between climate extremes with high economic impact and violent conflicts for various regions, although no coherent global signal emerges from our analysis. Our results indicate the importance of regional vulnerability and might aid to identify hot-spot regions for potential climate-triggered violent social conflicts.
NASA Astrophysics Data System (ADS)
Williams, Arnold C.; Pachowicz, Peter W.
2004-09-01
Current mine detection research indicates that no single sensor or single look from a sensor will detect mines/minefields in a real-time manner at a performance level suitable for a forward maneuver unit. Hence, the integrated development of detectors and fusion algorithms are of primary importance. A problem in this development process has been the evaluation of these algorithms with relatively small data sets, leading to anecdotal and frequently over trained results. These anecdotal results are often unreliable and conflicting among various sensors and algorithms. Consequently, the physical phenomena that ought to be exploited and the performance benefits of this exploitation are often ambiguous. The Army RDECOM CERDEC Night Vision Laboratory and Electron Sensors Directorate has collected large amounts of multisensor data such that statistically significant evaluations of detection and fusion algorithms can be obtained. Even with these large data sets care must be taken in algorithm design and data processing to achieve statistically significant performance results for combined detectors and fusion algorithms. This paper discusses statistically significant detection and combined multilook fusion results for the Ellipse Detector (ED) and the Piecewise Level Fusion Algorithm (PLFA). These statistically significant performance results are characterized by ROC curves that have been obtained through processing this multilook data for the high resolution SAR data of the Veridian X-Band radar. We discuss the implications of these results on mine detection and the importance of statistical significance, sample size, ground truth, and algorithm design in performance evaluation.
Reif, David M.; Israel, Mark A.; Moore, Jason H.
2007-01-01
The biological interpretation of gene expression microarray results is a daunting challenge. For complex diseases such as cancer, wherein the body of published research is extensive, the incorporation of expert knowledge provides a useful analytical framework. We have previously developed the Exploratory Visual Analysis (EVA) software for exploring data analysis results in the context of annotation information about each gene, as well as biologically relevant groups of genes. We present EVA as a flexible combination of statistics and biological annotation that provides a straightforward visual interface for the interpretation of microarray analyses of gene expression in the most commonly occuring class of brain tumors, glioma. We demonstrate the utility of EVA for the biological interpretation of statistical results by analyzing publicly available gene expression profiles of two important glial tumors. The results of a statistical comparison between 21 malignant, high-grade glioblastoma multiforme (GBM) tumors and 19 indolent, low-grade pilocytic astrocytomas were analyzed using EVA. By using EVA to examine the results of a relatively simple statistical analysis, we were able to identify tumor class-specific gene expression patterns having both statistical and biological significance. Our interactive analysis highlighted the potential importance of genes involved in cell cycle progression, proliferation, signaling, adhesion, migration, motility, and structure, as well as candidate gene loci on a region of Chromosome 7 that has been implicated in glioma. Because EVA does not require statistical or computational expertise and has the flexibility to accommodate any type of statistical analysis, we anticipate EVA will prove a useful addition to the repertoire of computational methods used for microarray data analysis. EVA is available at no charge to academic users and can be found at http://www.epistasis.org. PMID:19390666
NASA Astrophysics Data System (ADS)
Hacker, Joshua; Vandenberghe, Francois; Jung, Byoung-Jo; Snyder, Chris
2017-04-01
Effective assimilation of cloud-affected radiance observations from space-borne imagers, with the aim of improving cloud analysis and forecasting, has proven to be difficult. Large observation biases, nonlinear observation operators, and non-Gaussian innovation statistics present many challenges. Ensemble-variational data assimilation (EnVar) systems offer the benefits of flow-dependent background error statistics from an ensemble, and the ability of variational minimization to handle nonlinearity. The specific benefits of ensemble statistics, relative to static background errors more commonly used in variational systems, have not been quantified for the problem of assimilating cloudy radiances. A simple experiment framework is constructed with a regional NWP model and operational variational data assimilation system, to provide the basis understanding the importance of ensemble statistics in cloudy radiance assimilation. Restricting the observations to those corresponding to clouds in the background forecast leads to innovations that are more Gaussian. The number of large innovations is reduced compared to the more general case of all observations, but not eliminated. The Huber norm is investigated to handle the fat tails of the distributions, and allow more observations to be assimilated without the need for strict background checks that eliminate them. Comparing assimilation using only ensemble background error statistics with assimilation using only static background error statistics elucidates the importance of the ensemble statistics. Although the cost functions in both experiments converge to similar values after sufficient outer-loop iterations, the resulting cloud water, ice, and snow content are greater in the ensemble-based analysis. The subsequent forecasts from the ensemble-based analysis also retain more condensed water species, indicating that the local environment is more supportive of clouds. In this presentation we provide details that explain the apparent benefit from using ensembles for cloudy radiance assimilation in an EnVar context.
Perception of orthodontic treatment need in children and adolescents.
Spalj, Stjepan; Slaj, Martina; Varga, Suzana; Strujic, Mihovil; Slaj, Mladen
2010-08-01
Patients' and parents' perception of malocclusion are important in determining orthodontic treatment demand, motivation, and cooperation. The aim of this study was to investigate differences in perception of treatment need in currently orthodontically treated, previously treated, and untreated subjects. The sample comprised 3196 children and adolescents (1593 males and 1603 females) aged 8-19 years (mean age 13.0 +/- 3.6 years) from 24 randomly selected public schools in Zagreb, Croatia. Objective treatment need was assessed clinically using the Dental Aesthetic Index (DAI). Subjective treatment need was estimated separately by an orthodontic resident, the child/adolescent and his/her parent using the Standardized Continuum of Aesthetic Need (SCAN) procedure. The children/adolescents completed a questionnaire that had five questions with five-point Likert-type scale answers concerning satisfaction with dental appearance, importance of teeth for facial appearance, and malocclusion-related quality of life. Spearman correlation and logistic regression were used for statistical analysis. Associations between objective and subjective orthodontic treatment need were weak but statistically significant (Rho from 0.20 to 0.50; P < 0.05). Malocclusion-related quality of life was poorly associated with treatment need. Satisfaction with tooth appearance showed the most frequent statistically significant correlation (Rho from -0.14 to -0.35; P < 0.05), while importance of aligned teeth for facial appearance and social contacts had the weakest correlation with treatment need. Perception of treatment need was greater in previously treated subjects. Parents' perception had a low predictive value. The findings of this study show that malocclusion has more impact on emotional well-being than on function or social contacts.
A phylogenetic transform enhances analysis of compositional microbiota data
Silverman, Justin D; Washburne, Alex D; Mukherjee, Sayan; David, Lawrence A
2017-01-01
Surveys of microbial communities (microbiota), typically measured as relative abundance of species, have illustrated the importance of these communities in human health and disease. Yet, statistical artifacts commonly plague the analysis of relative abundance data. Here, we introduce the PhILR transform, which incorporates microbial evolutionary models with the isometric log-ratio transform to allow off-the-shelf statistical tools to be safely applied to microbiota surveys. We demonstrate that analyses of community-level structure can be applied to PhILR transformed data with performance on benchmarks rivaling or surpassing standard tools. Additionally, by decomposing distance in the PhILR transformed space, we identified neighboring clades that may have adapted to distinct human body sites. Decomposing variance revealed that covariation of bacterial clades within human body sites increases with phylogenetic relatedness. Together, these findings illustrate how the PhILR transform combines statistical and phylogenetic models to overcome compositional data challenges and enable evolutionary insights relevant to microbial communities. DOI: http://dx.doi.org/10.7554/eLife.21887.001 PMID:28198697
NASA Astrophysics Data System (ADS)
Röpke, G.
2018-01-01
One of the fundamental problems in physics that are not yet rigorously solved is the statistical mechanics of nonequilibrium processes. An important contribution to describing irreversible behavior starting from reversible Hamiltonian dynamics was given by D. N. Zubarev, who invented the method of the nonequilibrium statistical operator. We discuss this approach, in particular, the extended von Neumann equation, and as an example consider the electrical conductivity of a system of charged particles. We consider the selection of the set of relevant observables. We show the relation between kinetic theory and linear response theory. Using thermodynamic Green's functions, we present a systematic treatment of correlation functions, but the convergence needs investigation. We compare different expressions for the conductivity and list open questions.
Lambert, Nathaniel D.; Pankratz, V. Shane; Larrabee, Beth R.; Ogee-Nwankwo, Adaeze; Chen, Min-hsin; Icenogle, Joseph P.
2014-01-01
Rubella remains a social and economic burden due to the high incidence of congenital rubella syndrome (CRS) in some countries. For this reason, an accurate and efficient high-throughput measure of antibody response to vaccination is an important tool. In order to measure rubella-specific neutralizing antibodies in a large cohort of vaccinated individuals, a high-throughput immunocolorimetric system was developed. Statistical interpolation models were applied to the resulting titers to refine quantitative estimates of neutralizing antibody titers relative to the assayed neutralizing antibody dilutions. This assay, including the statistical methods developed, can be used to assess the neutralizing humoral immune response to rubella virus and may be adaptable for assessing the response to other viral vaccines and infectious agents. PMID:24391140
A genome-wide methylation study on obesity: differential variability and differential methylation.
Xu, Xiaojing; Su, Shaoyong; Barnes, Vernon A; De Miguel, Carmen; Pollock, Jennifer; Ownby, Dennis; Shi, Hidong; Zhu, Haidong; Snieder, Harold; Wang, Xiaoling
2013-05-01
Besides differential methylation, DNA methylation variation has recently been proposed and demonstrated to be a potential contributing factor to cancer risk. Here we aim to examine whether differential variability in methylation is also an important feature of obesity, a typical non-malignant common complex disease. We analyzed genome-wide methylation profiles of over 470,000 CpGs in peripheral blood samples from 48 obese and 48 lean African-American youth aged 14-20 y old. A substantial number of differentially variable CpG sites (DVCs), using statistics based on variances, as well as a substantial number of differentially methylated CpG sites (DMCs), using statistics based on means, were identified. Similar to the findings in cancers, DVCs generally exhibited an outlier structure and were more variable in cases than in controls. By randomly splitting the current sample into a discovery and validation set, we observed that both the DVCs and DMCs identified from the first set could independently predict obesity status in the second set. Furthermore, both the genes harboring DMCs and the genes harboring DVCs showed significant enrichment of genes identified by genome-wide association studies on obesity and related diseases, such as hypertension, dyslipidemia, type 2 diabetes and certain types of cancers, supporting their roles in the etiology and pathogenesis of obesity. We generalized the recent finding on methylation variability in cancer research to obesity and demonstrated that differential variability is also an important feature of obesity-related methylation changes. Future studies on the epigenetics of obesity will benefit from both statistics based on means and statistics based on variances.
Rushton, Paul R P; Grevitt, Michael P
2013-04-20
Review and statistical analysis of studies evaluating the effect of surgery on the health-related quality of life of adolescents with adolescent idiopathic scoliosis, using Scoliosis Research Society (SRS) outcomes. Apply published minimum clinical important differences (MCID) values for the SRS22r questionnaire to the literature to identify what areas of health-related quality of life are consistently affected by surgery and whether changes are clinically meaningful. The interpretation of published studies using the SRS outcomes has been limited by the lack of MCID values for the questionnaire domains. The recent publication of these data allows the clinical importance of any changes in these studies to be examined for the first time. A literature search was undertaken to locate suitable studies that were then analyzed. Statistically significant differences from baseline to 2 years postoperatively were ascertained by narratively reporting the analyses within included studies. When possible, clinically significant changes were assessed using 95% confidence intervals for the change in mean domain score. If the lower bound of the confidence intervals for the change exceeded the MCID for that domain, the change was considered clinically significant. The numbers of cohorts available for the different analyses varied (5-16). Eighty-one percent and 94% of included cohorts experienced statistically significant improvements in pain and self-image domains. In terms of clinical significance, it was only self-image that regularly improved by more than MCID, doing so in 4 of 5 included cohorts (80%) compared with 1 of 12 cohorts (8%) for pain. No clinically relevant changes occurred in mental health or activity domains. Evidence suggests that surgery can lead to clinically important improvement in patient self-image. Surgeons and patients should be aware of the limited evidence for improvements in domains other than self-image after surgery. Surgical decision-making will also be influenced by the natural history of adolescent idiopathic scoliosis.
Willits, Jon A.; Seidenberg, Mark S.; Saffran, Jenny R.
2014-01-01
What makes some words easy for infants to recognize, and other words difficult? We addressed this issue in the context of prior results suggesting that infants have difficulty recognizing verbs relative to nouns. In this work, we highlight the role played by the distributional contexts in which nouns and verbs occur. Distributional statistics predict that English nouns should generally be easier to recognize than verbs in fluent speech. However, there are situations in which distributional statistics provide similar support for verbs. The statistics for verbs that occur with the English morpheme –ing, for example, should facilitate verb recognition. In two experiments with 7.5- and 9.5-month-old infants, we tested the importance of distributional statistics for word recognition by varying the frequency of the contextual frames in which verbs occur. The results support the conclusion that distributional statistics are utilized by infant language learners and contribute to noun–verb differences in word recognition. PMID:24908342
Statistical Analysis of Large-Scale Structure of Universe
NASA Astrophysics Data System (ADS)
Tugay, A. V.
While galaxy cluster catalogs were compiled many decades ago, other structural elements of cosmic web are detected at definite level only in the newest works. For example, extragalactic filaments were described by velocity field and SDSS galaxy distribution during the last years. Large-scale structure of the Universe could be also mapped in the future using ATHENA observations in X-rays and SKA in radio band. Until detailed observations are not available for the most volume of Universe, some integral statistical parameters can be used for its description. Such methods as galaxy correlation function, power spectrum, statistical moments and peak statistics are commonly used with this aim. The parameters of power spectrum and other statistics are important for constraining the models of dark matter, dark energy, inflation and brane cosmology. In the present work we describe the growth of large-scale density fluctuations in one- and three-dimensional case with Fourier harmonics of hydrodynamical parameters. In result we get power-law relation for the matter power spectrum.
Two SPSS programs for interpreting multiple regression results.
Lorenzo-Seva, Urbano; Ferrando, Pere J; Chico, Eliseo
2010-02-01
When multiple regression is used in explanation-oriented designs, it is very important to determine both the usefulness of the predictor variables and their relative importance. Standardized regression coefficients are routinely provided by commercial programs. However, they generally function rather poorly as indicators of relative importance, especially in the presence of substantially correlated predictors. We provide two user-friendly SPSS programs that implement currently recommended techniques and recent developments for assessing the relevance of the predictors. The programs also allow the user to take into account the effects of measurement error. The first program, MIMR-Corr.sps, uses a correlation matrix as input, whereas the second program, MIMR-Raw.sps, uses the raw data and computes bootstrap confidence intervals of different statistics. The SPSS syntax, a short manual, and data files related to this article are available as supplemental materials from http://brm.psychonomic-journals.org/content/supplemental.
Basic statistics (the fundamental concepts).
Lim, Eric
2014-12-01
An appreciation and understanding of statistics is import to all practising clinicians, not simply researchers. This is because mathematics is the fundamental basis to which we base clinical decisions, usually with reference to the benefit in relation to risk. Unless a clinician has a basic understanding of statistics, he or she will never be in a position to question healthcare management decisions that have been handed down from generation to generation, will not be able to conduct research effectively nor evaluate the validity of published evidence (usually making an assumption that most published work is either all good or all bad). This article provides a brief introduction to basic statistical methods and illustrates its use in common clinical scenarios. In addition, pitfalls of incorrect usage have been highlighted. However, it is not meant to be a substitute for formal training or consultation with a qualified and experienced medical statistician prior to starting any research project.
Different Manhattan project: automatic statistical model generation
NASA Astrophysics Data System (ADS)
Yap, Chee Keng; Biermann, Henning; Hertzmann, Aaron; Li, Chen; Meyer, Jon; Pao, Hsing-Kuo; Paxia, Salvatore
2002-03-01
We address the automatic generation of large geometric models. This is important in visualization for several reasons. First, many applications need access to large but interesting data models. Second, we often need such data sets with particular characteristics (e.g., urban models, park and recreation landscape). Thus we need the ability to generate models with different parameters. We propose a new approach for generating such models. It is based on a top-down propagation of statistical parameters. We illustrate the method in the generation of a statistical model of Manhattan. But the method is generally applicable in the generation of models of large geographical regions. Our work is related to the literature on generating complex natural scenes (smoke, forests, etc) based on procedural descriptions. The difference in our approach stems from three characteristics: modeling with statistical parameters, integration of ground truth (actual map data), and a library-based approach for texture mapping.
Andersen, Lars Peter; Hogh, Annie; Biering, Karin; Gadegaard, Charlotte Ann
2018-01-01
Threats and violence at work are major concerns for employees in many human service sectors. The prevention of work-related violence is a major challenge for employees and management. The purpose of this study was to identify prospective associations between psycho-social work environment and work-related threats and violence in four high risk human service sectors. Questionnaire data was collected from 3011 employees working at psychiatric wards, in the elder sector, in the Prison and Probation Service and at Special Schools. Associations between psycho-social work environment and work-related violence and threats were then studied using a one-year follow-up design and multilevel logistic regression analyses. The analyses showed that quantitative demands, high emotional demands, low level of influence over own work-situation, low predictability, low rewards at work, low role clarity, many role conflicts, many work-family conflicts and low organizational justice had statistically significant associations with high levels of work-related threats. Furthermore, high emotional demands, low predictability, low role clarity, many role conflicts, many work-family conflicts, low supervisor quality and low support from nearest supervisor had statistically significant associations with high levels of work-related violence. Finally, across the four sectors both similar and different associations between psycho-social work environment and work-related violence and threats were found. The results of the study underline the importance of including the psycho-social work environment as a supplement to existing violence prevention methods and interventions aimed at reducing work-related violence and threats.
NASA Astrophysics Data System (ADS)
Tkačik, Gašper
2016-07-01
The article by O. Martin and colleagues provides a much needed systematic review of a body of work that relates the topological structure of genetic regulatory networks to evolutionary selection for function. This connection is very important. Using the current wealth of genomic data, statistical features of regulatory networks (e.g., degree distributions, motif composition, etc.) can be quantified rather easily; it is, however, often unclear how to interpret the results. On a graph theoretic level the statistical significance of the results can be evaluated by comparing observed graphs to ;randomized; ones (bravely ignoring the issue of how precisely to randomize!) and comparing the frequency of appearance of a particular network structure relative to a randomized null expectation. While this is a convenient operational test for statistical significance, its biological meaning is questionable. In contrast, an in-silico genotype-to-phenotype model makes explicit the assumptions about the network function, and thus clearly defines the expected network structures that can be compared to the case of no selection for function and, ultimately, to data.
Issues related to the detection of boundaries
M.-J. Fortin; Olson; R.J.; S. Ferson; L. Iverson; C. Hunsaker; G. Edwards; D. Levine; K. Butera; V. Klemas; V. Klemas
2000-01-01
Ecotones are inherent features of landscapes, transitional zones, and play more than one functional role in ecosystem dynamics. The delineation of ecotones and environmental boundaries is therefore an important step in land-use management planning. The delineation of ecotones depends on the phenomenon of interest and the statistical methods used as well as the...
ERIC Educational Resources Information Center
Women's Bureau (DOL), Washington, DC.
Statistical information pertaining to one of the most important changes in the American economy in this century--the increase in the number of women who work outside the home--is presented as an introduction to the broader range of topics which will be considered by the Advisory Committee on the Economic Role of Women. Job-related aspects of…
Conjoint Analysis: A Study of the Effects of Using Person Variables.
ERIC Educational Resources Information Center
Fraas, John W.; Newman, Isadore
Three statistical techniques--conjoint analysis, a multiple linear regression model, and a multiple linear regression model with a surrogate person variable--were used to estimate the relative importance of five university attributes for students in the process of selecting a college. The five attributes include: availability and variety of…
Using Candy Samples to Learn about Sampling Techniques and Statistical Data Evaluation
ERIC Educational Resources Information Center
Canaes, Larissa S.; Brancalion, Marcel L.; Rossi, Adriana V.; Rath, Susanne
2008-01-01
A classroom exercise for undergraduate and beginning graduate students that takes about one class period is proposed and discussed. It is an easy, interesting exercise that demonstrates important aspects of sampling techniques (sample amount, particle size, and the representativeness of the sample in relation to the bulk material). The exercise…
Ready-to-Use Simulation: Demystifying Statistical Process Control
ERIC Educational Resources Information Center
Sumukadas, Narendar; Fairfield-Sonn, James W.; Morgan, Sandra
2005-01-01
Business students are typically introduced to the concept of process management in their introductory course on operations management. A very important learning outcome here is an appreciation that the management of processes is a key to the management of quality. Some of the related concepts are qualitative, such as strategic and behavioral…
ERIC Educational Resources Information Center
Thomas, Michael S. C.; Forrester, Neil A.; Ronald, Angelica
2016-01-01
In the multidisciplinary field of developmental cognitive neuroscience, statistical associations between levels of description play an increasingly important role. One example of such associations is the observation of correlations between relatively common gene variants and individual differences in behavior. It is perhaps surprising that such…
Learning to Look: Probabilistic Variation and Noise Guide Infants' Eye Movements
ERIC Educational Resources Information Center
Tummeltshammer, Kristen Swan; Kirkham, Natasha Z.
2013-01-01
Young infants have demonstrated a remarkable sensitivity to probabilistic relations among visual features (Fiser & Aslin, 2002; Kirkham et al., 2002). Previous research has raised important questions regarding the usefulness of statistical learning in an environment filled with variability and noise, such as an infant's natural world. In…
Re-Entry, Recruitment, and Retention: A Community Relations Model for Sacramento City College.
ERIC Educational Resources Information Center
White, Maureen E.
Enrollment statistics and projections confirm the importance of focusing community college student recruitment and retention efforts on re-entry students. Re-entry students are a distinct and growing population whose educational requirements often differ from those of younger, traditional students. The literature on adult learners indicates that:…
Downscaling Indicators of Forest Habitat Structure from National Assessments
Kurt H. Riitters
2005-01-01
Downscaling is an important problem because consistent large-area assessments of forest habitat structure, while feasible, are only feasible when using relatively coarse data and indicators. Techniques are needed to enable more detailed and local interpretations of the national statistics. Using the results of national assessments from land-cover maps, this paper...
SOME APPLICATIONS OF SEISMIC SOURCE MECHANISM STUDIES TO ASSESSING UNDERGROUND HAZARD.
McGarr, A.; ,
1984-01-01
Various measures of the seismic source mechanism of mine tremors, such as magnitude, moment, stress drop, apparent stress, and seismic efficiency, can be related directly to several aspects of the problem of determining the underground hazard arising from strong ground motion of large seismic events. First, the relation between the sum of seismic moments of tremors and the volume of stope closure caused by mining during a given period can be used in conjunction with magnitude-frequency statistics and an empirical relation between moment and magnitude to estimate the maximum possible sized tremor for a given mining situation. Second, it is shown that the 'energy release rate,' a commonly-used parameter for predicting underground seismic hazard, may be misleading in that the importance of overburden stress, or depth, is overstated. Third, results involving the relation between peak velocity and magnitude, magnitude-frequency statistics, and the maximum possible magnitude are applied to the problem of estimating the frequency at which design limits of certain underground support equipment are likely to be exceeded.
Johnson, Eric D; Tubau, Elisabet
2017-06-01
Presenting natural frequencies facilitates Bayesian inferences relative to using percentages. Nevertheless, many people, including highly educated and skilled reasoners, still fail to provide Bayesian responses to these computationally simple problems. We show that the complexity of relational reasoning (e.g., the structural mapping between the presented and requested relations) can help explain the remaining difficulties. With a non-Bayesian inference that required identical arithmetic but afforded a more direct structural mapping, performance was universally high. Furthermore, reducing the relational demands of the task through questions that directed reasoners to use the presented statistics, as compared with questions that prompted the representation of a second, similar sample, also significantly improved reasoning. Distinct error patterns were also observed between these presented- and similar-sample scenarios, which suggested differences in relational-reasoning strategies. On the other hand, while higher numeracy was associated with better Bayesian reasoning, higher-numerate reasoners were not immune to the relational complexity of the task. Together, these findings validate the relational-reasoning view of Bayesian problem solving and highlight the importance of considering not only the presented task structure, but also the complexity of the structural alignment between the presented and requested relations.
Analysis of Loss-of-Offsite-Power Events 1997-2015
DOE Office of Scientific and Technical Information (OSTI.GOV)
Johnson, Nancy Ellen; Schroeder, John Alton
2016-07-01
Loss of offsite power (LOOP) can have a major negative impact on a power plant’s ability to achieve and maintain safe shutdown conditions. LOOP event frequencies and times required for subsequent restoration of offsite power are important inputs to plant probabilistic risk assessments. This report presents a statistical and engineering analysis of LOOP frequencies and durations at U.S. commercial nuclear power plants. The data used in this study are based on the operating experience during calendar years 1997 through 2015. LOOP events during critical operation that do not result in a reactor trip, are not included. Frequencies and durations weremore » determined for four event categories: plant-centered, switchyard-centered, grid-related, and weather-related. Emergency diesel generator reliability is also considered (failure to start, failure to load and run, and failure to run more than 1 hour). There is an adverse trend in LOOP durations. The previously reported adverse trend in LOOP frequency was not statistically significant for 2006-2015. Grid-related LOOPs happen predominantly in the summer. Switchyard-centered LOOPs happen predominantly in winter and spring. Plant-centered and weather-related LOOPs do not show statistically significant seasonality. The engineering analysis of LOOP data shows that human errors have been much less frequent since 1997 than in the 1986 -1996 time period.« less
Posada, David
2006-01-01
ModelTest server is a web-based application for the selection of models of nucleotide substitution using the program ModelTest. The server takes as input a text file with likelihood scores for the set of candidate models. Models can be selected with hierarchical likelihood ratio tests, or with the Akaike or Bayesian information criteria. The output includes several statistics for the assessment of model selection uncertainty, for model averaging or to estimate the relative importance of model parameters. The server can be accessed at . PMID:16845102
New Probe of Departures from General Relativity Using Minkowski Functionals.
Fang, Wenjuan; Li, Baojiu; Zhao, Gong-Bo
2017-05-05
The morphological properties of the large scale structure of the Universe can be fully described by four Minkowski functionals (MFs), which provide important complementary information to other statistical observables such as the widely used 2-point statistics in configuration and Fourier spaces. In this work, for the first time, we present the differences in the morphology of the large scale structure caused by modifications to general relativity (to address the cosmic acceleration problem), by measuring the MFs from N-body simulations of modified gravity and general relativity. We find strong statistical power when using the MFs to constrain modified theories of gravity: with a galaxy survey that has survey volume ∼0.125(h^{-1} Gpc)^{3} and galaxy number density ∼1/(h^{-1} Mpc)^{3}, the two normal-branch Dvali-Gabadadze-Porrati models and the F5 f(R) model that we simulated can be discriminated from the ΛCDM model at a significance level ≳5σ with an individual MF measurement. Therefore, the MF of the large scale structure is potentially a powerful probe of gravity, and its application to real data deserves active exploration.
On P values and effect modification.
Mayer, Martin
2017-12-01
A crucial element of evidence-based healthcare is the sound understanding and use of statistics. As part of instilling sound statistical knowledge and practice, it seems useful to highlight instances of unsound statistical reasoning or practice, not merely in captious or vitriolic spirit, but rather, to use such error as a springboard for edification by giving tangibility to the concepts at hand and highlighting the importance of avoiding such error. This article aims to provide an instructive overview of two key statistical concepts: effect modification and P values. A recent article published in the Journal of the American College of Cardiology on side effects related to statin therapy offers a notable example of errors in understanding effect modification and P values, and although not so critical as to entirely invalidate the article, the errors still demand considerable scrutiny and correction. In doing so, this article serves as an instructive overview of the statistical concepts of effect modification and P values. Judicious handling of statistics is imperative to avoid muddying their utility. This article contributes to the body of literature aiming to improve the use of statistics, which in turn will help facilitate evidence appraisal, synthesis, translation, and application.
Alhdiri, Maryam Ahmed; Samat, Nor Azah; Mohamed, Zulkifley
2017-03-01
Cancer is the most rapidly spreading disease in the world, especially in developing countries, including Libya. Cancer represents a significant burden on patients, families, and their societies. This disease can be controlled if detected early. Therefore, disease mapping has recently become an important method in the fields of public health research and disease epidemiology. The correct choice of statistical model is a very important step to producing a good map of a disease. Libya was selected to perform this work and to examine its geographical variation in the incidence of lung cancer. The objective of this paper is to estimate the relative risk for lung cancer. Four statistical models to estimate the relative risk for lung cancer and population censuses of the study area for the time period 2006 to 2011 were used in this work. They are initially known as Standardized Morbidity Ratio, which is the most popular statistic, which used in the field of disease mapping, Poisson-gamma model, which is one of the earliest applications of Bayesian methodology, Besag, York and Mollie (BYM) model and Mixture model. As an initial step, this study begins by providing a review of all proposed models, which we then apply to lung cancer data in Libya. Maps, tables and graph, goodness-of-fit (GOF) were used to compare and present the preliminary results. This GOF is common in statistical modelling to compare fitted models. The main general results presented in this study show that the Poisson-gamma model, BYM model, and Mixture model can overcome the problem of the first model (SMR) when there is no observed lung cancer case in certain districts. Results show that the Mixture model is most robust and provides better relative risk estimates across a range of models. Creative Commons Attribution License
Alhdiri, Maryam Ahmed; Samat, Nor Azah; Mohamed, Zulkifley
2017-01-01
Cancer is the most rapidly spreading disease in the world, especially in developing countries, including Libya. Cancer represents a significant burden on patients, families, and their societies. This disease can be controlled if detected early. Therefore, disease mapping has recently become an important method in the fields of public health research and disease epidemiology. The correct choice of statistical model is a very important step to producing a good map of a disease. Libya was selected to perform this work and to examine its geographical variation in the incidence of lung cancer. The objective of this paper is to estimate the relative risk for lung cancer. Four statistical models to estimate the relative risk for lung cancer and population censuses of the study area for the time period 2006 to 2011 were used in this work. They are initially known as Standardized Morbidity Ratio, which is the most popular statistic, which used in the field of disease mapping, Poisson-gamma model, which is one of the earliest applications of Bayesian methodology, Besag, York and Mollie (BYM) model and Mixture model. As an initial step, this study begins by providing a review of all proposed models, which we then apply to lung cancer data in Libya. Maps, tables and graph, goodness-of-fit (GOF) were used to compare and present the preliminary results. This GOF is common in statistical modelling to compare fitted models. The main general results presented in this study show that the Poisson-gamma model, BYM model, and Mixture model can overcome the problem of the first model (SMR) when there is no observed lung cancer case in certain districts. Results show that the Mixture model is most robust and provides better relative risk estimates across a range of models. PMID:28440974
Topographic relationships for design rainfalls over Australia
NASA Astrophysics Data System (ADS)
Johnson, F.; Hutchinson, M. F.; The, C.; Beesley, C.; Green, J.
2016-02-01
Design rainfall statistics are the primary inputs used to assess flood risk across river catchments. These statistics normally take the form of Intensity-Duration-Frequency (IDF) curves that are derived from extreme value probability distributions fitted to observed daily, and sub-daily, rainfall data. The design rainfall relationships are often required for catchments where there are limited rainfall records, particularly catchments in remote areas with high topographic relief and hence some form of interpolation is required to provide estimates in these areas. This paper assesses the topographic dependence of rainfall extremes by using elevation-dependent thin plate smoothing splines to interpolate the mean annual maximum rainfall, for periods from one to seven days, across Australia. The analyses confirm the important impact of topography in explaining the spatial patterns of these extreme rainfall statistics. Continent-wide residual and cross validation statistics are used to demonstrate the 100-fold impact of elevation in relation to horizontal coordinates in explaining the spatial patterns, consistent with previous rainfall scaling studies and observational evidence. The impact of the complexity of the fitted spline surfaces, as defined by the number of knots, and the impact of applying variance stabilising transformations to the data, were also assessed. It was found that a relatively large number of 3570 knots, suitably chosen from 8619 gauge locations, was required to minimise the summary error statistics. Square root and log data transformations were found to deliver marginally superior continent-wide cross validation statistics, in comparison to applying no data transformation, but detailed assessments of residuals in complex high rainfall regions with high topographic relief showed that no data transformation gave superior performance in these regions. These results are consistent with the understanding that in areas with modest topographic relief, as for most of the Australian continent, extreme rainfall is closely aligned with elevation, but in areas with high topographic relief the impacts of topography on rainfall extremes are more complex. The interpolated extreme rainfall statistics, using no data transformation, have been used by the Australian Bureau of Meteorology to produce new IDF data for the Australian continent. The comprehensive methods presented for the evaluation of gridded design rainfall statistics will be useful for similar studies, in particular the importance of balancing the need for a continentally-optimum solution that maintains sufficient definition at the local scale.
Severson, R.C.; Gough, L.P.
1979-01-01
In order to assess the contribution to plants and soils of certain elements emitted by phosphate processing, we sampled sagebrush, grasses, and A- and C-horizon soils along upwind and downwind transects at Pocatello and Soda Springs, Idaho. Analyses for 70 elements in plants showed that, statistically, the concentration of 7 environmentally important elements, cadmium, chromium, fluorine, selenium, uranium, vanadium, and zinc, were related to emissions from phosphate-processing operations. Two additional elements, lithium and nickel, show probable relationships. The literature on the effects of these elements on plant and animal health is briefly surveyed. Relations between element content in plants and distance from the phosphate-processing operations were stronger at Soda Springs than at Pocatello and, in general, stronger in sagebrush than in the grasses. Analyses for 58 elements in soils showed that, statistically, beryllium, fluorine, iron, lead, lithium, potassium, rubidium, thorium, and zinc were related to emissions only at Pocatello and only in the A horizon. Moreover, six additional elements, copper, mercury, nickel, titanium, uranium, and vanadium, probably are similarly related along the same transect. The approximate amounts of elements added to the soils by the emissions are estimated. In C-horizon soils, no statistically significant relations were observed between element concentrations and distance from the processing sites. At Soda Springs, the nonuniformity of soils at the sampling locations may have obscured the relationship between soil-element content and emissions from phosphate processing.
The Statistics of Urban Scaling and Their Connection to Zipf’s Law
Gomez-Lievano, Andres; Youn, HyeJin; Bettencourt, Luís M. A.
2012-01-01
Urban scaling relations characterizing how diverse properties of cities vary on average with their population size have recently been shown to be a general quantitative property of many urban systems around the world. However, in previous studies the statistics of urban indicators were not analyzed in detail, raising important questions about the full characterization of urban properties and how scaling relations may emerge in these larger contexts. Here, we build a self-consistent statistical framework that characterizes the joint probability distributions of urban indicators and city population sizes across an urban system. To develop this framework empirically we use one of the most granular and stochastic urban indicators available, specifically measuring homicides in cities of Brazil, Colombia and Mexico, three nations with high and fast changing rates of violent crime. We use these data to derive the conditional probability of the number of homicides per year given the population size of a city. To do this we use Bayes’ rule together with the estimated conditional probability of city size given their number of homicides and the distribution of total homicides. We then show that scaling laws emerge as expectation values of these conditional statistics. Knowledge of these distributions implies, in turn, a relationship between scaling and population size distribution exponents that can be used to predict Zipf’s exponent from urban indicator statistics. Our results also suggest how a general statistical theory of urban indicators may be constructed from the stochastic dynamics of social interaction processes in cities. PMID:22815745
NASA Astrophysics Data System (ADS)
Uhlemann, C.; Feix, M.; Codis, S.; Pichon, C.; Bernardeau, F.; L'Huillier, B.; Kim, J.; Hong, S. E.; Laigle, C.; Park, C.; Shin, J.; Pogosyan, D.
2018-02-01
Starting from a very accurate model for density-in-cells statistics of dark matter based on large deviation theory, a bias model for the tracer density in spheres is formulated. It adopts a mean bias relation based on a quadratic bias model to relate the log-densities of dark matter to those of mass-weighted dark haloes in real and redshift space. The validity of the parametrized bias model is established using a parametrization-independent extraction of the bias function. This average bias model is then combined with the dark matter PDF, neglecting any scatter around it: it nevertheless yields an excellent model for densities-in-cells statistics of mass tracers that is parametrized in terms of the underlying dark matter variance and three bias parameters. The procedure is validated on measurements of both the one- and two-point statistics of subhalo densities in the state-of-the-art Horizon Run 4 simulation showing excellent agreement for measured dark matter variance and bias parameters. Finally, it is demonstrated that this formalism allows for a joint estimation of the non-linear dark matter variance and the bias parameters using solely the statistics of subhaloes. Having verified that galaxy counts in hydrodynamical simulations sampled on a scale of 10 Mpc h-1 closely resemble those of subhaloes, this work provides important steps towards making theoretical predictions for density-in-cells statistics applicable to upcoming galaxy surveys like Euclid or WFIRST.
The Content of Statistical Requirements for Authors in Biomedical Research Journals
Liu, Tian-Yi; Cai, Si-Yu; Nie, Xiao-Lu; Lyu, Ya-Qi; Peng, Xiao-Xia; Feng, Guo-Shuang
2016-01-01
Background: Robust statistical designing, sound statistical analysis, and standardized presentation are important to enhance the quality and transparency of biomedical research. This systematic review was conducted to summarize the statistical reporting requirements introduced by biomedical research journals with an impact factor of 10 or above so that researchers are able to give statistical issues’ serious considerations not only at the stage of data analysis but also at the stage of methodological design. Methods: Detailed statistical instructions for authors were downloaded from the homepage of each of the included journals or obtained from the editors directly via email. Then, we described the types and numbers of statistical guidelines introduced by different press groups. Items of statistical reporting guideline as well as particular requirements were summarized in frequency, which were grouped into design, method of analysis, and presentation, respectively. Finally, updated statistical guidelines and particular requirements for improvement were summed up. Results: Totally, 21 of 23 press groups introduced at least one statistical guideline. More than half of press groups can update their statistical instruction for authors gradually relative to issues of new statistical reporting guidelines. In addition, 16 press groups, covering 44 journals, address particular statistical requirements. The most of the particular requirements focused on the performance of statistical analysis and transparency in statistical reporting, including “address issues relevant to research design, including participant flow diagram, eligibility criteria, and sample size estimation,” and “statistical methods and the reasons.” Conclusions: Statistical requirements for authors are becoming increasingly perfected. Statistical requirements for authors remind researchers that they should make sufficient consideration not only in regards to statistical methods during the research design, but also standardized statistical reporting, which would be beneficial in providing stronger evidence and making a greater critical appraisal of evidence more accessible. PMID:27748343
The Content of Statistical Requirements for Authors in Biomedical Research Journals.
Liu, Tian-Yi; Cai, Si-Yu; Nie, Xiao-Lu; Lyu, Ya-Qi; Peng, Xiao-Xia; Feng, Guo-Shuang
2016-10-20
Robust statistical designing, sound statistical analysis, and standardized presentation are important to enhance the quality and transparency of biomedical research. This systematic review was conducted to summarize the statistical reporting requirements introduced by biomedical research journals with an impact factor of 10 or above so that researchers are able to give statistical issues' serious considerations not only at the stage of data analysis but also at the stage of methodological design. Detailed statistical instructions for authors were downloaded from the homepage of each of the included journals or obtained from the editors directly via email. Then, we described the types and numbers of statistical guidelines introduced by different press groups. Items of statistical reporting guideline as well as particular requirements were summarized in frequency, which were grouped into design, method of analysis, and presentation, respectively. Finally, updated statistical guidelines and particular requirements for improvement were summed up. Totally, 21 of 23 press groups introduced at least one statistical guideline. More than half of press groups can update their statistical instruction for authors gradually relative to issues of new statistical reporting guidelines. In addition, 16 press groups, covering 44 journals, address particular statistical requirements. The most of the particular requirements focused on the performance of statistical analysis and transparency in statistical reporting, including "address issues relevant to research design, including participant flow diagram, eligibility criteria, and sample size estimation," and "statistical methods and the reasons." Statistical requirements for authors are becoming increasingly perfected. Statistical requirements for authors remind researchers that they should make sufficient consideration not only in regards to statistical methods during the research design, but also standardized statistical reporting, which would be beneficial in providing stronger evidence and making a greater critical appraisal of evidence more accessible.
The Quantitative Reasoning for College Science (QuaRCS) Assessment in non-Astro 101 Courses II
NASA Astrophysics Data System (ADS)
Kirkman, Thomas W.; Jensen, Ellen
2017-06-01
The Quantitative Reasoning for College Science (QuaRCS) Assessment[1] aims to measure the pre-algebra mathematical skills that are often part of "general education" science courses like Astro 101. In four majors STEM classes, we report comparisons between QuaRCS metrics, ACT math, GPAO, and the course grade. In three of four classes QuaRCS QR score and ACT math were statistically significantly correlated (with r˜.6), however in the fourth course —a senior-level microbiology course— there was no statistically significantly correlation (in fact, r<0). In all courses —even in courses with seemingly little quantitative content— course grade was statistically significantly correlated to GPAO and QR. A QuaRCS metric aiming to report the students belief in the importance of math in science was seen to grow with the course level. Pre/post QuaRCS testing in Physics courses showed fractional sigma gains in QR, self-estimated math fluency and math importance, but not all of those increases were statistically significant. Using a QuaRCS map relating the questions to skill areas, we found graph reading, percentages, and proportional reasoning to be the most misunderstood skills in all four courses.[1] QuaRCS, Follette, et al.,2015, DOI: http://dx.doi.org/10.5038/1936-4660.8.2.2
Ng, M L; Warlow, R S; Chrishanthan, N; Ellis, C; Walls, R
2000-09-01
The aim of this study is to formulate criteria for the definition of allergic rhinitis. Other studies have sought to develop scoring systems to categorize the severity of allergic rhinitis symptoms but it was never used for the formulation of diagnostic criteria. These other scoring systems were arbitrarily chosen and were not derived by any statistical analysis. To date, a study of this kind has not been performed. The hypothesis of this study is that it is possible to formulate criteria for the definition of allergic rhinitis. This is the first study to systematically examine and evaluate the relative importance of symptoms, signs and investigative tests in allergic rhinitis. We sought to statistically rank, from the most to the least important, the multiplicity of symptoms, signs and test results. Forty-seven allergic rhinitis and 23 normal subjects were evaluated with a detailed questionnaire and history, physical examination, serum total immunoglobulin E, skin prick tests and serum enzyme allergosorbent tests (EAST). Statistical ranking of variables indicated rhinitis symptoms (nasal, ocular and oronasal) were the most commonly occurring, followed by a history of allergen provocation, then serum total IgE, positive skin prick tests and positive EAST's to house dust mite, perennial rye and bermuda/couch grass. Throat symptoms ranked even lower whilst EAST's to cat epithelia, plantain and cockroach were the least important. Not all symptoms, signs and tests evaluated proved to be statistically significant when compared to a control group; this included symtoms and signs which had been considered historically to be traditionally associated with allergic rhinitis, e.g. sore throat and bleeding nose. In performing statistical analyses, we were able to rank from most to least important, the multiplicity of symptoms signs and test results. The most important symptoms and signs were identified for the first time, even though some of these were not included in our original selection criteria for defining the disease cohort i.e. sniffing, postnasal drip, oedematous nasal mucosa, impaired sense of smell, mouth breathing, itchy nose and many of the specific provocation factors.
NASA Technical Reports Server (NTRS)
Wilson, R. M.
1982-01-01
Based on 1349 hydrogen alpha flares with X-ray counterparts, an investigation into the relationship between the rise time, decay time, duration, latitude, hydrogen alpha importance, and X-ray class with 2800 MHz radio emission (F2800) was accomplished. An important finding is that during 1980 both the number of hydrogen alpha importance class 1 and number of X-ray class M (and M+X) flares appeared to be rather strongly related to F2800, in a positive sense; i.e., number of class 1 and class M events increased as F2800 increased.
Zheng, Jie; Harris, Marcelline R; Masci, Anna Maria; Lin, Yu; Hero, Alfred; Smith, Barry; He, Yongqun
2016-09-14
Statistics play a critical role in biological and clinical research. However, most reports of scientific results in the published literature make it difficult for the reader to reproduce the statistical analyses performed in achieving those results because they provide inadequate documentation of the statistical tests and algorithms applied. The Ontology of Biological and Clinical Statistics (OBCS) is put forward here as a step towards solving this problem. The terms in OBCS including 'data collection', 'data transformation in statistics', 'data visualization', 'statistical data analysis', and 'drawing a conclusion based on data', cover the major types of statistical processes used in basic biological research and clinical outcome studies. OBCS is aligned with the Basic Formal Ontology (BFO) and extends the Ontology of Biomedical Investigations (OBI), an OBO (Open Biological and Biomedical Ontologies) Foundry ontology supported by over 20 research communities. Currently, OBCS comprehends 878 terms, representing 20 BFO classes, 403 OBI classes, 229 OBCS specific classes, and 122 classes imported from ten other OBO ontologies. We discuss two examples illustrating how the ontology is being applied. In the first (biological) use case, we describe how OBCS was applied to represent the high throughput microarray data analysis of immunological transcriptional profiles in human subjects vaccinated with an influenza vaccine. In the second (clinical outcomes) use case, we applied OBCS to represent the processing of electronic health care data to determine the associations between hospital staffing levels and patient mortality. Our case studies were designed to show how OBCS can be used for the consistent representation of statistical analysis pipelines under two different research paradigms. Other ongoing projects using OBCS for statistical data processing are also discussed. The OBCS source code and documentation are available at: https://github.com/obcs/obcs . The Ontology of Biological and Clinical Statistics (OBCS) is a community-based open source ontology in the domain of biological and clinical statistics. OBCS is a timely ontology that represents statistics-related terms and their relations in a rigorous fashion, facilitates standard data analysis and integration, and supports reproducible biological and clinical research.
Alendronate for fracture prevention in postmenopause.
Holder, Kathryn K; Kerley, Sara Shelton
2008-09-01
Osteoporosis is an abnormal reduction in bone mass and bone deterioration leading to increased fracture risk. Alendronate (Fosamax) belongs to the bisphosphonate class of drugs, which act to inhibit bone resorption by interfering with the activity of osteoclasts. To assess the effectiveness of alendronate in the primary and secondary prevention of osteoporotic fractures in postmenopausal women. The authors searched Central, Medline, and EMBASE for relevant randomized controlled trials published from 1966 to 2007. The authors undertook study selection and data abstraction in duplicate. The authors performed meta-analysis of fracture outcomes using relative risks, and a relative change greater than 15 percent was considered clinically important. The authors assessed study quality through reporting of allocation concealment, blinding, and withdrawals. Eleven trials representing 12,068 women were included in the review. Relative and absolute risk reductions for the 10-mg dose were as follows. For vertebral fractures, a 45 percent relative risk reduction was found (relative risk [RR] = 0.55; 95% confidence interval [CI], 0.45 to 0.67). This was significant for primary prevention, with a 45 percent relative risk reduction (RR = 0.55; 95% CI, 0.38 to 0.80) and 2 percent absolute risk reduction; and for secondary prevention, with 45 percent relative risk reduction (RR = 0.55; 95% CI, 0.43 to 0.69) and 6 percent absolute risk reduction. For nonvertebral fractures, a 16 percent relative risk reduction was found (RR = 0.84; 95% CI, 0.74 to 0.94). This was significant for secondary prevention, with a 23 percent relative risk reduction (RR = 0.77; 95% CI, 0.64 to 0.92) and a 2 percent absolute risk reduction, but not for primary prevention (RR = 0.89; 95% CI, 0.76 to 1.04). There was a 40 percent relative risk reduction in hip fractures (RR = 0.60; 95% CI, 0.40 to 0.92), but only secondary prevention was significant, with a 53 percent relative risk reduction (RR = 0.47; 95% CI, 0.26 to 0.85) and a 1 percent absolute risk reduction. The only significance found for wrist fractures was in secondary prevention, with a 50 percent relative risk reduction (RR = 0.50; 95% CI, 0.34 to 0.73) and a 2 percent absolute risk reduction. For adverse events, the authors found no statistically significant difference in any included study. However, observational data raise concerns about potential risk for upper gastrointestinal injury and, less commonly, osteonecrosis of the jaw. At 10 mg of alendronate per day, clinically important and statistically significant reductions in vertebral, nonvertebral, hip, and wrist fractures were observed for secondary prevention. The authors found no statistically significant results for primary prevention, with the exception of vertebral fractures, for which the reduction was clinically important.
Tonkin, Matthew J.; Tiedeman, Claire; Ely, D. Matthew; Hill, Mary C.
2007-01-01
The OPR-PPR program calculates the Observation-Prediction (OPR) and Parameter-Prediction (PPR) statistics that can be used to evaluate the relative importance of various kinds of data to simulated predictions. The data considered fall into three categories: (1) existing observations, (2) potential observations, and (3) potential information about parameters. The first two are addressed by the OPR statistic; the third is addressed by the PPR statistic. The statistics are based on linear theory and measure the leverage of the data, which depends on the location, the type, and possibly the time of the data being considered. For example, in a ground-water system the type of data might be a head measurement at a particular location and time. As a measure of leverage, the statistics do not take into account the value of the measurement. As linear measures, the OPR and PPR statistics require minimal computational effort once sensitivities have been calculated. Sensitivities need to be calculated for only one set of parameter values; commonly these are the values estimated through model calibration. OPR-PPR can calculate the OPR and PPR statistics for any mathematical model that produces the necessary OPR-PPR input files. In this report, OPR-PPR capabilities are presented in the context of using the ground-water model MODFLOW-2000 and the universal inverse program UCODE_2005. The method used to calculate the OPR and PPR statistics is based on the linear equation for prediction standard deviation. Using sensitivities and other information, OPR-PPR calculates (a) the percent increase in the prediction standard deviation that results when one or more existing observations are omitted from the calibration data set; (b) the percent decrease in the prediction standard deviation that results when one or more potential observations are added to the calibration data set; or (c) the percent decrease in the prediction standard deviation that results when potential information on one or more parameters is added.
Preparing for the first meeting with a statistician.
De Muth, James E
2008-12-15
Practical statistical issues that should be considered when performing data collection and analysis are reviewed. The meeting with a statistician should take place early in the research development before any study data are collected. The process of statistical analysis involves establishing the research question, formulating a hypothesis, selecting an appropriate test, sampling correctly, collecting data, performing tests, and making decisions. Once the objectives are established, the researcher can determine the characteristics or demographics of the individuals required for the study, how to recruit volunteers, what type of data are needed to answer the research question(s), and the best methods for collecting the required information. There are two general types of statistics: descriptive and inferential. Presenting data in a more palatable format for the reader is called descriptive statistics. Inferential statistics involve making an inference or decision about a population based on results obtained from a sample of that population. In order for the results of a statistical test to be valid, the sample should be representative of the population from which it is drawn. When collecting information about volunteers, researchers should only collect information that is directly related to the study objectives. Important information that a statistician will require first is an understanding of the type of variables involved in the study and which variables can be controlled by researchers and which are beyond their control. Data can be presented in one of four different measurement scales: nominal, ordinal, interval, or ratio. Hypothesis testing involves two mutually exclusive and exhaustive statements related to the research question. Statisticians should not be replaced by computer software, and they should be consulted before any research data are collected. When preparing to meet with a statistician, the pharmacist researcher should be familiar with the steps of statistical analysis and consider several questions related to the study to be conducted.
Hagos, Seifu; Hailemariam, Damen; WoldeHanna, Tasew; Lindtjørn, Bernt
2017-01-01
Understanding the spatial distribution of stunting and underlying factors operating at meso-scale is of paramount importance for intervention designing and implementations. Yet, little is known about the spatial distribution of stunting and some discrepancies are documented on the relative importance of reported risk factors. Therefore, the present study aims at exploring the spatial distribution of stunting at meso- (district) scale, and evaluates the effect of spatial dependency on the identification of risk factors and their relative contribution to the occurrence of stunting and severe stunting in a rural area of Ethiopia. A community based cross sectional study was conducted to measure the occurrence of stunting and severe stunting among children aged 0-59 months. Additionally, we collected relevant information on anthropometric measures, dietary habits, parent and child-related demographic and socio-economic status. Latitude and longitude of surveyed households were also recorded. Local Anselin Moran's I was calculated to investigate the spatial variation of stunting prevalence and identify potential local pockets (hotspots) of high prevalence. Finally, we employed a Bayesian geo-statistical model, which accounted for spatial dependency structure in the data, to identify potential risk factors for stunting in the study area. Overall, the prevalence of stunting and severe stunting in the district was 43.7% [95%CI: 40.9, 46.4] and 21.3% [95%CI: 19.5, 23.3] respectively. We identified statistically significant clusters of high prevalence of stunting (hotspots) in the eastern part of the district and clusters of low prevalence (cold spots) in the western. We found out that the inclusion of spatial structure of the data into the Bayesian model has shown to improve the fit for stunting model. The Bayesian geo-statistical model indicated that the risk of stunting increased as the child's age increased (OR 4.74; 95% Bayesian credible interval [BCI]:3.35-6.58) and among boys (OR 1.28; 95%BCI; 1.12-1.45). However, maternal education and household food security were found to be protective against stunting and severe stunting. Stunting prevalence may vary across space at different scale. For this, it's important that nutrition studies and, more importantly, control interventions take into account this spatial heterogeneity in the distribution of nutritional deficits and their underlying associated factors. The findings of this study also indicated that interventions integrating household food insecurity in nutrition programs in the district might help to avert the burden of stunting.
Path statistics, memory, and coarse-graining of continuous-time random walks on networks
Kion-Crosby, Willow; Morozov, Alexandre V.
2015-01-01
Continuous-time random walks (CTRWs) on discrete state spaces, ranging from regular lattices to complex networks, are ubiquitous across physics, chemistry, and biology. Models with coarse-grained states (for example, those employed in studies of molecular kinetics) or spatial disorder can give rise to memory and non-exponential distributions of waiting times and first-passage statistics. However, existing methods for analyzing CTRWs on complex energy landscapes do not address these effects. Here we use statistical mechanics of the nonequilibrium path ensemble to characterize first-passage CTRWs on networks with arbitrary connectivity, energy landscape, and waiting time distributions. Our approach can be applied to calculating higher moments (beyond the mean) of path length, time, and action, as well as statistics of any conservative or non-conservative force along a path. For homogeneous networks, we derive exact relations between length and time moments, quantifying the validity of approximating a continuous-time process with its discrete-time projection. For more general models, we obtain recursion relations, reminiscent of transfer matrix and exact enumeration techniques, to efficiently calculate path statistics numerically. We have implemented our algorithm in PathMAN (Path Matrix Algorithm for Networks), a Python script that users can apply to their model of choice. We demonstrate the algorithm on a few representative examples which underscore the importance of non-exponential distributions, memory, and coarse-graining in CTRWs. PMID:26646868
Hawkins, Robert J; Kremer, Michael J; Swanson, Barbara; Fogg, Lou; Pierce, Penny; Pearson, Julie
2014-01-01
The level of patient satisfaction is a result of a complex set of interactions between the patient and the health care provider. It is important to quantify satisfaction with care because it involves the patient in the care experience and decreases the potential gap between expected and actual care delivered. We tested a preliminary 23-item instrument to measure patient satisfaction with general anesthesia care. The rating scale Rasch model was chosen as the framework. There were 10 items found to have sufficient evidence of stable fit statistics. Items included 2 questions related to information provided, 2 questions related to concern and kindness of the provider, and 1 question each for interpersonal skills of the provider, attention by the provider, feeling safe, well-being, privacy, and overall anesthesia satisfaction. Such actions as providing enough time to understand the anesthesia plan, answering questions related to the anesthetic, showing kindness and concern for the patient, displaying good interpersonal skills, providing adequate attention to the patient, providing a safe environment that maintains privacy and provides a sense of well-being are important actions that are well within the control of individual anesthesia providers and may lead to improved care from the perception of the patient.
Zhang, Kai; Li, Yun; Schwartz, Joel D.; O'Neill, Marie S.
2014-01-01
Hot weather increases risk of mortality. Previous studies used different sets of weather variables to characterize heat stress, resulting in variation in heat-mortality- associations depending on the metric used. We employed a statistical learning method – random forests – to examine which of various weather variables had the greatest impact on heat-related mortality. We compiled a summertime daily weather and mortality counts dataset from four U.S. cities (Chicago, IL; Detroit, MI; Philadelphia, PA; and Phoenix, AZ) from 1998 to 2006. A variety of weather variables were ranked in predicting deviation from typical daily all-cause and cause-specific death counts. Ranks of weather variables varied with city and health outcome. Apparent temperature appeared to be the most important predictor of heat-related mortality for all-cause mortality. Absolute humidity was, on average, most frequently selected one of the top variables for all-cause mortality and seven cause-specific mortality categories. Our analysis affirms that apparent temperature is a reasonable variable for activating heat alerts and warnings, which are commonly based on predictions of total mortality in next few days. Additionally, absolute humidity should be included in future heat-health studies. Finally, random forests can be used to guide choice of weather variables in heat epidemiology studies. PMID:24834832
A New Paradigm to Analyze Data Completeness of Patient Data.
Nasir, Ayan; Gurupur, Varadraj; Liu, Xinliang
2016-08-03
There is a need to develop a tool that will measure data completeness of patient records using sophisticated statistical metrics. Patient data integrity is important in providing timely and appropriate care. Completeness is an important step, with an emphasis on understanding the complex relationships between data fields and their relative importance in delivering care. This tool will not only help understand where data problems are but also help uncover the underlying issues behind them. Develop a tool that can be used alongside a variety of health care database software packages to determine the completeness of individual patient records as well as aggregate patient records across health care centers and subpopulations. The methodology of this project is encapsulated within the Data Completeness Analysis Package (DCAP) tool, with the major components including concept mapping, CSV parsing, and statistical analysis. The results from testing DCAP with Healthcare Cost and Utilization Project (HCUP) State Inpatient Database (SID) data show that this tool is successful in identifying relative data completeness at the patient, subpopulation, and database levels. These results also solidify a need for further analysis and call for hypothesis driven research to find underlying causes for data incompleteness. DCAP examines patient records and generates statistics that can be used to determine the completeness of individual patient data as well as the general thoroughness of record keeping in a medical database. DCAP uses a component that is customized to the settings of the software package used for storing patient data as well as a Comma Separated Values (CSV) file parser to determine the appropriate measurements. DCAP itself is assessed through a proof of concept exercise using hypothetical data as well as available HCUP SID patient data.
A New Paradigm to Analyze Data Completeness of Patient Data
Nasir, Ayan; Liu, Xinliang
2016-01-01
Summary Background There is a need to develop a tool that will measure data completeness of patient records using sophisticated statistical metrics. Patient data integrity is important in providing timely and appropriate care. Completeness is an important step, with an emphasis on understanding the complex relationships between data fields and their relative importance in delivering care. This tool will not only help understand where data problems are but also help uncover the underlying issues behind them. Objectives Develop a tool that can be used alongside a variety of health care database software packages to determine the completeness of individual patient records as well as aggregate patient records across health care centers and subpopulations. Methods The methodology of this project is encapsulated within the Data Completeness Analysis Package (DCAP) tool, with the major components including concept mapping, CSV parsing, and statistical analysis. Results The results from testing DCAP with Healthcare Cost and Utilization Project (HCUP) State Inpatient Database (SID) data show that this tool is successful in identifying relative data completeness at the patient, subpopulation, and database levels. These results also solidify a need for further analysis and call for hypothesis driven research to find underlying causes for data incompleteness. Conclusion DCAP examines patient records and generates statistics that can be used to determine the completeness of individual patient data as well as the general thoroughness of record keeping in a medical database. DCAP uses a component that is customized to the settings of the software package used for storing patient data as well as a Comma Separated Values (CSV) file parser to determine the appropriate measurements. DCAP itself is assessed through a proof of concept exercise using hypothetical data as well as available HCUP SID patient data. PMID:27484918
Kattainen, Eija; Sintonen, Harri; Kettunen, Raimo; Meriläinen, Pirkko
2005-01-01
The aim of the study was to compare the health-related quality of life (HRQoL) of patients undergoing coronary artery bypass grafting (CABG) or percutaneous transluminal coronary angioplasty (PTCA) before the interventions and 6 and 12 months afterward, and to compare their HRQoL also with that of the general population. The sample (n = 615) consisted of consecutive coronary artery disease patients treated with elective CABG (n=432) or PTCA (n=183). The baseline data before the treatments were collected by structured interview, the follow-up data mainly by mailed self-administered questionnaires. HRQoL was measured by the 15D. For comparisons, the groups were standardized for differences in socioeconomic and clinical characteristics with a regression analysis. At baseline, the average 15D scores of the patient groups were 0.752 (95 percent confidence interval [CI], 0.743-0.761) in CABG and 0.730 (95 percent CI, 0.716-0.744) in PTCA. After standardization, the difference between the groups was statistically significant but not clinically important. These scores were significantly worse (statistically and clinically) than the score of 0.883 (95 percent CI, 0.871-0.879) in the general population sample matched with the gender and age distribution of the patients. By 6 months, the CABG and PTCA patients had experienced a statistically significant and clinically important improvement to 0.858 (95 percent CI, 0.844-0.872) and 0.824 (95 percent CI, 0.806-0.842), respectively. No significant change took place in either group from 6 to 12 months. Both CABG and PTCA produces an approximately similar, clinically important improvement in HRQoL in 1-year follow-up.
Development of the Statistical Reasoning in Biology Concept Inventory (SRBCI)
Deane, Thomas; Nomme, Kathy; Jeffery, Erica; Pollock, Carol; Birol, Gülnur
2016-01-01
We followed established best practices in concept inventory design and developed a 12-item inventory to assess student ability in statistical reasoning in biology (Statistical Reasoning in Biology Concept Inventory [SRBCI]). It is important to assess student thinking in this conceptual area, because it is a fundamental requirement of being statistically literate and associated skills are needed in almost all walks of life. Despite this, previous work shows that non–expert-like thinking in statistical reasoning is common, even after instruction. As science educators, our goal should be to move students along a novice-to-expert spectrum, which could be achieved with growing experience in statistical reasoning. We used item response theory analyses (the one-parameter Rasch model and associated analyses) to assess responses gathered from biology students in two populations at a large research university in Canada in order to test SRBCI’s robustness and sensitivity in capturing useful data relating to the students’ conceptual ability in statistical reasoning. Our analyses indicated that SRBCI is a unidimensional construct, with items that vary widely in difficulty and provide useful information about such student ability. SRBCI should be useful as a diagnostic tool in a variety of biology settings and as a means of measuring the success of teaching interventions designed to improve statistical reasoning skills. PMID:26903497
Hammond, Matthew D; Cimpian, Andrei
2017-05-01
Stereotypes are typically defined as beliefs about groups, but this definition is underspecified. Beliefs about groups can be generic or statistical. Generic beliefs attribute features to entire groups (e.g., men are strong), whereas statistical beliefs encode the perceived prevalence of features (e.g., how common it is for men to be strong). In the present research, we sought to determine which beliefs-generic or statistical-are more central to the cognitive structure of stereotypes. Specifically, we tested whether generic or statistical beliefs are more influential in people's social judgments, on the assumption that greater functional importance indicates greater centrality in stereotype structure. Relative to statistical beliefs, generic beliefs about social groups were significantly stronger predictors of expectations (Studies 1-3) and explanations (Study 4) for unfamiliar individuals' traits. In addition, consistent with prior evidence that generic beliefs are cognitively simpler than statistical beliefs, generic beliefs were particularly predictive of social judgments for participants with more intuitive (vs. analytic) cognitive styles and for participants higher (vs. lower) in authoritarianism, who tend to view outgroups in simplistic, all-or-none terms. The present studies suggest that generic beliefs about groups are more central than statistical beliefs to the cognitive structure of stereotypes. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Statistical Perspectives on Stratospheric Transport
NASA Technical Reports Server (NTRS)
Sparling, L. C.
1999-01-01
Long-lived tropospheric source gases, such as nitrous oxide, enter the stratosphere through the tropical tropopause, are transported throughout the stratosphere by the Brewer-Dobson circulation, and are photochemically destroyed in the upper stratosphere. These chemical constituents, or "tracers" can be used to track mixing and transport by the stratospheric winds. Much of our understanding about the stratospheric circulation is based on large scale gradients and other spatial features in tracer fields constructed from satellite measurements. The point of view presented in this paper is different, but complementary, in that transport is described in terms of tracer probability distribution functions (PDFs). The PDF is computed from the measurements, and is proportional to the area occupied by tracer values in a given range. The flavor of this paper is tutorial, and the ideas are illustrated with several examples of transport-related phenomena, annotated with remarks that summarize the main point or suggest new directions. One example shows how the multimodal shape of the PDF gives information about the different branches of the circulation. Another example shows how the statistics of fluctuations from the most probable tracer value give insight into mixing between different regions of the atmosphere. Also included is an analysis of the time-dependence of the PDF during the onset and decline of the winter circulation, and a study of how "bursts" in the circulation are reflected in transient periods of rapid evolution of the PDF. The dependence of the statistics on location and time are also shown to be important for practical problems related to statistical robustness and satellite sampling. The examples illustrate how physically-based statistical analysis can shed some light on aspects of stratospheric transport that may not be obvious or quantifiable with other types of analyses. An important motivation for the work presented here is the need for synthesis of the large and growing database of observations of the atmosphere and the vast quantities of output generated by atmospheric models.
Bok, Harold G J; Teunissen, Pim W; Boerboom, Tobias B B; Rhind, Susan M; Baillie, Sarah; Tegzes, John; Annandale, Henry; Matthew, Susan; Torgersen, Anne; Hecker, Kent G; Härdi-Landerer, Christina M; Gomez-Lucia, Esperanza; Ahmad, Bashir; Muijtjens, Arno M M; Jaarsma, Debbie A D C; van der Vleuten, Cees P M; van Beukelen, Peter
2014-10-15
To determine the perceived importance of specific competencies in professional veterinary practice and education among veterinarians in several countries. Survey-based prospective study. 1,137 veterinarians in 10 countries. Veterinarians were invited via email to participate in the study. A framework of 18 competencies grouped into 7 domains (veterinary expertise, communication, collaboration, entrepreneurship, health and welfare, scholarship, and personal development) was used. Respondents rated the importance of each competency for veterinary professional practice and for veterinary education by use of a 9-point Likert scale in an online questionnaire. Quantitative statistical analyses were performed to assess the data. All described competencies were perceived as having importance (with overall mean ratings [all countries] ≥ 6.45/9) for professional practice and education. Competencies related to veterinary expertise had the highest ratings (overall mean, 8.33/9 for both professional practice and education). For the veterinary expertise, entrepreneurship, and scholarship domains, substantial differences (determined on the basis of statistical significance and effect size) were found in importance ratings among veterinarians in different countries. Results indicated a general consensus regarding the importance of specific types of competencies in veterinary professional practice and education. Further research into the definition of competencies essential for veterinary professionals is needed to help inform an international dialogue on the subject.
Ozge, C; Toros, F; Bayramkaya, E; Camdeviren, H; Sasmaz, T
2006-08-01
The purpose of this study is to evaluate the most important sociodemographic factors on smoking status of high school students using a broad randomised epidemiological survey. Using in-class, self administered questionnaire about their sociodemographic variables and smoking behaviour, a representative sample of total 3304 students of preparatory, 9th, 10th, and 11th grades, from 22 randomly selected schools of Mersin, were evaluated and discriminative factors have been determined using appropriate statistics. In addition to binary logistic regression analysis, the study evaluated combined effects of these factors using classification and regression tree methodology, as a new statistical method. The data showed that 38% of the students reported lifetime smoking and 16.9% of them reported current smoking with a male predominancy and increasing prevalence by age. Second hand smoking was reported at a 74.3% frequency with father predominance (56.6%). The significantly important factors that affect current smoking in these age groups were increased by household size, late birth rank, certain school types, low academic performance, increased second hand smoking, and stress (especially reported as separation from a close friend or because of violence at home). Classification and regression tree methodology showed the importance of some neglected sociodemographic factors with a good classification capacity. It was concluded that, as closely related with sociocultural factors, smoking was a common problem in this young population, generating important academic and social burden in youth life and with increasing data about this behaviour and using new statistical methods, effective coping strategies could be composed.
Federal Register 2010, 2011, 2012, 2013, 2014
2012-11-06
... exceeded will be effective July 1, 2013, unless the President grants a waiver before the exclusion goes... available in February 2013 on the Web site of the U.S. International Trade Commission at http://dataweb....10.05--Coniferous wood continuously shaped along any of its ends (Brazil) 7202.99.20--Calcium silicon...
ERIC Educational Resources Information Center
Onur, Arzu; Sahin, Elvan; Tekkaya, Ceren
2012-01-01
Environmental attitudes depend on the relative importance that individuals attach to themselves, other people, or all living things. These distinct bases have been found to predict environmental concern, and may act as statistically significant determinants of pro-environmental behaviours. We claim that examining the complex nature of value…
Workforce Education and Two Important Viewpoints
ERIC Educational Resources Information Center
Dennis, Dawn Holley; Hudson, Clemente Charles
2007-01-01
Workforce Education appears to be a goldmine in the state of Florida. As of December 2003, annual job growth statistics reveal that Florida ranked number one in the nation in new jobs and tied for number one in percent change relative to other populous states (Georgia, Texas, New Jersey, and New York). The purpose of Florida's Workforce Education…
ERIC Educational Resources Information Center
Walker, A. Adrienne; Jennings, Jeremy Kyle; Engelhard, George, Jr.
2018-01-01
Individual person fit analyses provide important information regarding the validity of test score inferences for an "individual" test taker. In this study, we use data from an undergraduate statistics test (N = 1135) to illustrate a two-step method that researchers and practitioners can use to examine individual person fit. First, person…
Eagles-Smith, Collin A.; Herring, Garth; Johnson, Branden L.; Graw, Rick
2016-01-01
Remote high-elevation lakes represent unique environments for evaluating the bioaccumulation of atmospherically deposited mercury through freshwater food webs, as well as for evaluating the relative importance of mercury loading versus landscape influences on mercury bioaccumulation. The increase in mercury deposition to these systems over the past century, coupled with their limited exposure to direct anthropogenic disturbance make them useful indicators for estimating how changes in mercury emissions may propagate to changes in Hg bioaccumulation and ecological risk. We evaluated mercury concentrations in resident fish from 28 high-elevation, sub-alpine lakes in the Pacific Northwest region of the United States. Fish total mercury (THg) concentrations ranged from 4 to 438 ng/g wet weight, with a geometric mean concentration (±standard error) of 43 ± 2 ng/g ww. Fish THg concentrations were negatively correlated with relative condition factor, indicating that faster growing fish that are in better condition have lower THg concentrations. Across the 28 study lakes, mean THg concentrations of resident salmonid fishes varied as much as 18-fold among lakes. We used a hierarchal statistical approach to evaluate the relative importance of physiological, limnological, and catchment drivers of fish Hg concentrations. Our top statistical model explained 87% of the variability in fish THg concentrations among lakes with four key landscape and limnological variables: catchment conifer density (basal area of conifers within a lake's catchment), lake surface area, aqueous dissolved sulfate, and dissolved organic carbon. Conifer density within a lake's catchment was the most important variable explaining fish THg concentrations across lakes, with THg concentrations differing by more than 400 percent across the forest density spectrum. These results illustrate the importance of landscape characteristics in controlling mercury bioaccumulation in fish.
Eagles-Smith, Collin A; Herring, Garth; Johnson, Branden; Graw, Rick
2016-05-01
Remote high-elevation lakes represent unique environments for evaluating the bioaccumulation of atmospherically deposited mercury through freshwater food webs, as well as for evaluating the relative importance of mercury loading versus landscape influences on mercury bioaccumulation. The increase in mercury deposition to these systems over the past century, coupled with their limited exposure to direct anthropogenic disturbance make them useful indicators for estimating how changes in mercury emissions may propagate to changes in Hg bioaccumulation and ecological risk. We evaluated mercury concentrations in resident fish from 28 high-elevation, sub-alpine lakes in the Pacific Northwest region of the United States. Fish total mercury (THg) concentrations ranged from 4 to 438 ng/g wet weight, with a geometric mean concentration (±standard error) of 43 ± 2 ng/g ww. Fish THg concentrations were negatively correlated with relative condition factor, indicating that faster growing fish that are in better condition have lower THg concentrations. Across the 28 study lakes, mean THg concentrations of resident salmonid fishes varied as much as 18-fold among lakes. We used a hierarchal statistical approach to evaluate the relative importance of physiological, limnological, and catchment drivers of fish Hg concentrations. Our top statistical model explained 87% of the variability in fish THg concentrations among lakes with four key landscape and limnological variables: catchment conifer density (basal area of conifers within a lake's catchment), lake surface area, aqueous dissolved sulfate, and dissolved organic carbon. Conifer density within a lake's catchment was the most important variable explaining fish THg concentrations across lakes, with THg concentrations differing by more than 400 percent across the forest density spectrum. These results illustrate the importance of landscape characteristics in controlling mercury bioaccumulation in fish. Published by Elsevier Ltd.
Games and gambling involvement among casino patrons.
LaPlante, Debi A; Afifi, Tracie O; Shaffer, Howard J
2013-06-01
A growing literature is addressing the nature of the relationships among gambling activity, gambling involvement, and gambling-related problems. This research suggests that among the general population, compared to playing any specific game, gambling involvement is a better predictor of gambling-related problems. To date, researchers have not examined these relationships among casino patrons, a population that differs from the general population in a variety of important ways. A survey of 1160 casino patrons at two Las Vegas resort casinos allowed us to determine relationships between the games that patrons played during the 12 months before their casino visit, the games that patrons played during their casino visit, and patrons' self-perceived history of gambling-related problems. Results indicate that playing specific gambling games onsite predicted (i.e., statistically significant odds ratios ranging from .5 to 4.51) self-perceived gambling-related problems. However, after controlling for involvement, operationally defined as the number of games played during the current casino visit and self-reported gambling frequency during the past 12 months, the relationships between games and gambling-related problems disappeared or were attenuated (i.e., odds ratios no longer statistically significant). These results extend the burgeoning literature related to gambling involvement and its relationship to gambling-related problems.
Application of real rock pore-threat statistics to a regular pore network model
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rakibul, M.; Sarker, H.; McIntyre, D.
2011-01-01
This work reports the application of real rock statistical data to a previously developed regular pore network model in an attempt to produce an accurate simulation tool with low computational overhead. A core plug from the St. Peter Sandstone formation in Indiana was scanned with a high resolution micro CT scanner. The pore-throat statistics of the three-dimensional reconstructed rock were extracted and the distribution of the pore-throat sizes was applied to the regular pore network model. In order to keep the equivalent model regular, only the throat area or the throat radius was varied. Ten realizations of randomly distributed throatmore » sizes were generated to simulate the drainage process and relative permeability was calculated and compared with the experimentally determined values of the original rock sample. The numerical and experimental procedures are explained in detail and the performance of the model in relation to the experimental data is discussed and analyzed. Petrophysical properties such as relative permeability are important in many applied fields such as production of petroleum fluids, enhanced oil recovery, carbon dioxide sequestration, ground water flow, etc. Relative permeability data are used for a wide range of conventional reservoir engineering calculations and in numerical reservoir simulation. Two-phase oil water relative permeability data are generated on the same core plug from both pore network model and experimental procedure. The shape and size of the relative permeability curves were compared and analyzed and good match has been observed for wetting phase relative permeability but for non-wetting phase, simulation results were found to be deviated from the experimental ones. Efforts to determine petrophysical properties of rocks using numerical techniques are to eliminate the necessity of regular core analysis, which can be time consuming and expensive. So a numerical technique is expected to be fast and to produce reliable results. In applied engineering, sometimes quick result with reasonable accuracy is acceptable than the more time consuming results. Present work is an effort to check the accuracy and validity of a previously developed pore network model for obtaining important petrophysical properties of rocks based on cutting-sized sample data.« less
Application of real rock pore-throat statistics to a regular pore network model
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sarker, M.R.; McIntyre, D.; Ferer, M.
2011-01-01
This work reports the application of real rock statistical data to a previously developed regular pore network model in an attempt to produce an accurate simulation tool with low computational overhead. A core plug from the St. Peter Sandstone formation in Indiana was scanned with a high resolution micro CT scanner. The pore-throat statistics of the three-dimensional reconstructed rock were extracted and the distribution of the pore-throat sizes was applied to the regular pore network model. In order to keep the equivalent model regular, only the throat area or the throat radius was varied. Ten realizations of randomly distributed throatmore » sizes were generated to simulate the drainage process and relative permeability was calculated and compared with the experimentally determined values of the original rock sample. The numerical and experimental procedures are explained in detail and the performance of the model in relation to the experimental data is discussed and analyzed. Petrophysical properties such as relative permeability are important in many applied fields such as production of petroleum fluids, enhanced oil recovery, carbon dioxide sequestration, ground water flow, etc. Relative permeability data are used for a wide range of conventional reservoir engineering calculations and in numerical reservoir simulation. Two-phase oil water relative permeability data are generated on the same core plug from both pore network model and experimental procedure. The shape and size of the relative permeability curves were compared and analyzed and good match has been observed for wetting phase relative permeability but for non-wetting phase, simulation results were found to be deviated from the experimental ones. Efforts to determine petrophysical properties of rocks using numerical techniques are to eliminate the necessity of regular core analysis, which can be time consuming and expensive. So a numerical technique is expected to be fast and to produce reliable results. In applied engineering, sometimes quick result with reasonable accuracy is acceptable than the more time consuming results. Present work is an effort to check the accuracy and validity of a previously developed pore network model for obtaining important petrophysical properties of rocks based on cutting-sized sample data. Introduction« less
Regojo Zapata, O; Lamata Hernández, F; Sánchez Zalabardo, J M; Elizalde Benito, A; Navarro Gil, J; Valdivia Uría, J G
2004-09-01
Studies about quality in thesis and investigation projects in biomedical sciences are unusual, but very important in university teaching because is necessary to improve the quality elaboration of the thesis. The objectives the study were to determine the project's quality of thesis in our department, according to the fulfillment of the scientific methodology and to establish, if it exists, a relation between the global quality of the project and the statistical used resources. Descriptive study of 273 thesis projects performed between 1995-2002 in surgery department of the Zaragoza University. The review realized for 15 observers that they analyzed 28 indicators of every project. Giving a value to each of the indicators, the projects qualified in a scale from 1 to 10 according to the quality in the fulfillment of the scientific methodology. The mean of the project's quality was 5.53 (D.E: 1.77). In 13.9% the thesis projects was concluded with the reading of the work. The three indicators of statistical used resources had a significant difference with the value of the quality projects. The quality of the statistical resources is very important when a project of thesis wants to be realized by good methodology, because it assures to come to certain conclusions. In our study we have thought that more of the third part of the variability in the quality of the project of thesis explains for three statistical above-mentioned articles.
Data-Driven Learning of Q-Matrix
Liu, Jingchen; Xu, Gongjun; Ying, Zhiliang
2013-01-01
The recent surge of interests in cognitive assessment has led to developments of novel statistical models for diagnostic classification. Central to many such models is the well-known Q-matrix, which specifies the item–attribute relationships. This article proposes a data-driven approach to identification of the Q-matrix and estimation of related model parameters. A key ingredient is a flexible T-matrix that relates the Q-matrix to response patterns. The flexibility of the T-matrix allows the construction of a natural criterion function as well as a computationally amenable algorithm. Simulations results are presented to demonstrate usefulness and applicability of the proposed method. Extension to handling of the Q-matrix with partial information is presented. The proposed method also provides a platform on which important statistical issues, such as hypothesis testing and model selection, may be formally addressed. PMID:23926363
Snakebite mortality in the world
Swaroop, S.; Grab, B.
1954-01-01
In examining the relative importance of snakebite mortality in different parts of the world, the authors review the information collected concerning both snakebite mortality and the species of snake incriminated. Available statistical data are known to be unreliable and at best can serve to provide only an approximate and highly conservative estimate of the relative magnitude of the snakebite problem. The sources of error inherent in the data are discussed, and estimates are made of the probable mortality from snakebite in various areas of the world. PMID:13150169
Applications of spatial statistical network models to stream data
Isaak, Daniel J.; Peterson, Erin E.; Ver Hoef, Jay M.; Wenger, Seth J.; Falke, Jeffrey A.; Torgersen, Christian E.; Sowder, Colin; Steel, E. Ashley; Fortin, Marie-Josée; Jordan, Chris E.; Ruesch, Aaron S.; Som, Nicholas; Monestiez, Pascal
2014-01-01
Streams and rivers host a significant portion of Earth's biodiversity and provide important ecosystem services for human populations. Accurate information regarding the status and trends of stream resources is vital for their effective conservation and management. Most statistical techniques applied to data measured on stream networks were developed for terrestrial applications and are not optimized for streams. A new class of spatial statistical model, based on valid covariance structures for stream networks, can be used with many common types of stream data (e.g., water quality attributes, habitat conditions, biological surveys) through application of appropriate distributions (e.g., Gaussian, binomial, Poisson). The spatial statistical network models account for spatial autocorrelation (i.e., nonindependence) among measurements, which allows their application to databases with clustered measurement locations. Large amounts of stream data exist in many areas where spatial statistical analyses could be used to develop novel insights, improve predictions at unsampled sites, and aid in the design of efficient monitoring strategies at relatively low cost. We review the topic of spatial autocorrelation and its effects on statistical inference, demonstrate the use of spatial statistics with stream datasets relevant to common research and management questions, and discuss additional applications and development potential for spatial statistics on stream networks. Free software for implementing the spatial statistical network models has been developed that enables custom applications with many stream databases.
Jacob, Laurent; Combes, Florence; Burger, Thomas
2018-06-18
We propose a new hypothesis test for the differential abundance of proteins in mass-spectrometry based relative quantification. An important feature of this type of high-throughput analyses is that it involves an enzymatic digestion of the sample proteins into peptides prior to identification and quantification. Due to numerous homology sequences, different proteins can lead to peptides with identical amino acid chains, so that their parent protein is ambiguous. These so-called shared peptides make the protein-level statistical analysis a challenge and are often not accounted for. In this article, we use a linear model describing peptide-protein relationships to build a likelihood ratio test of differential abundance for proteins. We show that the likelihood ratio statistic can be computed in linear time with the number of peptides. We also provide the asymptotic null distribution of a regularized version of our statistic. Experiments on both real and simulated datasets show that our procedures outperforms state-of-the-art methods. The procedures are available via the pepa.test function of the DAPAR Bioconductor R package.
Statistical aspects of solar flares
NASA Technical Reports Server (NTRS)
Wilson, Robert M.
1987-01-01
A survey of the statistical properties of 850 H alpha solar flares during 1975 is presented. Comparison of the results found here with those reported elsewhere for different epochs is accomplished. Distributions of rise time, decay time, and duration are given, as are the mean, mode, median, and 90th percentile values. Proportions by selected groupings are also determined. For flares in general, mean values for rise time, decay time, and duration are 5.2 + or - 0.4 min, and 18.1 + or 1.1 min, respectively. Subflares, accounting for nearly 90 percent of the flares, had mean values lower than those found for flares of H alpha importance greater than 1, and the differences are statistically significant. Likewise, flares of bright and normal relative brightness have mean values of decay time and duration that are significantly longer than those computed for faint flares, and mass-motion related flares are significantly longer than non-mass-motion related flares. Seventy-three percent of the mass-motion related flares are categorized as being a two-ribbon flare and/or being accompanied by a high-speed dark filament. Slow rise time flares (rise time greater than 5 min) have a mean value for duration that is significantly longer than that computed for fast rise time flares, and long-lived duration flares (duration greater than 18 min) have a mean value for rise time that is significantly longer than that computed for short-lived duration flares, suggesting a positive linear relationship between rise time and duration for flares. Monthly occurrence rates for flares in general and by group are found to be linearly related in a positive sense to monthly sunspot number. Statistical testing reveals the association between sunspot number and numbers of flares to be significant at the 95 percent level of confidence, and the t statistic for slope is significant at greater than 99 percent level of confidence. Dependent upon the specific fit, between 58 percent and 94 percent of the variation can be accounted for with the linear fits. A statistically significant Northern Hemisphere flare excess (P less than 1 percent) was found, as was a Western Hemisphere excess (P approx 3 percent). Subflares were more prolific within 45 deg of central meridian (P less than 1 percent), while flares of H alpha importance or = 1 were more prolific near the limbs greater than 45 deg from central meridian; P approx 2 percent). Two-ribbon flares were more frequent within 45 deg of central meridian (P less than 1 percent). Slow rise time flares occurred more frequently in the western hemisphere (P approx 2 percent), as did short-lived duration flares (P approx 9 percent), but fast rise time flares were not preferentially distributed (in terms of east-west or limb-disk). Long-lived duration flares occurred more often within 45 deg 0 central meridian (P approx 7 percent). Mean durations for subflares and flares of H alpha importance or + 1, found within 45 deg of central meridian, are 14 percent and 70 percent, respectively, longer than those found for flares closer to the limb. As compared to flares occurring near cycle maximum, the flares of 1975 (near solar minimum) have mean values of rise time, decay time, and duration that are significantly shorter. A flare near solar maximum, on average, is about 1.6 times longer than one occurring near solar minimum.
Dealing with dietary measurement error in nutritional cohort studies.
Freedman, Laurence S; Schatzkin, Arthur; Midthune, Douglas; Kipnis, Victor
2011-07-20
Dietary measurement error creates serious challenges to reliably discovering new diet-disease associations in nutritional cohort studies. Such error causes substantial underestimation of relative risks and reduction of statistical power for detecting associations. On the basis of data from the Observing Protein and Energy Nutrition Study, we recommend the following approaches to deal with these problems. Regarding data analysis of cohort studies using food-frequency questionnaires, we recommend 1) using energy adjustment for relative risk estimation; 2) reporting estimates adjusted for measurement error along with the usual relative risk estimates, whenever possible (this requires data from a relevant, preferably internal, validation study in which participants report intakes using both the main instrument and a more detailed reference instrument such as a 24-hour recall or multiple-day food record); 3) performing statistical adjustment of relative risks, based on such validation data, if they exist, using univariate (only for energy-adjusted intakes such as densities or residuals) or multivariate regression calibration. We note that whereas unadjusted relative risk estimates are biased toward the null value, statistical significance tests of unadjusted relative risk estimates are approximately valid. Regarding study design, we recommend increasing the sample size to remedy loss of power; however, it is important to understand that this will often be an incomplete solution because the attenuated signal may be too small to distinguish from unmeasured confounding in the model relating disease to reported intake. Future work should be devoted to alleviating the problem of signal attenuation, possibly through the use of improved self-report instruments or by combining dietary biomarkers with self-report instruments.
Hong, Peilong; Li, Liming; Liu, Jianji; Zhang, Guoquan
2016-03-29
Young's double-slit or two-beam interference is of fundamental importance to understand various interference effects, in which the stationary phase difference between two beams plays the key role in the first-order coherence. Different from the case of first-order coherence, in the high-order optical coherence the statistic behavior of the optical phase will play the key role. In this article, by employing a fundamental interfering configuration with two classical point sources, we showed that the high- order optical coherence between two classical point sources can be actively designed by controlling the statistic behavior of the relative phase difference between two point sources. Synchronous position Nth-order subwavelength interference with an effective wavelength of λ/M was demonstrated, in which λ is the wavelength of point sources and M is an integer not larger than N. Interestingly, we found that the synchronous position Nth-order interference fringe fingerprints the statistic trace of random phase fluctuation of two classical point sources, therefore, it provides an effective way to characterize the statistic properties of phase fluctuation for incoherent light sources.
Jezberová, Jitka; Jezbera, Jan; Brandt, Ulrike; Lindström, Eva S.; Langenheder, Silke; Hahn, Martin W.
2010-01-01
Summary We present a survey on the distribution and habitat range of P. necessarius subspecies asymbioticus (PnecC), an important taxon in the water column of freshwater systems. We systematically sampled stagnant freshwater habitats in a heterogeneous 2000 km2 area, together with ecologically different habitats outside this area. In total, 137 lakes, ponds and puddles were investigated, which represent an enormous diversity of habitats differing, i.e., in depth (<10 cm – 171 m) and pH (3.9 – 8.5). PnecC was detected by cultivation-independent methods in all investigated habitats, and its presence was confirmed by cultivation of strains from selected habitats including the most extreme ones. The determined relative abundance of the subspecies ranged from slightly above 0% to 67% (average 14.5% ± 14.3%), and the highest observed absolute abundance was 5.3×106 cells mL−1. Statistical analyses revealed that the abundance of PnecC is partially controlled by factors linked to concentrations of humic substances, which might support the hypothesis that these bacteria utilize photodegradation products of humic substances. . Statistical analyses revealed that the abundance of PnecC is partially controlled by low conductivity and pH and factors linked to concentrations of humic substances. Based on the revealed statistical relationships, an average relative abundance of this subspecies of 20% in global freshwater habitats was extrapolated. Our study provides important implications for the current debate on ubiquity and biogeography in microorganisms. PMID:20041938
Clinical significance in nursing research: A discussion and descriptive analysis.
Polit, Denise F
2017-08-01
It is widely understood that statistical significance should not be equated with clinical significance, but the topic of clinical significance has not received much attention in the nursing literature. By contrast, interest in conceptualizing and operationalizing clinical significance has been a "hot topic" in other health care fields for several decades. The major purpose of this paper is to briefly describe recent advances in defining and quantifying clinical significance. The overview covers both group-level indicators of clinical significance (e.g., effect size indexes), and individual-level benchmarks (e.g., the minimal important change index). A secondary purpose is to describe the extent to which developments in clinical significance have penetrated the nursing literature. A descriptive analysis of a sample of primary research articles published in three high-impact nursing research journals in 2016 was undertaken. A total of 362 articles were electronically searched for terms relating to statistical and clinical significance. Of the 362 articles, 261 were reports of quantitative studies, the vast majority of which (93%) included a formal evaluation of the statistical significance of the results. By contrast, the term "clinical significance" or related surrogate terms were found in only 33 papers, and most often the term was used informally, without explicit definition or assessment. Raising consciousness about clinical significance should be an important priority among nurse researchers. Several recommendations are offered to improve the visibility and salience of clinical significance in nursing science. Copyright © 2017 Elsevier Ltd. All rights reserved.
Lazarides, Rebecca; Watt, Helen M G
2017-12-01
According to Eccles and Jacobs' (1986) parent socialization model, parents' gendered ability and value beliefs influence girls' and boys' interpretations of those beliefs, and hence students' domain-specific valuing of tasks and competence beliefs and subsequent career plans. Studies have rarely analyzed how both student-perceived mothers' and fathers' beliefs affect girls' and boys' task values, success expectancies, and career plans across domains. This study analyzed survey data of 459 students (262 boys) assessed through Grades 9, 10, and 11 from three coeducational secondary schools in Sydney, Australia. Longitudinal structural equation models revealed gendered value transmission pathways for girls in mathematics. Although mathematics test scores did not vary statistically significantly, girls reported statistically significantly lower mothers' ability beliefs for them in mathematics than boys at Time 1, which led to their statistically significantly lower mathematics intrinsic value at Time 2 and mathematics-related career plans at Time 3. Such gendered pathways did not occur in English. Matched same-gender effects and gendered pathways in parent socialization processes were evident; perceived mothers' value beliefs were more strongly related to girls' than boys' importance values in English. Student-perceived fathers' ability beliefs positively predicted boys', not girls', importance value in mathematics. Implications for educational practice emphasize the need to target girls' and boys' interest when aiming to enhance their mathematical career motivations. © 2017 The Authors. Journal of Research on Adolescence © 2017 Society for Research on Adolescence.
Statistical universals reveal the structures and functions of human music.
Savage, Patrick E; Brown, Steven; Sakai, Emi; Currie, Thomas E
2015-07-21
Music has been called "the universal language of mankind." Although contemporary theories of music evolution often invoke various musical universals, the existence of such universals has been disputed for decades and has never been empirically demonstrated. Here we combine a music-classification scheme with statistical analyses, including phylogenetic comparative methods, to examine a well-sampled global set of 304 music recordings. Our analyses reveal no absolute universals but strong support for many statistical universals that are consistent across all nine geographic regions sampled. These universals include 18 musical features that are common individually as well as a network of 10 features that are commonly associated with one another. They span not only features related to pitch and rhythm that are often cited as putative universals but also rarely cited domains including performance style and social context. These cross-cultural structural regularities of human music may relate to roles in facilitating group coordination and cohesion, as exemplified by the universal tendency to sing, play percussion instruments, and dance to simple, repetitive music in groups. Our findings highlight the need for scientists studying music evolution to expand the range of musical cultures and musical features under consideration. The statistical universals we identified represent important candidates for future investigation.
Statistical universals reveal the structures and functions of human music
Savage, Patrick E.; Brown, Steven; Sakai, Emi; Currie, Thomas E.
2015-01-01
Music has been called “the universal language of mankind.” Although contemporary theories of music evolution often invoke various musical universals, the existence of such universals has been disputed for decades and has never been empirically demonstrated. Here we combine a music-classification scheme with statistical analyses, including phylogenetic comparative methods, to examine a well-sampled global set of 304 music recordings. Our analyses reveal no absolute universals but strong support for many statistical universals that are consistent across all nine geographic regions sampled. These universals include 18 musical features that are common individually as well as a network of 10 features that are commonly associated with one another. They span not only features related to pitch and rhythm that are often cited as putative universals but also rarely cited domains including performance style and social context. These cross-cultural structural regularities of human music may relate to roles in facilitating group coordination and cohesion, as exemplified by the universal tendency to sing, play percussion instruments, and dance to simple, repetitive music in groups. Our findings highlight the need for scientists studying music evolution to expand the range of musical cultures and musical features under consideration. The statistical universals we identified represent important candidates for future investigation. PMID:26124105
Ranganathan, Priya; Pramesh, C. S.; Buyse, Marc
2015-01-01
In the second part of a series on pitfalls in statistical analysis, we look at various ways in which a statistically significant study result can be expressed. We debunk some of the myths regarding the ‘P’ value, explain the importance of ‘confidence intervals’ and clarify the importance of including both values in a paper PMID:25878958
NASA Astrophysics Data System (ADS)
Agus, M.; Hitchcott, P. K.; Penna, M. P.; Peró-Cebollero, M.; Guàrdia-Olmos, J.
2016-11-01
Many studies have investigated the features of probabilistic reasoning developed in relation to different formats of problem presentation, showing that it is affected by various individual and contextual factors. Incomplete understanding of the identity and role of these factors may explain the inconsistent evidence concerning the effect of problem presentation format. Thus, superior performance has sometimes been observed for graphically, rather than verbally, presented problems. The present study was undertaken to address this issue. Psychology undergraduates without any statistical expertise (N = 173 in Italy; N = 118 in Spain; N = 55 in England) were administered statistical problems in two formats (verbal-numerical and graphical-pictorial) under a condition of time pressure. Students also completed additional measures indexing several potentially relevant individual dimensions (statistical ability, statistical anxiety, attitudes towards statistics and confidence). Interestingly, a facilitatory effect of graphical presentation was observed in the Italian and Spanish samples but not in the English one. Significantly, the individual dimensions predicting statistical performance also differed between the samples, highlighting a different role of confidence. Hence, these findings confirm previous observations concerning problem presentation format while simultaneously highlighting the importance of individual dimensions.
Statistical dielectronic recombination rates for multielectron ions in plasma
NASA Astrophysics Data System (ADS)
Demura, A. V.; Leont'iev, D. S.; Lisitsa, V. S.; Shurygin, V. A.
2017-10-01
We describe the general analytic derivation of the dielectronic recombination (DR) rate coefficient for multielectron ions in a plasma based on the statistical theory of an atom in terms of the spatial distribution of the atomic electron density. The dielectronic recombination rates for complex multielectron tungsten ions are calculated numerically in a wide range of variation of the plasma temperature, which is important for modern nuclear fusion studies. The results of statistical theory are compared with the data obtained using level-by-level codes ADPAK, FAC, HULLAC, and experimental results. We consider different statistical DR models based on the Thomas-Fermi distribution, viz., integral and differential with respect to the orbital angular momenta of the ion core and the trapped electron, as well as the Rost model, which is an analog of the Frank-Condon model as applied to atomic structures. In view of its universality and relative simplicity, the statistical approach can be used for obtaining express estimates of the dielectronic recombination rate coefficients in complex calculations of the parameters of the thermonuclear plasmas. The application of statistical methods also provides information for the dielectronic recombination rates with much smaller computer time expenditures as compared to available level-by-level codes.
Schmitter, Marc; Kress, Bodo; Leckel, Michael; Henschel, Volkmar; Ohlmann, Brigitte; Rammelsberg, Peter
2008-06-01
This hypothesis-generating study was performed to determine which items in the Research Diagnostic Criteria for Temporomandibular Disorders (RDC/TMD) and additional diagnostic tests have the best predictive accuracy for joint-related diagnoses. One hundred forty-nine TMD patients and 43 symptom-free subjects were examined in clinical examinations and with magnetic resonance imaging (MRI). The importance of each variable of the clinical examination for correct joint-related diagnosis was assessed by using MRI diagnoses. For this purpose, "random forest" statistical software (based on classification trees) was used. Maximum unassisted jaw opening, maximum assisted jaw opening, history of locked jaw, joint sound with and without compression, joint pain, facial pain, pain on palpation of the lateral pterygoid area, and overjet proved suitable for distinguishing between subtypes of joint-related TMD. Measurement of excursion, protrusion, and midline deviation were less important. The validity of clinical TMD examination procedures can be enhanced by using the 16 variables of greatest importance identified in this study. In addition to other variables, maximum unassisted and assisted opening and a history of locked jaw were important when assessing the status of the TMJ.
Needs of relatives of critically ill patients in an academic hospital in Chile.
Padilla-Fortunatti, C; Rojas-Silva, N; Amthauer-Rojas, M; Molina-Muñoz, Y
To identify the importance of the needs of family members of patients in an intensive care unit (ICU). Descriptive, comparative and cross-sectional study based on a secondary data analysis of 251 relatives of ICU patients at a university hospital in Santiago, Chile. Using non-random sampling, the 'Critical Care Family Needs Inventory' was used to establish the family needs, as well as a sociodemographic questionnaire that included: age, gender, educational level, patient relationship and previous ICU experience. A descriptive statistical analysis, Student's T test and ANOVA were performed. The most important family needs related to the dimensions of 'security' (mean=3.90) and 'information' (mean=3.76), while those of minor importance with 'support' (mean=3.09). In the latter, differences were observed at an older age (P<.05), an educational level (P<.001) and relationship with the patient (P<.05). The most relevant needs for family members in the ICU are related to safety and information. Less important needs are influenced by certain sociodemographic variables. Identifying the degree of importance of family needs will allow the health team to improve its relationship with families in ICUs. Copyright © 2017 Sociedad Española de Enfermería Intensiva y Unidades Coronarias (SEEIUC). Publicado por Elsevier España, S.L.U. All rights reserved.
Pang, Jingxiang; Fu, Jialei; Yang, Meina; Zhao, Xiaolei; van Wijk, Eduard; Wang, Mei; Fan, Hua; Han, Jinxiang
2016-03-01
In the practice and principle of Chinese medicine, herbal materials are classified according to their therapeutic properties. 'Cold' and 'heat' are the most important classes of Chinese medicinal herbs according to the theory of traditional Chinese medicine (TCM). In this work, delayed luminescence (DL) was measured for different samples of Chinese medicinal herbs using a sensitive photon multiplier detection system. A comparison of DL parameters, including mean intensity and statistic entropy, was undertaken to discriminate between the 'cold' and 'heat' properties of Chinese medicinal herbs. The results suggest that there are significant differences in mean intensity and statistic entropy and using this method combined with statistical analysis may provide novel parameters for the characterization of Chinese medicinal herbs in relation to their energetic properties. Copyright © 2015 John Wiley & Sons, Ltd.
How Many Is Enough?—Statistical Principles for Lexicostatistics
Zhang, Menghan; Gong, Tao
2016-01-01
Lexicostatistics has been applied in linguistics to inform phylogenetic relations among languages. There are two important yet not well-studied parameters in this approach: the conventional size of vocabulary list to collect potentially true cognates and the minimum matching instances required to confirm a recurrent sound correspondence. Here, we derive two statistical principles from stochastic theorems to quantify these parameters. These principles validate the practice of using the Swadesh 100- and 200-word lists to indicate degree of relatedness between languages, and enable a frequency-based, dynamic threshold to detect recurrent sound correspondences. Using statistical tests, we further evaluate the generality of the Swadesh 100-word list compared to the Swadesh 200-word list and other 100-word lists sampled randomly from the Swadesh 200-word list. All these provide mathematical support for applying lexicostatistics in historical and comparative linguistics. PMID:28018261
Gender and Age Related Effects While Watching TV Advertisements: An EEG Study.
Cartocci, Giulia; Cherubino, Patrizia; Rossi, Dario; Modica, Enrica; Maglione, Anton Giulio; di Flumeri, Gianluca; Babiloni, Fabio
2016-01-01
The aim of the present paper is to show how the variation of the EEG frontal cortical asymmetry is related to the general appreciation perceived during the observation of TV advertisements, in particular considering the influence of the gender and age on it. In particular, we investigated the influence of the gender on the perception of a car advertisement (Experiment 1) and the influence of the factor age on a chewing gum commercial (Experiment 2). Experiment 1 results showed statistically significant higher approach values for the men group throughout the commercial. Results from Experiment 2 showed significant lower values by older adults for the spot, containing scenes not very enjoyed by them. In both studies, there was no statistical significant difference in the scene relative to the product offering between the experimental populations, suggesting the absence in our study of a bias towards the specific product in the evaluated populations. These evidences state the importance of the creativity in advertising, in order to attract the target population.
Gender and Age Related Effects While Watching TV Advertisements: An EEG Study
Cartocci, Giulia; Cherubino, Patrizia; Rossi, Dario; Modica, Enrica; Maglione, Anton Giulio; di Flumeri, Gianluca; Babiloni, Fabio
2016-01-01
The aim of the present paper is to show how the variation of the EEG frontal cortical asymmetry is related to the general appreciation perceived during the observation of TV advertisements, in particular considering the influence of the gender and age on it. In particular, we investigated the influence of the gender on the perception of a car advertisement (Experiment 1) and the influence of the factor age on a chewing gum commercial (Experiment 2). Experiment 1 results showed statistically significant higher approach values for the men group throughout the commercial. Results from Experiment 2 showed significant lower values by older adults for the spot, containing scenes not very enjoyed by them. In both studies, there was no statistical significant difference in the scene relative to the product offering between the experimental populations, suggesting the absence in our study of a bias towards the specific product in the evaluated populations. These evidences state the importance of the creativity in advertising, in order to attract the target population. PMID:27313602
Brinton, Louise A; Cook, Michael B; McCormack, Valerie; Johnson, Kenneth C; Olsson, Håkan; Casagrande, John T; Cooke, Rosie; Falk, Roni T; Gapstur, Susan M; Gaudet, Mia M; Gaziano, J Michael; Gkiokas, Georgios; Guénel, Pascal; Henderson, Brian E; Hollenbeck, Albert; Hsing, Ann W; Kolonel, Laurence N; Isaacs, Claudine; Lubin, Jay H; Michels, Karin B; Negri, Eva; Parisi, Dominick; Petridou, Eleni Th; Pike, Malcolm C; Riboli, Elio; Sesso, Howard D; Snyder, Kirk; Swerdlow, Anthony J; Trichopoulos, Dimitrios; Ursin, Giske; van den Brandt, Piet A; Van Den Eeden, Stephen K; Weiderpass, Elisabete; Willett, Walter C; Ewertz, Marianne; Thomas, David B
2014-03-01
The etiology of male breast cancer is poorly understood, partly because of its relative rarity. Although genetic factors are involved, less is known regarding the role of anthropometric and hormonally related risk factors. In the Male Breast Cancer Pooling Project, a consortium of 11 case-control and 10 cohort investigations involving 2405 case patients (n = 1190 from case-control and n = 1215 from cohort studies) and 52013 control subjects, individual participant data were harmonized and pooled. Unconditional logistic regression generated study design-specific (case-control/cohort) odds ratios (ORs) and 95% confidence intervals (CIs), with exposure estimates combined using fixed effects meta-analysis. All statistical tests were two-sided. Risk was statistically significantly associated with weight (highest/lowest tertile: OR = 1.36; 95% CI = 1.18 to 1.57), height (OR = 1.18; 95% CI = 1.01 to 1.38), and body mass index (BMI; OR = 1.30; 95% CI = 1.12 to 1.51), with evidence that recent rather than distant BMI was the strongest predictor. Klinefelter syndrome (OR = 24.7; 95% CI = 8.94 to 68.4) and gynecomastia (OR = 9.78; 95% CI = 7.52 to 12.7) were also statistically significantly associated with risk, relations that were independent of BMI. Diabetes also emerged as an independent risk factor (OR = 1.19; 95% CI = 1.04 to 1.37). There were also suggestive relations with cryptorchidism (OR = 2.18; 95% CI = 0.96 to 4.94) and orchitis (OR = 1.43; 95% CI = 1.02 to 1.99). Although age at onset of puberty and histories of infertility were unrelated to risk, never having had children was statistically significantly related (OR = 1.29; 95% CI = 1.01 to 1.66). Among individuals diagnosed at older ages, a history of fractures was statistically significantly related (OR = 1.41; 95% CI = 1.07 to 1.86). Consistent findings across case-control and cohort investigations, complemented by pooled analyses, indicated important roles for anthropometric and hormonal risk factors in the etiology of male breast cancer. Further investigation should focus on potential roles of endogenous hormones.
Sedgley, Norman; Elmslie, Bruce
2011-01-01
Evidence of the importance of urban agglomeration and the offsetting effects of congestion are provided in a number of studies of productivity and wages. Little attention has been paid to this evidence in the economic growth literature, where the recent focus is on technological change. We extend the idea of agglomeration and congestion effects to the area of innovation by empirically looking for a nonlinear link between population density and patent activity. A panel data set consisting of observations on 302 USA metropolitan statistical areas (MSAs) over a 10-year period from 1990 to 1999 is utilized. Following the patent and R&D literature, models that account for the discreet nature of the dependent variable are employed. Strong evidence is found that agglomeration and congestion are important in explaining the vast differences in patent rates across US cities. The most important reason cities continue to exist, given the dramatic drop in transportation costs for physical goods over the last century, is probably related to the forces of agglomeration as they apply to knowledge spillovers. Therefore, the empirical investigation proposed here is an important part of understanding the viability of urban areas in the future.
The validity of multiphase DNS initialized on the basis of single--point statistics
NASA Astrophysics Data System (ADS)
Subramaniam, Shankar
1999-11-01
A study of the point--process statistical representation of a spray reveals that single--point statistical information contained in the droplet distribution function (ddf) is related to a sequence of single surrogate--droplet pdf's, which are in general different from the physical single--droplet pdf's. The results of this study have important consequences for the initialization and evolution of direct numerical simulations (DNS) of multiphase flows, which are usually initialized on the basis of single--point statistics such as the average number density in physical space. If multiphase DNS are initialized in this way, this implies that even the initial representation contains certain implicit assumptions concerning the complete ensemble of realizations, which are invalid for general multiphase flows. Also the evolution of a DNS initialized in this manner is shown to be valid only if an as yet unproven commutation hypothesis holds true. Therefore, it is questionable to what extent DNS that are initialized in this manner constitute a direct simulation of the physical droplets.
Statistical modeling of optical attenuation measurements in continental fog conditions
NASA Astrophysics Data System (ADS)
Khan, Muhammad Saeed; Amin, Muhammad; Awan, Muhammad Saleem; Minhas, Abid Ali; Saleem, Jawad; Khan, Rahimdad
2017-03-01
Free-space optics is an innovative technology that uses atmosphere as a propagation medium to provide higher data rates. These links are heavily affected by atmospheric channel mainly because of fog and clouds that act to scatter and even block the modulated beam of light from reaching the receiver end, hence imposing severe attenuation. A comprehensive statistical study of the fog effects and deep physical understanding of the fog phenomena are very important for suggesting improvements (reliability and efficiency) in such communication systems. In this regard, 6-months real-time measured fog attenuation data are considered and statistically investigated. A detailed statistical analysis related to each fog event for that period is presented; the best probability density functions are selected on the basis of Akaike information criterion, while the estimates of unknown parameters are computed by maximum likelihood estimation technique. The results show that most fog attenuation events follow normal mixture distribution and some follow the Weibull distribution.
On statistical independence of a contingency matrix
NASA Astrophysics Data System (ADS)
Tsumoto, Shusaku; Hirano, Shoji
2005-03-01
A contingency table summarizes the conditional frequencies of two attributes and shows how these two attributes are dependent on each other with the information on a partition of universe generated by these attributes. Thus, this table can be viewed as a relation between two attributes with respect to information granularity. This paper focuses on several characteristics of linear and statistical independence in a contingency table from the viewpoint of granular computing, which shows that statistical independence in a contingency table is a special form of linear dependence. The discussions also show that when a contingency table is viewed as a matrix, called a contingency matrix, its rank is equal to 1.0. Thus, the degree of independence, rank plays a very important role in extracting a probabilistic model from a given contingency table. Furthermore, it is found that in some cases, partial rows or columns will satisfy the condition of statistical independence, which can be viewed as a solving process of Diophatine equations.
Infant Directed Speech Enhances Statistical Learning in Newborn Infants: An ERP Study
Teinonen, Tuomas; Tervaniemi, Mari; Huotilainen, Minna
2016-01-01
Statistical learning and the social contexts of language addressed to infants are hypothesized to play important roles in early language development. Previous behavioral work has found that the exaggerated prosodic contours of infant-directed speech (IDS) facilitate statistical learning in 8-month-old infants. Here we examined the neural processes involved in on-line statistical learning and investigated whether the use of IDS facilitates statistical learning in sleeping newborns. Event-related potentials (ERPs) were recorded while newborns were exposed to12 pseudo-words, six spoken with exaggerated pitch contours of IDS and six spoken without exaggerated pitch contours (ADS) in ten alternating blocks. We examined whether ERP amplitudes for syllable position within a pseudo-word (word-initial vs. word-medial vs. word-final, indicating statistical word learning) and speech register (ADS vs. IDS) would interact. The ADS and IDS registers elicited similar ERP patterns for syllable position in an early 0–100 ms component but elicited different ERP effects in both the polarity and topographical distribution at 200–400 ms and 450–650 ms. These results provide the first evidence that the exaggerated pitch contours of IDS result in differences in brain activity linked to on-line statistical learning in sleeping newborns. PMID:27617967
Mar, Raymond A; Mason, Malia F; Litvack, Aubrey
2012-03-01
Daydreaming appears to have a complex relationship with life satisfaction and happiness. Here we demonstrate that the facets of daydreaming that predict life satisfaction differ between men and women (Study 1; N=421), that the content of daydreams tends to be social others (Study 2; N=17,556), and that who we daydream about influences the relation between daydreaming and happiness variables like life satisfaction, loneliness, and perceived social support (Study 3; N=361). Specifically, daydreaming about people not close to us predicts more loneliness and less perceived social support, whereas daydreaming about close others predicts greater life satisfaction. Importantly, these patterns hold even when actual social network depth and breadth are statistically controlled, although these associations tend to be small in magnitude. Individual differences and the content of daydreams are thus important to consider when examining how happiness relates to spontaneous thoughts. Copyright © 2011 Elsevier Inc. All rights reserved.
The role of causal criteria in causal inferences: Bradford Hill's "aspects of association".
Ward, Andrew C
2009-06-17
As noted by Wesley Salmon and many others, causal concepts are ubiquitous in every branch of theoretical science, in the practical disciplines and in everyday life. In the theoretical and practical sciences especially, people often base claims about causal relations on applications of statistical methods to data. However, the source and type of data place important constraints on the choice of statistical methods as well as on the warrant attributed to the causal claims based on the use of such methods. For example, much of the data used by people interested in making causal claims come from non-experimental, observational studies in which random allocations to treatment and control groups are not present. Thus, one of the most important problems in the social and health sciences concerns making justified causal inferences using non-experimental, observational data. In this paper, I examine one method of justifying such inferences that is especially widespread in epidemiology and the health sciences generally - the use of causal criteria. I argue that while the use of causal criteria is not appropriate for either deductive or inductive inferences, they do have an important role to play in inferences to the best explanation. As such, causal criteria, exemplified by what Bradford Hill referred to as "aspects of [statistical] associations", have an indispensible part to play in the goal of making justified causal claims.
El Dib, Regina; Tikkinen, Kari A O; Akl, Elie A; Gomaa, Huda A; Mustafa, Reem A; Agarwal, Arnav; Carpenter, Christopher R; Zhang, Yuchen; Jorge, Eliane C; Almeida, Ricardo A M B; do Nascimento Junior, Paulo; Doles, Joao Vitor P; Mustafa, Ahmad A; Sadeghirad, Behnam; Lopes, Luciane C; Bergamaschi, Cristiane C; Suzumura, Erica A; Cardoso, Marília M A; Corrente, José Eduardo; Stone, Samuel B; Schunemann, Holger J; Guyatt, Gordon H
2017-04-01
To provide a perspective on the current practice of randomized clinical trials (RCTs) of diagnostic strategies focusing on patient-important outcomes. We conducted a comprehensive search of MEDLINE and included RCTs published in full-text reports that evaluated alternative diagnostic strategies. Of 56,912 unique citations, we sampled 7,500 and included 103 eligible RCTs, therefore suggesting that MEDLINE includes approximately 781 diagnostic RCTs. The 103 eligible trials reported on: mortality (n = 41; 39.8%); morbidities (n = 63; 61.2%); symptoms/quality of life/functional status (n = 14; 13.6%); and on composite end points (n = 10; 9.7%). Of the studies that reported statistically significant results (n = 12; 11.6%), we judged 7 (58.3%) as at low risk of bias with respect to missing outcome data and 4 (33.3%) as at low risk of bias regarding blinding. Of the 41 RCTs that reported on mortality, only one (2.4%) reported statistically significant results. Of 63 RCTs addressing morbidity outcomes, 11 (17.5%) reported statistically significant results, all of which reported relative effects of greater than 20%. RCTs of diagnostic tests are not uncommon, and sometimes suggest benefits on patient-important outcomes but often suffer from limitations in sample size and conduct. Copyright © 2017 Elsevier Inc. All rights reserved.
The role of causal criteria in causal inferences: Bradford Hill's "aspects of association"
Ward, Andrew C
2009-01-01
As noted by Wesley Salmon and many others, causal concepts are ubiquitous in every branch of theoretical science, in the practical disciplines and in everyday life. In the theoretical and practical sciences especially, people often base claims about causal relations on applications of statistical methods to data. However, the source and type of data place important constraints on the choice of statistical methods as well as on the warrant attributed to the causal claims based on the use of such methods. For example, much of the data used by people interested in making causal claims come from non-experimental, observational studies in which random allocations to treatment and control groups are not present. Thus, one of the most important problems in the social and health sciences concerns making justified causal inferences using non-experimental, observational data. In this paper, I examine one method of justifying such inferences that is especially widespread in epidemiology and the health sciences generally – the use of causal criteria. I argue that while the use of causal criteria is not appropriate for either deductive or inductive inferences, they do have an important role to play in inferences to the best explanation. As such, causal criteria, exemplified by what Bradford Hill referred to as "aspects of [statistical] associations", have an indispensible part to play in the goal of making justified causal claims. PMID:19534788
Development of a Notational Analysis System for Selected Soccer Skills of a Women's College Team
ERIC Educational Resources Information Center
Thomas, Camille; Fellingham, Gilbert; Vehrs, Pat
2009-01-01
The purposes of this study were to develop a notational system to evaluate passing, dribbling, first touch, and individual defensive skills as they relate to success during women's soccer games and to develop a statistical model to weigh the importance of each skill on creating scoring opportunities. Sequences of skills in ten games of a National…
ERIC Educational Resources Information Center
Fagginger Auer, Marije F.; Hickendorff, Marian; Van Putten, Cornelis M.; Béguin, Anton A.; Heiser, Willem J.
2016-01-01
A first application of multilevel latent class analysis (MLCA) to educational large-scale assessment data is demonstrated. This statistical technique addresses several of the challenges that assessment data offers. Importantly, MLCA allows modeling of the often ignored teacher effects and of the joint influence of teacher and student variables.…
ERIC Educational Resources Information Center
Ng, Siew Foen; Confessore, Gary J.
2015-01-01
Malaysia currently has about three million senior citizens. United Nations statistics show that Malaysia is likely to reach "aging nation" status by the year 2035. It is important to address the issues that may have impact on the needs and concerns of this growing population. This study examined the relationships of life satisfaction,…
The production and consumption of minor timber products in Oregon and Washington, 1930.
Herman M. Johnson
1932-01-01
A thorough canvass of the producers of minor timber products in Oregon and Washington has brought together or the first time authoritative statistics as to quantities of each currently produced. This information is valuable in gauging the importance of those industries, the trend in production and consumption, their relation to the major timber industries, and their...
ERIC Educational Resources Information Center
Al-Maamari, Faisal
2015-01-01
It is important to consider the question of whether teacher-, course-, and student-related factors affect student ratings of instructors in Student Evaluation of Teaching (SET) in English Language Teaching (ELT). This paper reports on a statistical analysis of SET in two large EFL programmes at a university setting in the Sultanate of Oman. I…
ERIC Educational Resources Information Center
Versace, Amelia; Ladouceur, Cecile D.; Romero, Soledad; Birmaher, Boris; Axelson, David A.; Kupfer, David J.; Phillips, Mary L.
2010-01-01
Objective: To study white matter (WM) development in youth at high familial risk for bipolar disorder (BD). WM alterations are reported in youth and adults with BD. WM undergoes important maturational changes in adolescence. Age-related changes in WM microstructure using diffusion tensor imaging with tract-based spatial statistics in healthy…
Murray, Christopher J L
2007-03-10
Health statistics are at the centre of an increasing number of worldwide health controversies. Several factors are sharpening the tension between the supply and demand for high quality health information, and the health-related Millennium Development Goals (MDGs) provide a high-profile example. With thousands of indicators recommended but few measured well, the worldwide health community needs to focus its efforts on improving measurement of a small set of priority areas. Priority indicators should be selected on the basis of public-health significance and several dimensions of measurability. Health statistics can be divided into three types: crude, corrected, and predicted. Health statistics are necessary inputs to planning and strategic decision making, programme implementation, monitoring progress towards targets, and assessment of what works and what does not. Crude statistics that are biased have no role in any of these steps; corrected statistics are preferred. For strategic decision making, when corrected statistics are unavailable, predicted statistics can play an important part. For monitoring progress towards agreed targets and assessment of what works and what does not, however, predicted statistics should not be used. Perhaps the most effective method to decrease controversy over health statistics and to encourage better primary data collection and the development of better analytical methods is a strong commitment to provision of an explicit data audit trail. This initiative would make available the primary data, all post-data collection adjustments, models including covariates used for farcasting and forecasting, and necessary documentation to the public.
Volberg, Rachel A; McNamara, Lauren M; Carris, Kari L
2018-06-01
While population surveys have been carried out in numerous jurisdictions internationally, little has been done to assess the relative strength of different risk factors that may contribute to the development of problem gambling. This is an important preparatory step for future research on the etiology of problem gambling. Using data from the 2006 California Problem Gambling Prevalence Survey, a telephone survey of adult California residents that used the NODS to assess respondents for gambling problems, binary logistic regression analysis was used to identify demographic characteristics, health-related behaviors, and gambling participation variables that statistically predicted the odds of being a problem or pathological gambler. In a separate approach, linear regression analysis was used to assess the impact of changes in these variables on the severity of the disorder. In both of the final models, the greatest statistical predictor of problem gambling status was past year Internet gambling. Furthermore, the unique finding of a significant interaction between physical or mental disability, Internet gambling, and problem gambling highlights the importance of exploring the interactions between different forms of gambling, the experience of mental and physical health issues, and the development of problem gambling using a longitudinal lens.
Relationships between digit ratio (2D:4D) and basketball performance in Australian men.
Frick, Nathan A; Hull, Melissa J; Manning, John T; Tomkinson, Grant R
2017-05-06
To investigate relationships between the digit ratio (2D:4D) and competitive basketball performance in Australian men. Using an observational cross-sectional design a total of 221 Australian basketball players who competed in the Olympic Games, International Basketball Federation World Championships/Cup, Australian National Basketball League, Central Australian Basketball League or socially had their 2D:4Ds measured. Analysis of variance was used to assess differences in mean 2D:4Ds between men playing at different competitive standards, with relationships between 2D:4Ds and basketball game-related statistics assessed using Pearson's product moment correlations in men playing at a single competitive standard. There were significant differences between competitive standards for the left 2D:4D following Bonferroni correction, but not for the right 2D:4D, with basketballers who achieved higher competitive standards tending to have lower left 2D:4Ds. No important correlations between 2D:4D and basketball game-related statistics were found, with correlations typically negligible. This study indicated that the 2D:4D can discriminate between basketballers competing at different standards, but not between basketballers within a single competitive standard using objective game-related statistics. © 2016 Wiley Periodicals, Inc.
ERIC Educational Resources Information Center
Sullivan, Sharon G.; Barr, Catherine; Grabois, Andrew
2002-01-01
Includes six articles that report on prices of U.S. and foreign published materials; book title output and average prices; book sales statistics; book exports and imports; book outlets in the U.S. and Canada; and review media statistics. (LRW)
Hilgers, Ralf-Dieter; Bogdan, Malgorzata; Burman, Carl-Fredrik; Dette, Holger; Karlsson, Mats; König, Franz; Male, Christoph; Mentré, France; Molenberghs, Geert; Senn, Stephen
2018-05-11
IDeAl (Integrated designs and analysis of small population clinical trials) is an EU funded project developing new statistical design and analysis methodologies for clinical trials in small population groups. Here we provide an overview of IDeAl findings and give recommendations to applied researchers. The description of the findings is broken down by the nine scientific IDeAl work packages and summarizes results from the project's more than 60 publications to date in peer reviewed journals. In addition, we applied text mining to evaluate the publications and the IDeAl work packages' output in relation to the design and analysis terms derived from in the IRDiRC task force report on small population clinical trials. The results are summarized, describing the developments from an applied viewpoint. The main result presented here are 33 practical recommendations drawn from the work, giving researchers a comprehensive guidance to the improved methodology. In particular, the findings will help design and analyse efficient clinical trials in rare diseases with limited number of patients available. We developed a network representation relating the hot topics developed by the IRDiRC task force on small population clinical trials to IDeAl's work as well as relating important methodologies by IDeAl's definition necessary to consider in design and analysis of small-population clinical trials. These network representation establish a new perspective on design and analysis of small-population clinical trials. IDeAl has provided a huge number of options to refine the statistical methodology for small-population clinical trials from various perspectives. A total of 33 recommendations developed and related to the work packages help the researcher to design small population clinical trial. The route to improvements is displayed in IDeAl-network representing important statistical methodological skills necessary to design and analysis of small-population clinical trials. The methods are ready for use.
Monitoring Method of Cow Anthrax Based on Gis and Spatial Statistical Analysis
NASA Astrophysics Data System (ADS)
Li, Lin; Yang, Yong; Wang, Hongbin; Dong, Jing; Zhao, Yujun; He, Jianbin; Fan, Honggang
Geographic information system (GIS) is a computer application system, which possesses the ability of manipulating spatial information and has been used in many fields related with the spatial information management. Many methods and models have been established for analyzing animal diseases distribution models and temporal-spatial transmission models. Great benefits have been gained from the application of GIS in animal disease epidemiology. GIS is now a very important tool in animal disease epidemiological research. Spatial analysis function of GIS can be widened and strengthened by using spatial statistical analysis, allowing for the deeper exploration, analysis, manipulation and interpretation of spatial pattern and spatial correlation of the animal disease. In this paper, we analyzed the cow anthrax spatial distribution characteristics in the target district A (due to the secret of epidemic data we call it district A) based on the established GIS of the cow anthrax in this district in combination of spatial statistical analysis and GIS. The Cow anthrax is biogeochemical disease, and its geographical distribution is related closely to the environmental factors of habitats and has some spatial characteristics, and therefore the correct analysis of the spatial distribution of anthrax cow for monitoring and the prevention and control of anthrax has a very important role. However, the application of classic statistical methods in some areas is very difficult because of the pastoral nomadic context. The high mobility of livestock and the lack of enough suitable sampling for the some of the difficulties in monitoring currently make it nearly impossible to apply rigorous random sampling methods. It is thus necessary to develop an alternative sampling method, which could overcome the lack of sampling and meet the requirements for randomness. The GIS computer application software ArcGIS9.1 was used to overcome the lack of data of sampling sites.Using ArcGIS 9.1 and GEODA to analyze the cow anthrax spatial distribution of district A. we gained some conclusions about cow anthrax' density: (1) there is a spatial clustering model. (2) there is an intensely spatial autocorrelation. We established a prediction model to estimate the anthrax distribution based on the spatial characteristic of the density of cow anthrax. Comparing with the true distribution, the prediction model has a well coincidence and is feasible to the application. The method using a GIS tool facilitates can be implemented significantly in the cow anthrax monitoring and investigation, and the space statistics - related prediction model provides a fundamental use for other study on space-related animal diseases.
NASA Technical Reports Server (NTRS)
Thomas, R. E.; Gaines, G. B.
1978-01-01
Recommended design procedures to reduce the complete factorial design by retaining information on anticipated important interaction effects, and by generally giving up information on unconditional main effects are discussed. A hypothetical photovoltaic module used in the test design is presented. Judgments were made of the relative importance of various environmental stresses such as UV radiation, abrasion, chemical attack, temperature, mechanical stress, relative humidity and voltage. Consideration is given to a complete factorial design and its graphical representation, elimination of selected test conditions, examination and improvement of an engineering design, and parametric study. The resulting design consists of a mix of conditional main effects and conditional interactions and represents a compromise between engineering and statistical requirements.
Özge, C; Toros, F; Bayramkaya, E; Çamdeviren, H; Şaşmaz, T
2006-01-01
Background The purpose of this study is to evaluate the most important sociodemographic factors on smoking status of high school students using a broad randomised epidemiological survey. Methods Using in‐class, self administered questionnaire about their sociodemographic variables and smoking behaviour, a representative sample of total 3304 students of preparatory, 9th, 10th, and 11th grades, from 22 randomly selected schools of Mersin, were evaluated and discriminative factors have been determined using appropriate statistics. In addition to binary logistic regression analysis, the study evaluated combined effects of these factors using classification and regression tree methodology, as a new statistical method. Results The data showed that 38% of the students reported lifetime smoking and 16.9% of them reported current smoking with a male predominancy and increasing prevalence by age. Second hand smoking was reported at a 74.3% frequency with father predominance (56.6%). The significantly important factors that affect current smoking in these age groups were increased by household size, late birth rank, certain school types, low academic performance, increased second hand smoking, and stress (especially reported as separation from a close friend or because of violence at home). Classification and regression tree methodology showed the importance of some neglected sociodemographic factors with a good classification capacity. Conclusions It was concluded that, as closely related with sociocultural factors, smoking was a common problem in this young population, generating important academic and social burden in youth life and with increasing data about this behaviour and using new statistical methods, effective coping strategies could be composed. PMID:16891446
Tiedeman, Claire; Ely, D. Matthew; Hill, Mary C.; O'Brien, Grady M.
2004-01-01
We develop a new observation‐prediction (OPR) statistic for evaluating the importance of system state observations to model predictions. The OPR statistic measures the change in prediction uncertainty produced when an observation is added to or removed from an existing monitoring network, and it can be used to guide refinement and enhancement of the network. Prediction uncertainty is approximated using a first‐order second‐moment method. We apply the OPR statistic to a model of the Death Valley regional groundwater flow system (DVRFS) to evaluate the importance of existing and potential hydraulic head observations to predicted advective transport paths in the saturated zone underlying Yucca Mountain and underground testing areas on the Nevada Test Site. Important existing observations tend to be far from the predicted paths, and many unimportant observations are in areas of high observation density. These results can be used to select locations at which increased observation accuracy would be beneficial and locations that could be removed from the network. Important potential observations are mostly in areas of high hydraulic gradient far from the paths. Results for both existing and potential observations are related to the flow system dynamics and coarse parameter zonation in the DVRFS model. If system properties in different locations are as similar as the zonation assumes, then the OPR results illustrate a data collection opportunity whereby observations in distant, high‐gradient areas can provide information about properties in flatter‐gradient areas near the paths. If this similarity is suspect, then the analysis produces a different type of data collection opportunity involving testing of model assumptions critical to the OPR results.
Wartberg, Lutz; Kriston, Levente; Kammerl, Rudolf
2017-07-01
Internet Gaming Disorder (IGD) has been included in the current edition of the Diagnostic and Statistical Manual of Mental Disorders-Fifth Edition (DSM-5). In the present study, the relationship among social support, friends only known through the Internet, health-related quality of life, and IGD in adolescence was explored for the first time. For this purpose, 1,095 adolescents aged from 12 to 14 years were surveyed with a standardized questionnaire concerning IGD, self-perceived social support, proportion of friends only known through the Internet, and health-related quality of life. The authors conducted unpaired t-tests, a chi-square test, as well as correlation and logistic regression analyses. According to the statistical analyses, adolescents with IGD reported lower self-perceived social support, more friends only known through the Internet, and a lower health-related quality of life compared with the group without IGD. Both in bivariate and multivariate logistic regression models, statistically significant associations between IGD and male gender, a higher proportion of friends only known through the Internet, and a lower health-related quality of life (multivariate model: Nagelkerke's R 2 = 0.37) were revealed. Lower self-perceived social support was related to IGD in the bivariate model only. In summary, quality of life and social aspects seem to be important factors for IGD in adolescence and therefore should be incorporated in further (longitudinal) studies. The findings of the present survey may provide starting points for the development of prevention and intervention programs for adolescents affected by IGD.
Chern, Yahn-Bor; Ho, Pei-Shan; Kuo, Li-Chueh; Chen, Jin-Bor
2013-01-01
Peritoneal dialysis (PD)-related peritonitis remains an important complication in PD patients, potentially causing technique failure and influencing patient outcome. To date, no comprehensive study in the Taiwanese PD population has used a time-dependent statistical method to analyze the factors associated with PD-related peritonitis. Our single-center retrospective cohort study, conducted in southern Taiwan between February 1999 and July 2010, used time-dependent statistical methods to analyze the factors associated with PD-related peritonitis. The study recruited 404 PD patients for analysis, 150 of whom experienced at least 1 episode of peritonitis during the follow-up period. The incidence rate of peritonitis was highest during the first 6 months after PD start. A comparison of patients in the two groups (peritonitis vs null-peritonitis) by univariate analysis showed that the peritonitis group included fewer men (p = 0.048) and more patients of older age (≥65 years, p = 0.049). In addition, patients who had never received compulsory education showed a statistically higher incidence of PD-related peritonitis in the univariate analysis (p = 0.04). A proportional hazards model identified education level (less than elementary school vs any higher education level) as having an independent association with PD-related peritonitis [hazard ratio (HR): 1.45; 95% confidence interval (CI): 1.01 to 2.06; p = 0.045). Comorbidities measured using the Charlson comorbidity index (score >2 vs ≤2) showed borderline statistical significance (HR: 1.44; 95% CI: 1.00 to 2.13; p = 0.053). A lower education level is a major risk factor for PD-related peritonitis independent of age, sex, hypoalbuminemia, and comorbidities. Our study emphasizes that a comprehensive PD education program is crucial for PD patients with a lower education level.
The Importance of Attitudes in Statistics Education
ERIC Educational Resources Information Center
Ramirez, Caroline; Schau, Candace; Emmioglu, Esma
2012-01-01
People forget what they do not use. But attitudes "stick." Our article emphasizes the importance of students' attitudes toward statistics. We examine 15 surveys that purport to assess these attitudes and then describe the "Survey of Attitudes Toward Statistics," a commonly used attitude survey. We present our conceptual model…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Slepian, Zachary; Slosar, Anze; Eisenstein, Daniel J.
We search for a galaxy clustering bias due to a modulation of galaxy number with the baryon-dark matter relative velocity resulting from recombination-era physics. We find no detected signal and place the constraint bv <0.01 on the relative velocity bias for the CMASS galaxies. This bias is an important potential systematic of Baryon Acoustic Oscillation (BAO) method measurements of the cosmic distance scale using the 2-point clustering. Our limit on the relative velocity bias indicates a systematic shift of no more than 0.3% rms in the distance scale inferred from the BAO feature in the BOSS 2-point clustering, well belowmore » the 1% statistical error of this measurement. In conclusion, this constraint is the most stringent currently available and has important implications for the ability of upcoming large-scale structure surveys such as DESI to self-protect against the relative velocity as a possible systematic.« less
Slepian, Zachary; Slosar, Anze; Eisenstein, Daniel J.; ...
2017-10-24
We search for a galaxy clustering bias due to a modulation of galaxy number with the baryon-dark matter relative velocity resulting from recombination-era physics. We find no detected signal and place the constraint bv <0.01 on the relative velocity bias for the CMASS galaxies. This bias is an important potential systematic of Baryon Acoustic Oscillation (BAO) method measurements of the cosmic distance scale using the 2-point clustering. Our limit on the relative velocity bias indicates a systematic shift of no more than 0.3% rms in the distance scale inferred from the BAO feature in the BOSS 2-point clustering, well belowmore » the 1% statistical error of this measurement. In conclusion, this constraint is the most stringent currently available and has important implications for the ability of upcoming large-scale structure surveys such as DESI to self-protect against the relative velocity as a possible systematic.« less
NASA Astrophysics Data System (ADS)
Slepian, Zachary; Eisenstein, Daniel J.; Blazek, Jonathan A.; Brownstein, Joel R.; Chuang, Chia-Hsun; Gil-Marín, Héctor; Ho, Shirley; Kitaura, Francisco-Shu; McEwen, Joseph E.; Percival, Will J.; Ross, Ashley J.; Rossi, Graziano; Seo, Hee-Jong; Slosar, Anže; Vargas-Magaña, Mariana
2018-02-01
We search for a galaxy clustering bias due to a modulation of galaxy number with the baryon-dark matter relative velocity resulting from recombination-era physics. We find no detected signal and place the constraint bv < 0.01 on the relative velocity bias for the CMASS galaxies. This bias is an important potential systematic of baryon acoustic oscillation (BAO) method measurements of the cosmic distance scale using the two-point clustering. Our limit on the relative velocity bias indicates a systematic shift of no more than 0.3 per cent rms in the distance scale inferred from the BAO feature in the BOSS two-point clustering, well below the 1 per cent statistical error of this measurement. This constraint is the most stringent currently available and has important implications for the ability of upcoming large-scale structure surveys such as the Dark Energy Spectroscopic Instrument (DESI) to self-protect against the relative velocity as a possible systematic.
NASA Astrophysics Data System (ADS)
Moschandreas, D. J.; Vuilleumier, K. L.
Prior to this study, indoor air constituent levels and ventilation rates of hospitality environments had not been measured simultaneously. This investigation measured indoor Environmental Tobacco Smoke-related (ETS-related) constituent levels in two restaurants, a billiard hall and a casino. The objective of this study was to characterize ETS-related constituent levels inside hospitality environments when the ventilation rates satisfy the requirements of the ASHRAE 62-1989 Ventilation Standard. The ventilation rate of each selected hospitality environment was measured and adjusted. The study advanced only if the requirements of the ASHRAE 62-1989 Ventilation Standard - the pertinent standard of the American Society of Heating, Refrigeration and Air Conditioning Engineers - were satisfied. The supply rates of outdoor air and occupant density were measured intermittently to assure that the ventilation rate of each facility satisfied the standard under occupied conditions. Six ETS-related constituents were measured: respirable suspended particulate (RSP) matter, fluorescent particulate matter (FPM, an estimate of the ETS particle concentrations), ultraviolet particulate matter (UVPM, a second estimate of the ETS particle concentrations), solanesol, nicotine and 3-ethenylpyridine (3-EP). ETS-related constituent levels in smoking sections, non-smoking sections and outdoors were sampled daily for eight consecutive days at each hospitality environment. This study found that the difference between the concentrations of ETS-related constituents in indoor smoking and non-smoking sections was statistically significant. Differences between indoor non-smoking sections and outdoor ETS-related constituent levels were identified but were not statistically significant. Similarly, differences between weekday and weekend evenings were identified but were not statistically significant. The difference between indoor smoking sections and outdoors was statistically significant. Most importantly, ETS-related constituent concentrations measured indoors did not exceed existing occupational standards. It was concluded that if the measured ventilation rates of the sampled facilities satisfied the ASHRAE 62-1989 Ventilation Standard requirements, the corresponding ETS-related constituents were measured at concentrations below known harmful levels as specified by the American Conference of Governmental Industrial Hygiene (ACGIH).
The More the Merrier?. Entropy and Statistics of Asexual Reproduction in Freshwater Planarians
NASA Astrophysics Data System (ADS)
Quinodoz, Sofia; Thomas, Michael A.; Dunkel, Jörn; Schötz, Eva-Maria
2011-04-01
The trade-off between traits in life-history strategies has been widely studied for sexual and parthenogenetic organisms, but relatively little is known about the reproduction strategies of asexual animals. Here, we investigate clonal reproduction in the freshwater planarian Schmidtea mediterranea, an important model organism for regeneration and stem cell research. We find that these flatworms adopt a randomized reproduction strategy that comprises both asymmetric binary fission and fragmentation (generation of multiple offspring during a reproduction cycle). Fragmentation in planarians has primarily been regarded as an abnormal behavior in the past; using a large-scale experimental approach, we now show that about one third of the reproduction events in S. mediterranea are fragmentations, implying that fragmentation is part of their normal reproductive behavior. Our analysis further suggests that certain characteristic aspects of the reproduction statistics can be explained in terms of a maximum relative entropy principle.
The impact of winter heating on air pollution in China.
Xiao, Qingyang; Ma, Zongwei; Li, Shenshen; Liu, Yang
2015-01-01
Fossil-fuel combustion related winter heating has become a major air quality and public health concern in northern China recently. We analyzed the impact of winter heating on aerosol loadings over China using the MODIS-Aqua Collection 6 aerosol product from 2004-2012. Absolute humidity (AH) and planetary boundary layer height (PBL) -adjusted aerosol optical depth (AOD*) was constructed to reflect ground-level PM2.5 concentrations. GIS analysis, standard statistical tests, and statistical modeling indicate that winter heating is an important factor causing increased PM2.5 levels in more than three-quarters of central and eastern China. The heating season AOD* was more than five times higher as the non-heating season AOD*, and the increase in AOD* in the heating areas was greater than in the non-heating areas. Finally, central heating tend to contribute less to air pollution relative to other means of household heating.
Statistical and methodological issues in the evaluation of case management studies.
Lesser, M L; Robertson, S; Kohn, N; Cooper, D J; Dlugacz, Y D
1996-01-01
For the past 3 years, the nursing case management team at North Shore University Hospital in Manhasset, NY, has been involved in a project to implement more than 50 clinical pathways, which provide a written "time line" for clinical events that should occur during a patient's hospital stay. A major objective of this project was to evaluate the efficacy of these pathways with respect to a number of important outcomes, such as length of stay, hospital costs, quality of patient care, and nursing and patient satisfaction. This article discusses several statistics-related issues in the design and evaluation of such case management studies. In particular, the role of a research approach in implementing and evaluating hospital programs, the choice of a comparison (control) group, the exclusion of selected patients from analysis, and the problems of equating pathways with diagnosis-related groups are addressed.
The Impact of Winter Heating on Air Pollution in China
Xiao, Qingyang; Ma, Zongwei; Li, Shenshen; Liu, Yang
2015-01-01
Fossil-fuel combustion related winter heating has become a major air quality and public health concern in northern China recently. We analyzed the impact of winter heating on aerosol loadings over China using the MODIS-Aqua Collection 6 aerosol product from 2004–2012. Absolute humidity (AH) and planetary boundary layer height (PBL) -adjusted aerosol optical depth (AOD*) was constructed to reflect ground-level PM2.5 concentrations. GIS analysis, standard statistical tests, and statistical modeling indicate that winter heating is an important factor causing increased PM2.5 levels in more than three-quarters of central and eastern China. The heating season AOD* was more than five times higher as the non-heating season AOD*, and the increase in AOD* in the heating areas was greater than in the non-heating areas. Finally, central heating tend to contribute less to air pollution relative to other means of household heating. PMID:25629878
Adult sports-related traumatic brain injury in United States trauma centers.
Winkler, Ethan A; Yue, John K; Burke, John F; Chan, Andrew K; Dhall, Sanjay S; Berger, Mitchel S; Manley, Geoffrey T; Tarapore, Phiroz E
2016-04-01
OBJECTIVE Sports-related traumatic brain injury (TBI) is an important public health concern estimated to affect 300,000 to 3.8 million people annually in the United States. Although injuries to professional athletes dominate the media, this group represents only a small proportion of the overall population. Here, the authors characterize the demographics of sports-related TBI in adults from a community-based trauma population and identify predictors of prolonged hospitalization and increased morbidity and mortality rates. METHODS Utilizing the National Sample Program of the National Trauma Data Bank (NTDB), the authors retrospectively analyzed sports-related TBI data from adults (age ≥ 18 years) across 5 sporting categories-fall or interpersonal contact (FIC), roller sports, skiing/snowboarding, equestrian sports, and aquatic sports. Multivariable regression analysis was used to identify predictors of prolonged hospital length of stay (LOS), medical complications, inpatient mortality rates, and hospital discharge disposition. Statistical significance was assessed at α < 0.05, and the Bonferroni correction for multiple comparisons was applied for each outcome analysis. RESULTS From 2003 to 2012, in total, 4788 adult sports-related TBIs were documented in the NTDB, which represented 18,310 incidents nationally. Equestrian sports were the greatest contributors to sports-related TBI (45.2%). Mild TBI represented nearly 86% of injuries overall. Mean (± SEM) LOSs in the hospital or intensive care unit (ICU) were 4.25 ± 0.09 days and 1.60 ± 0.06 days, respectively. The mortality rate was 3.0% across all patients, but was statistically higher in TBI from roller sports (4.1%) and aquatic sports (7.7%). Age, hypotension on admission to the emergency department (ED), and the severity of head and extracranial injuries were statistically significant predictors of prolonged hospital and ICU LOSs, medical complications, failure to discharge to home, and death. Traumatic brain injury during aquatic sports was similarly associated with prolonged ICU and hospital LOSs, medical complications, and failure to be discharged to home. CONCLUSIONS Age, hypotension on ED admission, severity of head and extracranial injuries, and sports mechanism of injury are important prognostic variables in adult sports-related TBI. Increasing TBI awareness and helmet use-particularly in equestrian and roller sports-are critical elements for decreasing sports-related TBI events in adults.
2016 Workplace and Gender Relations Survey of Active Duty Members: Statistical Methodology Report
2017-03-01
2016 Workplace and Gender Relations Survey of Active Duty Members Statistical Methodology Report Additional copies of this report may be...MEMBERS: STATISTICAL METHODOLOGY REPORT Office of People Analytics (OPA) Defense Research, Surveys, and Statistics Center 4800 Mark Center Drive...20 1 2016 WORKPLACE AND GENDER RELATIONS SURVEY OF ACTIVE DUTY MEMBERS: STATISTICAL METHODOLOGY REPORT
Halo models of HI selected galaxies
NASA Astrophysics Data System (ADS)
Paul, Niladri; Choudhury, Tirthankar Roy; Paranjape, Aseem
2018-06-01
Modelling the distribution of neutral hydrogen (HI) in dark matter halos is important for studying galaxy evolution in the cosmological context. We use a novel approach to infer the HI-dark matter connection at the massive end (m_H{I} > 10^{9.8} M_{⊙}) from radio HI emission surveys, using optical properties of low-redshift galaxies as an intermediary. In particular, we use a previously calibrated optical HOD describing the luminosity- and colour-dependent clustering of SDSS galaxies and describe the HI content using a statistical scaling relation between the optical properties and HI mass. This allows us to compute the abundance and clustering properties of HI-selected galaxies and compare with data from the ALFALFA survey. We apply an MCMC-based statistical analysis to constrain the free parameters related to the scaling relation. The resulting best-fit scaling relation identifies massive HI galaxies primarily with optically faint blue centrals, consistent with expectations from galaxy formation models. We compare the Hi-stellar mass relation predicted by our model with independent observations from matched Hi-optical galaxy samples, finding reasonable agreement. As a further application, we make some preliminary forecasts for future observations of HI and optical galaxies in the expected overlap volume of SKA and Euclid/LSST.
Nieri, Michele; Clauser, Carlo; Franceschi, Debora; Pagliaro, Umberto; Saletta, Daniele; Pini-Prato, Giovanpaolo
2007-08-01
The aim of the present study was to investigate the relationships among reported methodological, statistical, clinical and paratextual variables of randomized clinical trials (RCTs) in implant therapy, and their influence on subsequent research. The material consisted of the RCTs in implant therapy published through the end of the year 2000. Methodological, statistical, clinical and paratextual features of the articles were assessed and recorded. The perceived clinical relevance was subjectively evaluated by an experienced clinician on anonymous abstracts. The impact on research was measured by the number of citations found in the Science Citation Index. A new statistical technique (Structural learning of Bayesian Networks) was used to assess the relationships among the considered variables. Descriptive statistics revealed that the reported methodology and statistics of RCTs in implant therapy were defective. Follow-up of the studies was generally short. The perceived clinical relevance appeared to be associated with the objectives of the studies and with the number of published images in the original articles. The impact on research was related to the nationality of the involved institutions and to the number of published images. RCTs in implant therapy (until 2000) show important methodological and statistical flaws and may not be appropriate for guiding clinicians in their practice. The methodological and statistical quality of the studies did not appear to affect their impact on practice and research. Bayesian Networks suggest new and unexpected relationships among the methodological, statistical, clinical and paratextual features of RCTs.
Defining the ecological hydrology of Taiwan Rivers using multivariate statistical methods
NASA Astrophysics Data System (ADS)
Chang, Fi-John; Wu, Tzu-Ching; Tsai, Wen-Ping; Herricks, Edwin E.
2009-09-01
SummaryThe identification and verification of ecohydrologic flow indicators has found new support as the importance of ecological flow regimes is recognized in modern water resources management, particularly in river restoration and reservoir management. An ecohydrologic indicator system reflecting the unique characteristics of Taiwan's water resources and hydrology has been developed, the Taiwan ecohydrological indicator system (TEIS). A major challenge for the water resources community is using the TEIS to provide environmental flow rules that improve existing water resources management. This paper examines data from the extensive network of flow monitoring stations in Taiwan using TEIS statistics to define and refine environmental flow options in Taiwan. Multivariate statistical methods were used to examine TEIS statistics for 102 stations representing the geographic and land use diversity of Taiwan. The Pearson correlation coefficient showed high multicollinearity between the TEIS statistics. Watersheds were separated into upper and lower-watershed locations. An analysis of variance indicated significant differences between upstream, more natural, and downstream, more developed, locations in the same basin with hydrologic indicator redundancy in flow change and magnitude statistics. Issues of multicollinearity were examined using a Principal Component Analysis (PCA) with the first three components related to general flow and high/low flow statistics, frequency and time statistics, and quantity statistics. These principle components would explain about 85% of the total variation. A major conclusion is that managers must be aware of differences among basins, as well as differences within basins that will require careful selection of management procedures to achieve needed flow regimes.
Code of Federal Regulations, 2011 CFR
2011-01-01
... and related statistics by small certificated air carriers. 298.63 Section 298.63 Aeronautics and Space... aircraft operating expenses and related statistics by small certificated air carriers. (a) Each small... Related Statistics.” This schedule shall be filed quarterly as prescribed in § 298.60. Data reported on...
Code of Federal Regulations, 2010 CFR
2010-01-01
... and related statistics by small certificated air carriers. 298.63 Section 298.63 Aeronautics and Space... aircraft operating expenses and related statistics by small certificated air carriers. (a) Each small... Related Statistics.” This schedule shall be filed quarterly as prescribed in § 298.60. Data reported on...
The Importance of Teaching Power in Statistical Hypothesis Testing
ERIC Educational Resources Information Center
Olinsky, Alan; Schumacher, Phyllis; Quinn, John
2012-01-01
In this paper, we discuss the importance of teaching power considerations in statistical hypothesis testing. Statistical power analysis determines the ability of a study to detect a meaningful effect size, where the effect size is the difference between the hypothesized value of the population parameter under the null hypothesis and the true value…
The Importance of Statistical Modeling in Data Analysis and Inference
ERIC Educational Resources Information Center
Rollins, Derrick, Sr.
2017-01-01
Statistical inference simply means to draw a conclusion based on information that comes from data. Error bars are the most commonly used tool for data analysis and inference in chemical engineering data studies. This work demonstrates, using common types of data collection studies, the importance of specifying the statistical model for sound…
Artificial Intelligence Approach to Support Statistical Quality Control Teaching
ERIC Educational Resources Information Center
Reis, Marcelo Menezes; Paladini, Edson Pacheco; Khator, Suresh; Sommer, Willy Arno
2006-01-01
Statistical quality control--SQC (consisting of Statistical Process Control, Process Capability Studies, Acceptance Sampling and Design of Experiments) is a very important tool to obtain, maintain and improve the Quality level of goods and services produced by an organization. Despite its importance, and the fact that it is taught in technical and…
Using Data from Climate Science to Teach Introductory Statistics
ERIC Educational Resources Information Center
Witt, Gary
2013-01-01
This paper shows how the application of simple statistical methods can reveal to students important insights from climate data. While the popular press is filled with contradictory opinions about climate science, teachers can encourage students to use introductory-level statistics to analyze data for themselves on this important issue in public…
Young addicted men hormone profile detection
NASA Astrophysics Data System (ADS)
Zieliński, Paweł; Wasiewicz, Piotr; Leszczyńska, Bożena; Gromadzka-Ostrowska, Joanna
2010-09-01
Hormone parameters were determined in the serum of young addicted men in order to compare them with those obtained from the group of healthy subjects. Three groups were investigated which were named opiates, mixed and control group. Statistical and data mining methods were applied to obtain significant differences. R package was used for all computation. The determination of hormones parameters provide important information relative to impact of addiction.
Do Assimilated Drifter Velocities Improve Lagrangian Predictability in an Operational Ocean Model?
2015-05-01
extended Kalman filter . Molcard et al. (2005) used a statistical method to cor- relate model and drifter velocities. Taillandier et al. (2006) describe the... temperature and salinity observations. Trajectory angular differ- ences are also reduced. 1. Introduction The importance of Lagrangian forecasts was seen... Temperature , salinity, and sea surface height (SSH, measured along-track by satellite altimeters) observa- tions are typically assimilated in
The Black-White Difference in Youth Employment: Evidence for Demand-Side Factors.
ERIC Educational Resources Information Center
Cain, Glen G.; Finnie, Ross
The 1980 Census of the United States is used, first, to illustrate the serious lag in employment performance of young black men relative to young white men and, second, to test for the importance of demand-side causes of this lag. Aggregate data for 94 standard metropolitan statistical areas (SMSAs) contain data on the annual hours worked in 1979…
[How reliable is the monitoring for doping?].
Hüsler, J
1990-12-01
The reliability of the dope control, of the chemical analysis of the urine probes in the accredited laboratories and their decisions, is discussed using probabilistic and statistical methods. Basically, we evaluated and estimated the positive predictive value which means the probability that an urine probe contains prohibited dope substances given a positive test decision. Since there are not statistical data and evidence for some important quantities in relation to the predictive value, an exact evaluation is not possible, only conservative, lower bounds can be given. We found that the predictive value is at least 90% or 95% with respect to the analysis and decision based on the A-probe only, and at least 99% with respect to both A- and B-probes. A more realistic observation, but without sufficient statistical confidence, points to the fact that the true predictive value is significantly larger than these lower estimates.
Association of ED with chronic periodontal disease.
Matsumoto, S; Matsuda, M; Takekawa, M; Okada, M; Hashizume, K; Wada, N; Hori, J; Tamaki, G; Kita, M; Iwata, T; Kakizaki, H
2014-01-01
To examine the relationship between chronic periodontal disease (CPD) and ED, the interview sheet including the CPD self-checklist (CPD score) and the five-item version of the International Index of Erectile Function (IIEF-5) was distributed to 300 adult men who received a comprehensive dental examination. Statistical analyses were performed by the Spearman's rank correlation coefficient and other methods. Statistical significance was accepted at the level of P<0.05. The interview sheets were collected from 88 men (response rate 29.3%, 50.9±16.6 years old). There was a statistically significant correlation between the CPD score and the presence of ED (P=0.0415). The results in the present study suggest that ED is related to the damage caused by endothelial dysfunction and the systematic inflammatory changes associated with CPD. The present study also suggests that dental health is important as a preventive medicine for ED.
Application of multivariable statistical techniques in plant-wide WWTP control strategies analysis.
Flores, X; Comas, J; Roda, I R; Jiménez, L; Gernaey, K V
2007-01-01
The main objective of this paper is to present the application of selected multivariable statistical techniques in plant-wide wastewater treatment plant (WWTP) control strategies analysis. In this study, cluster analysis (CA), principal component analysis/factor analysis (PCA/FA) and discriminant analysis (DA) are applied to the evaluation matrix data set obtained by simulation of several control strategies applied to the plant-wide IWA Benchmark Simulation Model No 2 (BSM2). These techniques allow i) to determine natural groups or clusters of control strategies with a similar behaviour, ii) to find and interpret hidden, complex and casual relation features in the data set and iii) to identify important discriminant variables within the groups found by the cluster analysis. This study illustrates the usefulness of multivariable statistical techniques for both analysis and interpretation of the complex multicriteria data sets and allows an improved use of information for effective evaluation of control strategies.
Interactions between statistical and semantic information in infant language development
Lany, Jill; Saffran, Jenny R.
2013-01-01
Infants can use statistical regularities to form rudimentary word categories (e.g. noun, verb), and to learn the meanings common to words from those categories. Using an artificial language methodology, we probed the mechanisms by which two types of statistical cues (distributional and phonological regularities) affect word learning. Because linking distributional cues vs. phonological information to semantics make different computational demands on learners, we also tested whether their use is related to language proficiency. We found that 22-month-old infants with smaller vocabularies generalized using phonological cues; however, infants with larger vocabularies showed the opposite pattern of results, generalizing based on distributional cues. These findings suggest that both phonological and distributional cues marking word categories promote early word learning. Moreover, while correlations between these cues are important to forming word categories, we found infants’ weighting of these cues in subsequent word-learning tasks changes over the course of early language development. PMID:21884336
Carvajal, Thaddeus M; Viacrusis, Katherine M; Hernandez, Lara Fides T; Ho, Howell T; Amalin, Divina M; Watanabe, Kozo
2018-04-17
Several studies have applied ecological factors such as meteorological variables to develop models and accurately predict the temporal pattern of dengue incidence or occurrence. With the vast amount of studies that investigated this premise, the modeling approaches differ from each study and only use a single statistical technique. It raises the question of whether which technique would be robust and reliable. Hence, our study aims to compare the predictive accuracy of the temporal pattern of Dengue incidence in Metropolitan Manila as influenced by meteorological factors from four modeling techniques, (a) General Additive Modeling, (b) Seasonal Autoregressive Integrated Moving Average with exogenous variables (c) Random Forest and (d) Gradient Boosting. Dengue incidence and meteorological data (flood, precipitation, temperature, southern oscillation index, relative humidity, wind speed and direction) of Metropolitan Manila from January 1, 2009 - December 31, 2013 were obtained from respective government agencies. Two types of datasets were used in the analysis; observed meteorological factors (MF) and its corresponding delayed or lagged effect (LG). After which, these datasets were subjected to the four modeling techniques. The predictive accuracy and variable importance of each modeling technique were calculated and evaluated. Among the statistical modeling techniques, Random Forest showed the best predictive accuracy. Moreover, the delayed or lag effects of the meteorological variables was shown to be the best dataset to use for such purpose. Thus, the model of Random Forest with delayed meteorological effects (RF-LG) was deemed the best among all assessed models. Relative humidity was shown to be the top-most important meteorological factor in the best model. The study exhibited that there are indeed different predictive outcomes generated from each statistical modeling technique and it further revealed that the Random forest model with delayed meteorological effects to be the best in predicting the temporal pattern of Dengue incidence in Metropolitan Manila. It is also noteworthy that the study also identified relative humidity as an important meteorological factor along with rainfall and temperature that can influence this temporal pattern.
Sorrell, Jeanne M
2013-03-01
The fifth edition of the Diagnostic and Statistical Manual of Mental Disorders (DSM-5) is targeted for publication in May 2013. Older adults and their families should be aware of the potential impact that changes in this important document may have on diagnosis and treatment of mental health concerns. Two specific changes related to a new category of Neurocognitive Disorders and a new interpretation of criteria for depression after bereavement are discussed in this article. Nurses can help older adults and their families understand the new DSM-5 terminology and encourage them to discuss risks, benefits, and likely outcomes of diagnoses, procedures, and treatments that may seem unfamiliar. Copyright 2013, SLACK Incorporated.
Pereira, Tiago Veiga; Rudnicki, Martina; Pereira, Alexandre Costa; Pombo-de-Oliveira, Maria S; Franco, Rendrik França
2006-01-01
Meta-analysis has become an important statistical tool in genetic association studies, since it may provide more powerful and precise estimates. However, meta-analytic studies are prone to several potential biases not only because the preferential publication of "positive'' studies but also due to difficulties in obtaining all relevant information during the study selection process. In this letter, we point out major problems in meta-analysis that may lead to biased conclusions, illustrating an empirical example of two recent meta-analyses on the relation between MTHFR polymorphisms and risk of acute lymphoblastic leukemia that, despite the similarity in statistical methods and period of study selection, provided partially conflicting results.
Madhuri, Sonawane; Hegde, Shetiya Sahana; Ravi, Shirahatti; Deepti, Agarwal; Simpy, Mahuli
2014-07-01
The relationship between tooth loss and nutritional intake is important. As people age, their diminished physical capacity and decreased income adversely affect their ability to maintain their teeth. The aim of the study was to assess and compare the chewing ability, oral health related quality of life and nutritional status before and after fabrication and insertion of complete denture amongst edentulous participants in a dental college. Non Randomized Intervention study. The study population consisted of 42 participants (16 females and 26 males), aged 50 years and above. Prior to commencement of the study, informed consent was obtained and validation and reliability test of the questionnaire were done. The data for chewing ability, GOHAI and nutritional status assessment was recorded at baseline, 3(rd), 6(th) and 12(th) month after denture fabrication and insertion. The statistical comparisons were performed by repeated measure ANOVA and Chi-square test. P value<0.05 was considered as statistically significant. Chewing ability, GOHAI, BMI (Body Mass Index) and data from Food-intake questionnaire showed statistically significant improvement from baseline to 6(th) month but no statistically significant improvement was observed from 6(th) month to 12(th) month. Nutritive value of food (protein, energy and fat) showed no significant difference over a period of 12 months (p<0.05). Thus, it was concluded that the intervention (denture insertion) was effective in increasing the chewing ability, body weight, food-intake, and oral health related quality of life.
Hamel, Jean-Francois; Saulnier, Patrick; Pe, Madeline; Zikos, Efstathios; Musoro, Jammbe; Coens, Corneel; Bottomley, Andrew
2017-09-01
Over the last decades, Health-related Quality of Life (HRQoL) end-points have become an important outcome of the randomised controlled trials (RCTs). HRQoL methodology in RCTs has improved following international consensus recommendations. However, no international recommendations exist concerning the statistical analysis of such data. The aim of our study was to identify and characterise the quality of the statistical methods commonly used for analysing HRQoL data in cancer RCTs. Building on our recently published systematic review, we analysed a total of 33 published RCTs studying the HRQoL methods reported in RCTs since 1991. We focussed on the ability of the methods to deal with the three major problems commonly encountered when analysing HRQoL data: their multidimensional and longitudinal structure and the commonly high rate of missing data. All studies reported HRQoL being assessed repeatedly over time for a period ranging from 2 to 36 months. Missing data were common, with compliance rates ranging from 45% to 90%. From the 33 studies considered, 12 different statistical methods were identified. Twenty-nine studies analysed each of the questionnaire sub-dimensions without type I error adjustment. Thirteen studies repeated the HRQoL analysis at each assessment time again without type I error adjustment. Only 8 studies used methods suitable for repeated measurements. Our findings show a lack of consistency in statistical methods for analysing HRQoL data. Problems related to multiple comparisons were rarely considered leading to a high risk of false positive results. It is therefore critical that international recommendations for improving such statistical practices are developed. Copyright © 2017. Published by Elsevier Ltd.
Statistical representation of a spray as a point process
NASA Astrophysics Data System (ADS)
Subramaniam, S.
2000-10-01
The statistical representation of a spray as a finite point process is investigated. One objective is to develop a better understanding of how single-point statistical information contained in descriptions such as the droplet distribution function (ddf), relates to the probability density functions (pdfs) associated with the droplets themselves. Single-point statistical information contained in the droplet distribution function (ddf) is shown to be related to a sequence of single surrogate-droplet pdfs, which are in general different from the physical single-droplet pdfs. It is shown that the ddf contains less information than the fundamental single-point statistical representation of the spray, which is also described. The analysis shows which events associated with the ensemble of spray droplets can be characterized by the ddf, and which cannot. The implications of these findings for the ddf approach to spray modeling are discussed. The results of this study also have important consequences for the initialization and evolution of direct numerical simulations (DNS) of multiphase flows, which are usually initialized on the basis of single-point statistics such as the droplet number density in physical space. If multiphase DNS are initialized in this way, this implies that even the initial representation contains certain implicit assumptions concerning the complete ensemble of realizations, which are invalid for general multiphase flows. Also the evolution of a DNS initialized in this manner is shown to be valid only if an as yet unproven commutation hypothesis holds true. Therefore, it is questionable to what extent DNS that are initialized in this manner constitute a direct simulation of the physical droplets. Implications of these findings for large eddy simulations of multiphase flows are also discussed.
The ICF has made a difference to functioning and disability measurement and statistics.
Madden, Rosamond H; Bundy, Anita
2018-02-12
Fifteen years after the publication of the International Classification of Functioning, Disability and Health (ICF), we investigated: How ICF applications align with ICF aims, contents and principles, and how the ICF has been used to improve measurement of functioning and related statistics. In a scoping review, we investigated research published 2001-2015 relating to measurement and statistics for evidence of: a change in thinking; alignment of applications with ICF specifications and philosophy; and the emergence of new knowledge. The ICF is used in diverse applications, settings and countries, with processes largely aligned with the ICF and intended to improve measurement and statistics: new national surveys, information systems and ICF-based instruments; and international efforts to improve disability data. Knowledge is growing about the components and interactions of the ICF model, the diverse effects of the environment on functioning, and the meaning and measurement of participation. The ICF provides specificity and a common language in the complex world of functioning and disability and is stimulating new thinking, new applications in measurement and statistics, and the assembling of new knowledge. Nevertheless, the field needs to mature. Identified gaps suggest ways to improve measurement and statistics to underpin policies, services and outcomes. Implications for Rehabilitation The ICF offers a conceptualization of functioning and disability that can underpin assessment and documentation in rehabilitation, with a growing body of experience to draw on for guidance. Experience with the ICF reminds practitioners to consider all the domains of participation, the effect of the environment on participation and the importance of involving clients/patients in assessment and service planning. Understanding the variability of functioning within everyday environments and designing interventions for removing barriers in various environments is a vital part of rehabilitation planning.
From the necessary to the possible: the genesis of the spin-statistics theorem
NASA Astrophysics Data System (ADS)
Blum, Alexander
2014-12-01
The spin-statistics theorem, which relates the intrinsic angular momentum of a single particle to the type of quantum statistics obeyed by a system of many such particles, is one of the central theorems in quantum field theory and the physics of elementary particles. It was first formulated in 1939/40 by Wolfgang Pauli and his assistant Markus Fierz. This paper discusses the developments that led up to this first formulation, starting from early attempts in the late 1920s to explain why charged matter particles obey Fermi-Dirac statistics, while photons obey Bose-Einstein statistics. It is demonstrated how several important developments paved the way from such general philosophical musings to a general (and provable) theorem, most notably the use of quantum field theory, the discovery of new elementary particles, and the generalization of the notion of spin. It is also discussed how the attempts to prove a spin-statistics connection were driven by Pauli from formal to more physical arguments, culminating in Pauli's 1940 proof. This proof was a major success for the beleaguered theory of quantum field theory and the methods Pauli employed proved essential for the renaissance of quantum field theory and the development of renormalization techniques in the late 1940s.
Agić, Ante
2007-06-01
Knowledge of the foot morphometry is important for proper foot structure and function. Foot structure as a vital part of human body is important for many reasons. The foot anthropometric and morphology phenomena are analyzed together with hidden biomechanical descriptors in order to fully characterize foot functionality. For Croatian student population the scatter data of the individual foot variables were interpolated by multivariate statistics. Foot morphometric descriptors are influenced by many factors, such as life style, climate, and things of great importance in human society. Dominant descriptors related to fit and comfort are determined by the use 3D foot shape and advanced foot biomechanics. Some practical recommendations and conclusions for medical, sportswear and footwear practice are highlighted.
Neuroendocrine Tumor: Statistics
... Tumor > Neuroendocrine Tumor: Statistics Request Permissions Neuroendocrine Tumor: Statistics Approved by the Cancer.Net Editorial Board , 01/ ... the body. It is important to remember that statistics on the survival rates for people with a ...
Ecological statistics of Gestalt laws for the perceptual organization of contours.
Elder, James H; Goldberg, Richard M
2002-01-01
Although numerous studies have measured the strength of visual grouping cues for controlled psychophysical stimuli, little is known about the statistical utility of these various cues for natural images. In this study, we conducted experiments in which human participants trace perceived contours in natural images. These contours are automatically mapped to sequences of discrete tangent elements detected in the image. By examining relational properties between pairs of successive tangents on these traced curves, and between randomly selected pairs of tangents, we are able to estimate the likelihood distributions required to construct an optimal Bayesian model for contour grouping. We employed this novel methodology to investigate the inferential power of three classical Gestalt cues for contour grouping: proximity, good continuation, and luminance similarity. The study yielded a number of important results: (1) these cues, when appropriately defined, are approximately uncorrelated, suggesting a simple factorial model for statistical inference; (2) moderate image-to-image variation of the statistics indicates the utility of general probabilistic models for perceptual organization; (3) these cues differ greatly in their inferential power, proximity being by far the most powerful; and (4) statistical modeling of the proximity cue indicates a scale-invariant power law in close agreement with prior psychophysics.
The image recognition based on neural network and Bayesian decision
NASA Astrophysics Data System (ADS)
Wang, Chugege
2018-04-01
The artificial neural network began in 1940, which is an important part of artificial intelligence. At present, it has become a hot topic in the fields of neuroscience, computer science, brain science, mathematics, and psychology. Thomas Bayes firstly reported the Bayesian theory in 1763. After the development in the twentieth century, it has been widespread in all areas of statistics. In recent years, due to the solution of the problem of high-dimensional integral calculation, Bayesian Statistics has been improved theoretically, which solved many problems that cannot be solved by classical statistics and is also applied to the interdisciplinary fields. In this paper, the related concepts and principles of the artificial neural network are introduced. It also summarizes the basic content and principle of Bayesian Statistics, and combines the artificial neural network technology and Bayesian decision theory and implement them in all aspects of image recognition, such as enhanced face detection method based on neural network and Bayesian decision, as well as the image classification based on the Bayesian decision. It can be seen that the combination of artificial intelligence and statistical algorithms has always been the hot research topic.
A Virtual Study of Grid Resolution on Experiments of a Highly-Resolved Turbulent Plume
NASA Astrophysics Data System (ADS)
Maisto, Pietro M. F.; Marshall, Andre W.; Gollner, Michael J.; Fire Protection Engineering Department Collaboration
2017-11-01
An accurate representation of sub-grid scale turbulent mixing is critical for modeling fire plumes and smoke transport. In this study, PLIF and PIV diagnostics are used with the saltwater modeling technique to provide highly-resolved instantaneous field measurements in unconfined turbulent plumes useful for statistical analysis, physical insight, and model validation. The effect of resolution was investigated employing a virtual interrogation window (of varying size) applied to the high-resolution field measurements. Motivated by LES low-pass filtering concepts, the high-resolution experimental data in this study can be analyzed within the interrogation windows (i.e. statistics at the sub-grid scale) and on interrogation windows (i.e. statistics at the resolved scale). A dimensionless resolution threshold (L/D*) criterion was determined to achieve converged statistics on the filtered measurements. Such a criterion was then used to establish the relative importance between large and small-scale turbulence phenomena while investigating specific scales for the turbulent flow. First order data sets start to collapse at a resolution of 0.3D*, while for second and higher order statistical moments the interrogation window size drops down to 0.2D*.
Ohtana, Yuki; Abdullah, Azian Azamimi; Altaf-Ul-Amin, Md; Huang, Ming; Ono, Naoaki; Sato, Tetsuo; Sugiura, Tadao; Horai, Hisayuki; Nakamura, Yukiko; Morita Hirai, Aki; Lange, Klaus W; Kibinge, Nelson K; Katsuragi, Tetsuo; Shirai, Tsuyoshi; Kanaya, Shigehiko
2014-12-01
Developing database systems connecting diverse species based on omics is the most important theme in big data biology. To attain this purpose, we have developed KNApSAcK Family Databases, which are utilized in a number of researches in metabolomics. In the present study, we have developed a network-based approach to analyze relationships between 3D structure and biological activity of metabolites consisting of four steps as follows: construction of a network of metabolites based on structural similarity (Step 1), classification of metabolites into structure groups (Step 2), assessment of statistically significant relations between structure groups and biological activities (Step 3), and 2-dimensional clustering of the constructed data matrix based on statistically significant relations between structure groups and biological activities (Step 4). Applying this method to a data set consisting of 2072 secondary metabolites and 140 biological activities reported in KNApSAcK Metabolite Activity DB, we obtained 983 statistically significant structure group-biological activity pairs. As a whole, we systematically analyzed the relationship between 3D-chemical structures of metabolites and biological activities. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Suicidal behavior in the Ukraine, 1988-1998.
Kryzhanovskaya, L; Pilyagina, G
1999-01-01
This report studies the available data concerning suicide rates in the Ukraine and points to the importance of appropriate monitoring of suicides and attempted suicides. It illustrates the necessity of collecting this information and of developing "The Ukrainian National Program on Suicide Prevention." Unfortunately, suicide research and publications about suicide rates were prohibited in the former Soviet Union, so some of the data about suicidal behavior in the Ukraine is incomplete. We used the official suicide death statistics of the Ukraine from the Center of Statistics (Ukrainian Ministry of Health) for the period 1988-1998. The overall rate of suicide in the Ukraine is relatively high. Official statistics in the Ukraine show that there were 29.6 suicides per 100,000 population in 1998. The frequency of completed suicide differs in the various regions of the country, suicides being more frequent in the industrially developed regions and in the rural areas of the country than in the cities. In the western part of the Ukraine the frequency of suicide is relatively low (11.1 per 100,000). Between 1988 and 1997 the suicide rate increased by 57%. In 1998 the suicide rate for women was approximately five times lower than that for men.
Comparison of Housing Construction Development in Selected Regions of Central Europe
NASA Astrophysics Data System (ADS)
Dvorský, Ján; Petráková, Zora; Hollý, Ján
2017-12-01
In fast-growing countries, the economic growth, which came after the global financial crisis, ought to be manifested in the development of housing policy. The development of the region is directly related to the increase of the quality of living of its inhabitants. Housing construction and its relation with the availability of housing is a key issue for population overall. Comparison of its development in selected regions is important for experts in the field of construction, mayors of the regions, the state, but especially for the inhabitants themselves. The aim of the article is to compare the number of new dwellings with building permits and completed dwellings with final building approval between selected regions by using a mathematical statistics method - “Analysis of variance”. The article also uses the tools of descriptive statistics such as a point graph, a graph of deviations from the average, basic statistical characteristics of mean and variability. Qualitative factors influencing the construction of flats as well as the causes of quantitative differences in the number of started apartments under construction and completed apartments in selected regions of Central Europe are the subjects of the article’s conclusions.
Key statistical and analytical issues for evaluating treatment effects in periodontal research.
Tu, Yu-Kang; Gilthorpe, Mark S
2012-06-01
Statistics is an indispensible tool for evaluating treatment effects in clinical research. Due to the complexities of periodontal disease progression and data collection, statistical analyses for periodontal research have been a great challenge for both clinicians and statisticians. The aim of this article is to provide an overview of several basic, but important, statistical issues related to the evaluation of treatment effects and to clarify some common statistical misconceptions. Some of these issues are general, concerning many disciplines, and some are unique to periodontal research. We first discuss several statistical concepts that have sometimes been overlooked or misunderstood by periodontal researchers. For instance, decisions about whether to use the t-test or analysis of covariance, or whether to use parametric tests such as the t-test or its non-parametric counterpart, the Mann-Whitney U-test, have perplexed many periodontal researchers. We also describe more advanced methodological issues that have sometimes been overlooked by researchers. For instance, the phenomenon of regression to the mean is a fundamental issue to be considered when evaluating treatment effects, and collinearity amongst covariates is a conundrum that must be resolved when explaining and predicting treatment effects. Quick and easy solutions to these methodological and analytical issues are not always available in the literature, and careful statistical thinking is paramount when conducting useful and meaningful research. © 2012 John Wiley & Sons A/S.
ERIC Educational Resources Information Center
Mirman, Daniel; Estes, Katharine Graf; Magnuson, James S.
2010-01-01
Statistical learning mechanisms play an important role in theories of language acquisition and processing. Recurrent neural network models have provided important insights into how these mechanisms might operate. We examined whether such networks capture two key findings in human statistical learning. In Simulation 1, a simple recurrent network…
Relation of intelligence to ego functioning in an adult psychiatric population.
Allen, J G; Coyne, L; David, E
1986-01-01
Wechsler Adult Intelligence Scale-Revised (WAIS-R) IQs and clinical ratings of 10 ego functions in a diagnostically heterogeneous sample of 60 adult psychiatric inpatients were correlated. With severity of pathology statistically controlled, higher intelligence was associated with more adequate ego functioning in several spheres: primary autonomous functions, thought processes, object relations, and mastery-competence. There were also some clinically meaningful differences between the Verbal and Performance IQs in the pattern of correlations. Extending Hartmann's original views, the authors employ an ethological framework to conceptualize intelligence in relation to the ego's role in adaptation, emphasizing that intelligence is an important-albeit neglected-aspect of ego functioning.
Nourhashemi, Mina; Kongolo, Guy; Mahmoudzadeh, Mahdi; Goudjil, Sabrina; Wallois, Fabrice
2017-04-01
The mechanisms responsible for coupling between relative cerebral blood flow (rCBF), relative cerebral blood volume (rCBV), and relative cerebral metabolic rate of oxygen ([Formula: see text]), an important function of the microcirculation in preterm infants, remain unclear. Identification of a causal relationship between rCBF-rCBV and [Formula: see text] in preterms may, therefore, help to elucidate the principles of cortical hemodynamics during development. We simultaneously recorded rCBF and rCBV and estimated [Formula: see text] by two independent acquisition systems: diffuse correlation spectroscopy and near-infrared spectroscopy, respectively, in 10 preterms aged between 28 and 35 weeks of gestational age. Transfer entropy was calculated in order to determine the directionality between rCBF-rCBV and [Formula: see text]. The surrogate method was applied to determine statistical significance. The results show that rCBV and [Formula: see text] have a predominant driving influence on rCBF at the resting state in the preterm neonatal brain. Statistical analysis robustly detected the correct directionality of rCBV on rCBF and [Formula: see text] on rCBF. This study helps to clarify the early organization of the rCBV-rCBF and [Formula: see text] inter-relationship in the immature cortex.
Doherty, Kevin E.; Evans, Jeffrey S.; Coates, Peter S.; Juliusson, Lara; Fedy, Bradley C.
2016-01-01
We developed rangewide population and habitat models for Greater Sage-Grouse (Centrocercus urophasianus) that account for regional variation in habitat selection and relative densities of birds for use in conservation planning and risk assessments. We developed a probabilistic model of occupied breeding habitat by statistically linking habitat characteristics within 4 miles of an occupied lek using a nonlinear machine learning technique (Random Forests). Habitat characteristics used were quantified in GIS and represent standard abiotic and biotic variables related to sage-grouse biology. Statistical model fit was high (mean correctly classified = 82.0%, range = 75.4–88.0%) as were cross-validation statistics (mean = 80.9%, range = 75.1–85.8%). We also developed a spatially explicit model to quantify the relative density of breeding birds across each Greater Sage-Grouse management zone. The models demonstrate distinct clustering of relative abundance of sage-grouse populations across all management zones. On average, approximately half of the breeding population is predicted to be within 10% of the occupied range. We also found that 80% of sage-grouse populations were contained in 25–34% of the occupied range within each management zone. Our rangewide population and habitat models account for regional variation in habitat selection and the relative densities of birds, and thus, they can serve as a consistent and common currency to assess how sage-grouse habitat and populations overlap with conservation actions or threats over the entire sage-grouse range. We also quantified differences in functional habitat responses and disturbance thresholds across the Western Association of Fish and Wildlife Agencies (WAFWA) management zones using statistical relationships identified during habitat modeling. Even for a species as specialized as Greater Sage-Grouse, our results show that ecological context matters in both the strength of habitat selection (i.e., functional response curves) and response to disturbance.
Bachmann, Roger W; Hoyer, Mark V; Croteau, Amanda C; Canfield, Daniel E
2017-05-01
A probabilistic sample of lakes in the 48 coterminous US lakes was made by the United States Environmental Protection Agency in the 2007 National Lakes Assessment. Because of the statistical design, the results of our analyses of Secchi depths (SD) apply to a population of 45,265 lakes. We found statistically significant differences in mean Secchi depths between natural (1.57 m) and man-made lakes (1.18 m). The most important variable correlated with SD was turbidity, an optical measure related to suspended particles in the water column. For most lakes, chlorophyll a was highly correlated with both turbidity and SD, but several lakes had more turbidity and lower SD than expected based on chlorophyll a alone, indicating that non-algal suspended solids were an important factor. On an ecoregion basis, the non-algal suspended solids in the lake waters were related to the average levels of suspended solids in streams located in that ecoregion, and the non-algal suspended solids were more important in man-made than natural lakes. Phosphorus and nitrogen were directly correlated with chlorophyll a and turbidity and inversely correlated with SD. Based on diatom-inferred Secchi depths for the tops and bottoms of sediment cores from lakes in Ecoregions VIII and VII (excluding lakes in Minnesota) representing 40% of the natural lakes in the US, there has been no decrease in water transparency in that population of lakes in the past 70 or more years when the US population increased by 134%. We do not have information to determine if the other 60% of lakes have or have not changed.
The relative influence of nutrients and habitat on stream metabolism in agricultural streams
Frankforter, J.D.; Weyers, H.S.; Bales, J.D.; Moran, P.W.; Calhoun, D.L.
2010-01-01
Stream metabolism was measured in 33 streams across a gradient of nutrient concentrations in four agricultural areas of the USA to determine the relative influence of nutrient concentrations and habitat on primary production (GPP) and respiration (CR-24). In conjunction with the stream metabolism estimates, water quality and algal biomass samples were collected, as was an assessment of habitat in the sampling reach. When data for all study areas were combined, there were no statistically significant relations between gross primary production or community respiration and any of the independent variables. However, significant regression models were developed for three study areas for GPP (r 2 = 0.79-0.91) and CR-24 (r 2 = 0.76-0.77). Various forms of nutrients (total phosphorus and area-weighted total nitrogen loading) were significant for predicting GPP in two study areas, with habitat variables important in seven significant models. Important physical variables included light availability, precipitation, basin area, and in-stream habitat cover. Both benthic and seston chlorophyll were not found to be important explanatory variables in any of the models; however, benthic ash-free dry weight was important in two models for GPP. ?? 2009 The Author(s).
Yan, Zhengbing; Kuang, Te-Hui; Yao, Yuan
2017-09-01
In recent years, multivariate statistical monitoring of batch processes has become a popular research topic, wherein multivariate fault isolation is an important step aiming at the identification of the faulty variables contributing most to the detected process abnormality. Although contribution plots have been commonly used in statistical fault isolation, such methods suffer from the smearing effect between correlated variables. In particular, in batch process monitoring, the high autocorrelations and cross-correlations that exist in variable trajectories make the smearing effect unavoidable. To address such a problem, a variable selection-based fault isolation method is proposed in this research, which transforms the fault isolation problem into a variable selection problem in partial least squares discriminant analysis and solves it by calculating a sparse partial least squares model. As different from the traditional methods, the proposed method emphasizes the relative importance of each process variable. Such information may help process engineers in conducting root-cause diagnosis. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.
Biomarkers and surrogate endpoints in kidney disease.
Hartung, Erum A
2016-03-01
Kidney disease and its related comorbidities impose a large public health burden. Despite this, the number of clinical trials in nephrology lags behind many other fields. An important factor contributing to the relatively slow pace of nephrology trials is that existing clinical endpoints have significant limitations. "Hard" endpoints for chronic kidney disease, such as progression to end-stage renal disease, may not be reached for decades. Traditional biomarkers, such as serum creatinine in acute kidney injury, may lack sensitivity and predictive value. Finding new biomarkers to serve as surrogate endpoints is therefore an important priority in kidney disease research and may help to accelerate nephrology clinical trials. In this paper, I first review key concepts related to the selection of clinical trial endpoints and discuss statistical and regulatory considerations related to the evaluation of biomarkers as surrogate endpoints. This is followed by a discussion of the challenges and opportunities in developing novel biomarkers and surrogate endpoints in three major areas of nephrology research: acute kidney injury, chronic kidney disease, and autosomal dominant polycystic kidney disease.
The Sport Students’ Ability of Literacy and Statistical Reasoning
NASA Astrophysics Data System (ADS)
Hidayah, N.
2017-03-01
The ability of literacy and statistical reasoning is very important for the students of sport education college due to the materials of statistical learning can be taken from their many activities such as sport competition, the result of test and measurement, predicting achievement based on training, finding connection among variables, and others. This research tries to describe the sport education college students’ ability of literacy and statistical reasoning related to the identification of data type, probability, table interpretation, description and explanation by using bar or pie graphic, explanation of variability, interpretation, the calculation and explanation of mean, median, and mode through an instrument. This instrument is tested to 50 college students majoring in sport resulting only 26% of all students have the ability above 30% while others still below 30%. Observing from all subjects; 56% of students have the ability of identification data classification, 49% of students have the ability to read, display and interpret table through graphic, 27% students have the ability in probability, 33% students have the ability to describe variability, and 16.32% students have the ability to read, count and describe mean, median and mode. The result of this research shows that the sport students’ ability of literacy and statistical reasoning has not been adequate and students’ statistical study has not reached comprehending concept, literary ability trining and statistical rasoning, so it is critical to increase the sport students’ ability of literacy and statistical reasoning
Brehony, C; Cullinan, J; Cormican, M; Morris, D
2018-10-01
Shiga toxigenic Escherichia coli (STEC) are pathogenic E. coli that cause infectious diarrhoea. In some cases infection may be complicated by renal failure and death. The incidence of human infection with STEC in Ireland is the highest in Europe. The objective of the study was to examine the spatial incidence of human STEC infection in a region of Ireland with significantly higher rates of STEC incidence than the national average and to identify possible risk factors of STEC incidence at area level. Anonymised laboratory records (n = 379) from 2009 to 2015 were obtained from laboratories serving three counties in the West of Ireland. Data included location and sample date. Population and electoral division (ED) data were obtained from the Irish 2011 Census of Population. STEC incidence was calculated for each ED (n = 498) and used to map hotspots/coldspots using the Getis-Ord Gi* spatial statistic and significant spatial clustering using the Anselin's Local Moran's I statistic. Multivariable regression analysis was used to consider the importance of a number of potential predictors of STEC incidence. Incidence rates for the seven-year period ranged from 0 to 10.9 cases per 1000. A number of areas with significant local clustering of STEC incidence as well as variation in the spatial distribution of the two main serogroups associated with disease in the region i.e. O26 and O157 were identified. Cattle density was found to be a statistically significant predictor of STEC in the region. GIS analysis of routine data indicates that cattle density is associated STEC infection in this high incidence region. This finding points to the importance of agricultural practices for human health and the importance of a "one-health" approach to public policy in relation to agriculture, health and environment. Copyright © 2018 Elsevier B.V. All rights reserved.
CRACK GROWTH ANALYSIS OF SOLID OXIDE FUEL CELL ELECTROLYTES
DOE Office of Scientific and Technical Information (OSTI.GOV)
S. Bandopadhyay; N. Nagabhushana
2003-10-01
Defects and Flaws control the structural and functional property of ceramics. In determining the reliability and lifetime of ceramics structures it is very important to quantify the crack growth behavior of the ceramics. In addition, because of the high variability of the strength and the relatively low toughness of ceramics, a statistical design approach is necessary. The statistical nature of the strength of ceramics is currently well recognized, and is usually accounted for by utilizing Weibull or similar statistical distributions. Design tools such as CARES using a combination of strength measurements, stress analysis, and statistics are available and reasonably wellmore » developed. These design codes also incorporate material data such as elastic constants as well as flaw distributions and time-dependent properties. The fast fracture reliability for ceramics is often different from their time-dependent reliability. Further confounding the design complexity, the time-dependent reliability varies with the environment/temperature/stress combination. Therefore, it becomes important to be able to accurately determine the behavior of ceramics under simulated application conditions to provide a better prediction of the lifetime and reliability for a given component. In the present study, Yttria stabilized Zirconia (YSZ) of 9.6 mol% Yttria composition was procured in the form of tubes of length 100 mm. The composition is of interest as tubular electrolytes for Solid Oxide Fuel Cells. Rings cut from the tubes were characterized for microstructure, phase stability, mechanical strength (Weibull modulus) and fracture mechanisms. The strength at operating condition of SOFCs (1000 C) decreased to 95 MPa as compared to room temperature strength of 230 MPa. However, the Weibull modulus remains relatively unchanged. Slow crack growth (SCG) parameter, n = 17 evaluated at room temperature in air was representative of well studied brittle materials. Based on the results, further work was planned to evaluate the strength degradation, modulus and failure in more representative environment of the SOFCs.« less
History of water quality parameters - a study on the Sinos River/Brazil.
Konzen, G B; Figueiredo, J A S; Quevedo, D M
2015-05-01
Water is increasingly becoming a valuable resource, constituting one of the central themes of environmental, economic and social discussions. The Sinos River, located in southern Brazil, is the main river from the Sinos River Basin, representing a source of drinking water supply for a highly populated region. Considering its size and importance, it becomes necessary to conduct a study to follow up the water quality of this river, which is considered by some experts as one of the most polluted rivers in Brazil. As for this study, its great importance lies in the historical analysis of indicators. In this sense, we sought to develop aspects related to the management of water resources by performing a historical analysis of the Water Quality Index (WQI) of the Sinos River, using statistical methods. With regard to the methodological procedures, it should be pointed out that this study performs a time analysis of monitoring data on parameters related to a punctual measurement that is variable in time, using statistical tools. The data used refer to analyses of the water quality of the Sinos River (WQI) from the State Environmental Protection Agency Henrique Luiz Roessler (Fundação Estadual de Proteção Ambiental Henrique Luiz Roessler, FEPAM) covering the period between 2000 and 2008, as well as to a theoretical analysis focusing on the management of water resources. The study of WQI and its parameters by statistical analysis has shown to be effective, ensuring its effectiveness as a tool for the management of water resources. The descriptive analysis of the WQI and its parameters showed that the water quality of the Sinos River is concerning low, which reaffirms that it is one of the most polluted rivers in Brazil. It should be highlighted that there was an overall difficulty in obtaining data with the appropriate periodicity, as well as a long complete series, which limited the conduction of statistical studies such as the present one.
Investigation of Pre-Earthquake Ionospheric Disturbances by 3D Tomographic Analysis
NASA Astrophysics Data System (ADS)
Yagmur, M.
2016-12-01
Ionospheric variations before earthquakes have been widely discussed phenomena in ionospheric studies. To clarify the source and mechanism of these phenomena is highly important for earthquake forecasting. To well understanding the mechanical and physical processes of pre-seismic Ionospheric anomalies that might be related even with Lithosphere-Atmosphere-Ionosphere-Magnetosphere Coupling, both statistical and 3D modeling analysis are needed. For these purpose, firstly we have investigated the relation between Ionospheric TEC Anomalies and potential source mechanisms such as space weather activity and lithospheric phenomena like positive surface electric charges. To distinguish their effects on Ionospheric TEC, we have focused on pre-seismically active days. Then, we analyzed the statistical data of 54 earthquakes that M≽6 between 2000 and 2013 as well as the 2011 Tohoku and the 2016 Kumamoto Earthquakes in Japan. By comparing TEC anomaly and Solar activity by Dst Index, we have found that 28 events that might be related with Earthquake activity. Following the statistical analysis, we also investigate the Lithospheric effect on TEC change on selected days. Among those days, we have chosen two case studies as the 2011 Tohoku and the 2016 Kumamoto Earthquakes to make 3D reconstructed images by utilizing 3D Tomography technique with Neural Networks. The results will be presented in our presentation. Keywords : Earthquake, 3D Ionospheric Tomography, Positive and Negative Anomaly, Geomagnetic Storm, Lithosphere
Using expert informed GIS to locate important marine social-ecological hotspots.
Mahboubi, Pouyan; Parkes, Margot; Stephen, Craig; Chan, Hing Man
2015-09-01
The marine environment provides significant benefits to many local communities. Pressure to develop coastal waterways worldwide creates an urgent need for tools to locate marine spaces that have important social or ecological values, and to quantify their relative importance. The primary objective of this study was to develop, apply and critically assess a tool to identify important social-ecological hotspots in the marine environment. The study was conducted in a typical coastal community in northern British Columbia, Canada. This expert-informed GIS, or xGIS, tool used a survey instrument to draw on the knowledge of local experts from a range of backgrounds with respect to a series of 12 social-ecological value attributes, such as biodiversity, cultural and economic values. We identified approximately 1500 polygons on marine maps and assigned relative values to them using a token distribution exercise. A series of spatial statistical analyses were performed to locate and quantify the relative social-ecological importance of marine spaces and the results were ultimately summarized in a single hotspot map of the entire study area. This study demonstrates the utility of xGIS as a useful tool for stakeholders and environmental managers engaged in the planning and management of marine resources at the local and regional levels. Copyright © 2015 Elsevier Ltd. All rights reserved.
Code of Federal Regulations, 2011 CFR
2011-07-01
..., and Statistics Procedures Relating to the Implementation of the National Environmental Policy Act D... Assistance, Research, and Statistics Procedures Relating to the Implementation of the National Environmental... Statistics (OJARS) assists State and local units of government in strengthening and improving law enforcement...
Code of Federal Regulations, 2010 CFR
2010-07-01
..., and Statistics Procedures Relating to the Implementation of the National Environmental Policy Act D... Assistance, Research, and Statistics Procedures Relating to the Implementation of the National Environmental... Statistics (OJARS) assists State and local units of government in strengthening and improving law enforcement...
Millet, Juan Pablo; Garcia de Olalla, Patricia; Carrillo-Santisteve, Paloma; Gascón, Joaquim; Treviño, Begoña; Muñoz, José; Gómez I Prat, Jordi; Cabezos, Juan; González Cordón, Anna; Caylà, Joan A
2008-04-08
International travel and migration have been related with an increase of imported malaria cases. There has been considerable immigration to Barcelona from low-income countries (LIC) in recent years. The objective is to describe the epidemiology and to determine the trends of the disease in Barcelona. Analysis of the cases notified among city residents between 1989 and 2005. Patients were classified as: tourists, voluntary workers, resident immigrants (visiting friends and relatives, VFR) and recently arrived immigrants. An analysis was conducted using the chi2 test and comparison of means. As a measure of association we calculated the Relative Risk (RR) and Odds Ratio (OR) with a Confidence Interval of 95% (CI) and carried out a trends analysis. Of the total of 1,579 imported cases notified, 997 (63.1%) lived in Barcelona city, and 55.1% were male. The mean age of patients was 32.7 years. The incidence increased from 2.4 cases/100,000 in 1989 to 3.5 cases/100,000 in 2005 (RR 1.46 CI:1.36-1.55). This increase was not statistically significant (trends analysis, p = 0.36). In terms of reason for travelling, 40.7% were VFR, 33.6% tourists, 12.1% voluntary workers and 13.6% were recently arrived immigrants. The most frequent species found was Plasmodium falciparum (71.3%), mainly in visitors to Africa (OR = 2.3, CI = 1.7-3.2). The vast majority (82.2%) had had some contact with Africa (35.9% with Equatorial Guinea, a Spanish ex-colony) and 96.6% had not completed chemoprophylaxis. Six deaths were observed, all tourists who had travelled to Africa and not taken chemoprophylaxis (3.9% fatality rate). Over the period studied there is an increase in malaria incidence, however the trend is not statistically significant. Lack of chemoprophylaxis compliance and the association between Africa and P. falciparum are very clear in the imported cases. Most of the patients with malaria did not take chemoprophylaxis.
Non-statistical effects in bond fission reactions of 1,2-difluoroethane
NASA Astrophysics Data System (ADS)
Schranz, Harold W.; Raff, Lionel M.; Thompson, Donald L.
1991-08-01
A microcanonical, classical variational transition-state theory based on the use of the efficient microcanonical sampling (EMS) procedure is applied to simple bond fission in 1,2-difluoroethane. Comparison is made with results of trajectory calculations performed on the same global potential-energy surface. Agreement between the statistical theory and trajectory results for CC CF and CH bond fissions is poor with differences as large as a factor of 125. Most importantly, at the lower energy studied, 6.0 eV, the statistical calculations predict considerably slower rates than those computed from trajectories. We conclude from these results that the statistical assumptions inherent in the transition-state theory method are not valid for 1,2-difluoroethane in spite of the fact that the total intramolecular energy transfer rate out of CH and CC normal and local modes is large relative to the bond fission rates. The IVR rate is not globally rapid and the trajectories do not access all of the energetically available phase space uniformly on the timescale of the reactions.
Bourlier, Christophe
2005-07-10
The emissivity of two-dimensional anisotropic rough sea surfaces with non-Gaussian statistics is investigated. The emissivity derivation is of importance for retrieval of the sea-surface temperature or equivalent temperature of a rough sea surface by infrared thermal imaging. The well-known Cox-Munk slope probability-density function, considered non-Gaussian, is used for the emissivity derivation, in which the skewness and the kurtosis (related to the third- and fourth-order statistics, respectively) are included. The shadowing effect, which is significant for grazing angles, is also taken into account. The geometric optics approximation is assumed to be valid, which means that the rough surface is modeled as a collection of facets reflecting locally the light in the specular direction. In addition, multiple reflections are ignored. Numerical results of the emissivity are presented for Gaussian and non-Gaussian statistics, for moderate wind speeds, for near-infrared wavelengths, for emission angles ranging from 0 degrees (nadir) to 90 degrees (horizon), and according to the wind direction. In addition, the emissivity is compared with both measurements and a Monte Carlo ray-tracing method.
Statistical moments of the Strehl ratio
NASA Astrophysics Data System (ADS)
Yaitskova, Natalia; Esselborn, Michael; Gladysz, Szymon
2012-07-01
Knowledge of the statistical characteristics of the Strehl ratio is essential for the performance assessment of the existing and future adaptive optics systems. For full assessment not only the mean value of the Strehl ratio but also higher statistical moments are important. Variance is related to the stability of an image and skewness reflects the chance to have in a set of short exposure images more or less images with the quality exceeding the mean. Skewness is a central parameter in the domain of lucky imaging. We present a rigorous theory for the calculation of the mean value, the variance and the skewness of the Strehl ratio. In our approach we represent the residual wavefront as being formed by independent cells. The level of the adaptive optics correction defines the number of the cells and the variance of the cells, which are the two main parameters of our theory. The deliverables are the values of the three moments as the functions of the correction level. We make no further assumptions except for the statistical independence of the cells.
NASA Astrophysics Data System (ADS)
Jalali, Mohammad; Ramazi, Hamidreza
2018-06-01
Earthquake catalogues are the main source of statistical seismology for the long term studies of earthquake occurrence. Therefore, studying the spatiotemporal problems is important to reduce the related uncertainties in statistical seismology studies. A statistical tool, time normalization method, has been determined to revise time-frequency relationship in one of the most active regions of Asia, Eastern Iran and West of Afghanistan, (a and b were calculated around 8.84 and 1.99 in the exponential scale, not logarithmic scale). Geostatistical simulation method has been further utilized to reduce the uncertainties in the spatial domain. A geostatistical simulation produces a representative, synthetic catalogue with 5361 events to reduce spatial uncertainties. The synthetic database is classified using a Geographical Information System, GIS, based on simulated magnitudes to reveal the underlying seismicity patterns. Although some regions with highly seismicity correspond to known faults, significantly, as far as seismic patterns are concerned, the new method highlights possible locations of interest that have not been previously identified. It also reveals some previously unrecognized lineation and clusters in likely future strain release.
Statistical modeling of natural backgrounds in hyperspectral LWIR data
NASA Astrophysics Data System (ADS)
Truslow, Eric; Manolakis, Dimitris; Cooley, Thomas; Meola, Joseph
2016-09-01
Hyperspectral sensors operating in the long wave infrared (LWIR) have a wealth of applications including remote material identification and rare target detection. While statistical models for modeling surface reflectance in visible and near-infrared regimes have been well studied, models for the temperature and emissivity in the LWIR have not been rigorously investigated. In this paper, we investigate modeling hyperspectral LWIR data using a statistical mixture model for the emissivity and surface temperature. Statistical models for the surface parameters can be used to simulate surface radiances and at-sensor radiance which drives the variability of measured radiance and ultimately the performance of signal processing algorithms. Thus, having models that adequately capture data variation is extremely important for studying performance trades. The purpose of this paper is twofold. First, we study the validity of this model using real hyperspectral data, and compare the relative variability of hyperspectral data in the LWIR and visible and near-infrared (VNIR) regimes. Second, we illustrate how materials that are easily distinguished in the VNIR, may be difficult to separate when imaged in the LWIR.
ERIC Educational Resources Information Center
Yang, Christine K.
2011-01-01
Mathematics education has emerged to be of prime importance in the United States, as American students' performance has shown to be consistently and significantly lower than many other nations in the world (Rampey, Dion, & Donahue, 2009; National Center for Education Statistics, n.d.; OECD Programme for International Student Assessment, n.d.).…
Model of aircraft passenger acceptance
NASA Technical Reports Server (NTRS)
Jacobson, I. D.
1978-01-01
A technique developed to evaluate the passenger response to a transportation system environment is described. Reactions to motion, noise, temperature, seating, ventilation, sudden jolts and descents are modeled. Statistics are presented for the age, sex, occupation, and income distributions of the candidates analyzed. Values are noted for the relative importance of system variables such as time savings, on-time arrival, convenience, comfort, safety, the ability to read and write, and onboard services.
Fall 2014 SEI Research Review Probabilistic Analysis of Time Sensitive Systems
2014-10-28
Osmosis SMC Tool Osmosis is a tool for Statistical Model Checking (SMC) with Semantic Importance Sampling. • Input model is written in subset of C...ASSERT() statements in model indicate conditions that must hold. • Input probability distributions defined by the user. • Osmosis returns the...on: – Target relative error, or – Set number of simulations Osmosis Main Algorithm 1 http://dreal.cs.cmu.edu/ (?⃑?): Indicator
Statistics for wildlifers: how much and what kind?
Johnson, D.H.; Shaffer, T.L.; Newton, W.E.
2001-01-01
Quantitative methods are playing increasingly important roles in wildlife ecology and, ultimately, management. This change poses a challenge for wildlife practitioners and students who are not well-educated in mathematics and statistics. Here we give our opinions on what wildlife biologists should know about statistics, while recognizing that not everyone is inclined mathematically. For those who are, we recommend that they take mathematics coursework at least through calculus and linear algebra. They should take statistics courses that are focused conceptually , stressing the Why rather than the How of doing statistics. For less mathematically oriented wildlifers, introductory classes in statistical techniques will furnish some useful background in basic methods but may provide little appreciation of when the methods are appropriate. These wildlifers will have to rely much more on advice from statisticians. Far more important than knowing how to analyze data is an understanding of how to obtain and recognize good data. Regardless of the statistical education they receive, all wildlife biologists should appreciate the importance of controls, replication, and randomization in studies they conduct. Understanding these concepts requires little mathematical sophistication, but is critical to advancing the science of wildlife ecology.
Armijos, Rodrigo X.; Weigel, M. Margaret; Myers, Orrin B.; Li, Wen-Whai; Racines, Marcia; Berwick, Marianne
2015-01-01
Chronic exposure to urban traffic pollution is documented to promote atherosclerosis in adults but little is known about its potential effects in children. Our study examined the association of long-term exposure to traffic with carotid intima-media thickness (cIMT) in 287 healthy children. Residential proximity and distance-weighted traffic density (DWTD) were used as proximity markers for traffic-related air pollution exposure. The multivariable analyses revealed that children residing <100 meters from the nearest heavily trafficked road had cIMT mean and maximum measurements that were increased by 15% and 11% compared to those living ≥ 200 meters away (P = 0.0001). Similar increases in cIMT were identified for children in the highest versus lowest DWTD tertile. Children who resided 100–199 meters from traffic or in the middle DWTD tertile also exhibited increased cIMT but these differences were not statistically significant. No statistically significant differences were identified between residential distance to traffic or DWTD and systemic inflammation indicators (CRP, IL-6). The study results suggest that exposure to urban traffic promotes arterial remodeling in children. This finding is important since even small increases in cIMT over time can potentially lead to earlier progression to atherosclerosis. It is also important because traffic-related pollution is potentially modifiable. PMID:25685160
Armijos, Rodrigo X; Weigel, M Margaret; Myers, Orrin B; Li, Wen-Whai; Racines, Marcia; Berwick, Marianne
2015-01-01
Chronic exposure to urban traffic pollution is documented to promote atherosclerosis in adults but little is known about its potential effects in children. Our study examined the association of long-term exposure to traffic with carotid intima-media thickness (cIMT) in 287 healthy children. Residential proximity and distance-weighted traffic density (DWTD) were used as proximity markers for traffic-related air pollution exposure. The multivariable analyses revealed that children residing <100 meters from the nearest heavily trafficked road had cIMT mean and maximum measurements that were increased by 15% and 11% compared to those living ≥ 200 meters away (P = 0.0001). Similar increases in cIMT were identified for children in the highest versus lowest DWTD tertile. Children who resided 100-199 meters from traffic or in the middle DWTD tertile also exhibited increased cIMT but these differences were not statistically significant. No statistically significant differences were identified between residential distance to traffic or DWTD and systemic inflammation indicators (CRP, IL-6). The study results suggest that exposure to urban traffic promotes arterial remodeling in children. This finding is important since even small increases in cIMT over time can potentially lead to earlier progression to atherosclerosis. It is also important because traffic-related pollution is potentially modifiable.
Worry, Intolerance of Uncertainty, and Statistics Anxiety
ERIC Educational Resources Information Center
Williams, Amanda S.
2013-01-01
Statistics anxiety is a problem for most graduate students. This study investigates the relationship between intolerance of uncertainty, worry, and statistics anxiety. Intolerance of uncertainty was significantly related to worry, and worry was significantly related to three types of statistics anxiety. Six types of statistics anxiety were…
Wu, Robert; Glen, Peter; Ramsay, Tim; Martel, Guillaume
2014-06-28
Observational studies dominate the surgical literature. Statistical adjustment is an important strategy to account for confounders in observational studies. Research has shown that published articles are often poor in statistical quality, which may jeopardize their conclusions. The Statistical Analyses and Methods in the Published Literature (SAMPL) guidelines have been published to help establish standards for statistical reporting.This study will seek to determine whether the quality of statistical adjustment and the reporting of these methods are adequate in surgical observational studies. We hypothesize that incomplete reporting will be found in all surgical observational studies, and that the quality and reporting of these methods will be of lower quality in surgical journals when compared with medical journals. Finally, this work will seek to identify predictors of high-quality reporting. This work will examine the top five general surgical and medical journals, based on a 5-year impact factor (2007-2012). All observational studies investigating an intervention related to an essential component area of general surgery (defined by the American Board of Surgery), with an exposure, outcome, and comparator, will be included in this systematic review. Essential elements related to statistical reporting and quality were extracted from the SAMPL guidelines and include domains such as intent of analysis, primary analysis, multiple comparisons, numbers and descriptive statistics, association and correlation analyses, linear regression, logistic regression, Cox proportional hazard analysis, analysis of variance, survival analysis, propensity analysis, and independent and correlated analyses. Each article will be scored as a proportion based on fulfilling criteria in relevant analyses used in the study. A logistic regression model will be built to identify variables associated with high-quality reporting. A comparison will be made between the scores of surgical observational studies published in medical versus surgical journals. Secondary outcomes will pertain to individual domains of analysis. Sensitivity analyses will be conducted. This study will explore the reporting and quality of statistical analyses in surgical observational studies published in the most referenced surgical and medical journals in 2013 and examine whether variables (including the type of journal) can predict high-quality reporting.
NASA Astrophysics Data System (ADS)
Pipień, M.
2008-09-01
We present the results of an application of Bayesian inference in testing the relation between risk and return on the financial instruments. On the basis of the Intertemporal Capital Asset Pricing Model, proposed by Merton we built a general sampling distribution suitable in analysing this relationship. The most important feature of our assumptions is that the skewness of the conditional distribution of returns is used as an alternative source of relation between risk and return. This general specification relates to Skewed Generalized Autoregressive Conditionally Heteroscedastic-in-Mean model. In order to make conditional distribution of financial returns skewed we considered the unified approach based on the inverse probability integral transformation. In particular, we applied hidden truncation mechanism, inverse scale factors, order statistics concept, Beta and Bernstein distribution transformations and also a constructive method. Based on the daily excess returns on the Warsaw Stock Exchange Index we checked the empirical importance of the conditional skewness assumption on the relation between risk and return on the Warsaw Stock Market. We present posterior probabilities of all competing specifications as well as the posterior analysis of the positive sign of the tested relationship.
Palagini, Laura; Faraguna, Ugo; Mauri, Mauro; Gronchi, Alessia; Morin, Charles M; Riemann, Dieter
2016-03-01
Stress-related sleep reactivity, sleep-related cognitions, and psychological factors play an important role in insomnia. The aim was to investigate their possible association in Insomnia Disorder, insomnia subgroups, and healthy subjects. The cross-sectional study consisted of 93 subjects who met diagnostic criteria for Insomnia Disorder according to Diagnostic and Statistical Manual of Mental Disorders, 5th Edition (DSM-5) and of 30 healthy subjects. Survey instruments included the Insomnia Severity Index (ISI), Pittsburgh Sleep Quality Index (PSQI), Ford Insomnia Response to Stress Test (FIRST), Dysfunctional Beliefs about Sleep scale (DBAS), Beck Depression Inventory (BDI), and Zung Self-Rating Anxiety Scale (SAS). Descriptive statistics, Pearson correlations, χ(2)-test, and multiple linear regression were performed. FIRST and SAS best determined the insomnia subjects vs good sleepers (FIRST χ(2) = 109.6, p <0.001, SAS χ(2) = 120.3, p <0.001). FIRST was best predicted by DBAS (p <0.001), PSQI (p <0.001), and SAS by PSQI (p <0.001), ISI (p <0.05), BDI (p <0.001). In the sleep onset subgroup FIRST was related to ISI, PSQI, and DBAS and in the combined subgroup with DBAS. In both subgroups SAS was related to PSQI, ISI, and BDI. Findings suggest potential implications: (1) among the factors that may contribute to insomnia, stress-related sleep reactivity, and psychological factors, such as anxiety symptoms, may distinguish insomnia subjects from good sleepers; (2) sleep reactivity and sleep-related cognitions seem interrelated, unhelpful beliefs may affect the stress reactivity; (3) psychological factors may influence sleep quality and the severity of insomnia; (4) these important sleep-related variables may have similar associations in insomnia subgroups; they may constitute the core factors for insomnia development and maintenance. Copyright © 2015 Elsevier B.V. All rights reserved.
Nilsen, Charlotta; Andel, Ross; Fors, Stefan; Meinow, Bettina; Darin Mattsson, Alexander; Kåreholt, Ingemar
2014-08-27
People spend a considerable amount of time at work over the course of their lives, which makes the workplace important to health and aging. However, little is known about the potential long-term effects of work-related stress on late-life health. This study aims to examine work-related stress in late midlife and educational attainment in relation to serious health problems in old age. Data from nationally representative Swedish surveys were used in the analyses (n = 1,502). Follow-up time was 20-24 years. Logistic regressions were used to examine work-related stress (self-reported job demands, job control, and job strain) in relation to serious health problems measured as none, serious problems in one health domain, and serious problems in two or three health domains (complex health problems). While not all results were statistically significant, high job demands were associated with higher odds of serious health problems among women but lower odds of serious health problems among men. Job control was negatively associated with serious health problems. The strongest association in this study was between high job strain and complex health problems. After adjustment for educational attainment some of the associations became statistically nonsignificant. However, high job demands, remained related to lower odds of serious problems in one health domain among men, and low job control remained associated with higher odds of complex health problems among men. High job demands were associated with lower odds of complex health problems among men with low education, but not among men with high education, or among women regardless of level of education. The results underscore the importance of work-related stress for long-term health. Modification to work environment to reduce work stress (e.g., providing opportunities for self-direction/monitoring levels of psychological job demands) may serve as a springboard for the development of preventive strategies to improve public health both before and after retirement.
NASA Astrophysics Data System (ADS)
Ionita, M.; Grosfeld, K.; Scholz, P.; Lohmann, G.
2016-12-01
Sea ice in both Polar Regions is an important indicator for the expression of global climate change and its polar amplification. Consequently, a broad information interest exists on sea ice, its coverage, variability and long term change. Knowledge on sea ice requires high quality data on ice extent, thickness and its dynamics. However, its predictability depends on various climate parameters and conditions. In order to provide insights into the potential development of a monthly/seasonal signal, we developed a robust statistical model based on ocean heat content, sea surface temperature and atmospheric variables to calculate an estimate of the September minimum sea ice extent for every year. Although previous statistical attempts at monthly/seasonal forecasts of September sea ice minimum show a relatively reduced skill, here it is shown that more than 97% (r = 0.98) of the September sea ice extent can predicted three months in advance by using previous months conditions via a multiple linear regression model based on global sea surface temperature (SST), mean sea level pressure (SLP), air temperature at 850hPa (TT850), surface winds and sea ice extent persistence. The statistical model is based on the identification of regions with stable teleconnections between the predictors (climatological parameters) and the predictand (here sea ice extent). The results based on our statistical model contribute to the sea ice prediction network for the sea ice outlook report (https://www.arcus.org/sipn) and could provide a tool for identifying relevant regions and climate parameters that are important for the sea ice development in the Arctic and for detecting sensitive and critical regions in global coupled climate models with focus on sea ice formation.
Silver, Matt; Montana, Giovanni
2012-01-01
Where causal SNPs (single nucleotide polymorphisms) tend to accumulate within biological pathways, the incorporation of prior pathways information into a statistical model is expected to increase the power to detect true associations in a genetic association study. Most existing pathways-based methods rely on marginal SNP statistics and do not fully exploit the dependence patterns among SNPs within pathways. We use a sparse regression model, with SNPs grouped into pathways, to identify causal pathways associated with a quantitative trait. Notable features of our “pathways group lasso with adaptive weights” (P-GLAW) algorithm include the incorporation of all pathways in a single regression model, an adaptive pathway weighting procedure that accounts for factors biasing pathway selection, and the use of a bootstrap sampling procedure for the ranking of important pathways. P-GLAW takes account of the presence of overlapping pathways and uses a novel combination of techniques to optimise model estimation, making it fast to run, even on whole genome datasets. In a comparison study with an alternative pathways method based on univariate SNP statistics, our method demonstrates high sensitivity and specificity for the detection of important pathways, showing the greatest relative gains in performance where marginal SNP effect sizes are small. PMID:22499682
On the gas phase fragmentation of protonated uracil: a statistical perspective.
Rossich Molina, Estefanía; Salpin, Jean-Yves; Spezia, Riccardo; Martínez-Núñez, Emilio
2016-06-01
The potential energy surface of protonated uracil has been explored by an automated transition state search procedure, resulting in the finding of 1398 stationary points and 751 reactive channels, which can be categorized into isomerizations between pairs of isomers, unimolecular fragmentations and bimolecular reactions. The use of statistical Rice-Ramsperger-Kassel-Marcus (RRKM) theory and Kinetic Monte Carlo (KMC) simulations allowed us to determine the relative abundances of each fragmentation channel as a function of the ion's internal energy. The KMC/RRKM product abundances are compared with novel mass spectrometry (MS) experiments in the collision energy range 1-6 eV. To facilitate the comparison between theory and experiments, further dynamics simulations are carried out to determine the fraction of collision energy converted into the ion's internal energy. The KMC simulations show that the major fragmentation channels are isocyanic acid and ammonia losses, in good agreement with experiments. The third predominant channel is water loss according to both theory and experiments, although the abundance obtained in the KMC simulations is very low, suggesting that non-statistical dynamics might play an important role in this channel. Isocyanic acid (HNCOH(+)) is also an important product in the KMC simulations, although its abundance is only significant at internal energies not accessible in the MS experiments.
Mariani, Bette
2012-01-01
Mentoring is important in the career development of novice and experienced nurses. With the anticipated shortage in nursing, it is important to explore factors such as mentoring that may contribute to career satisfaction and intent to stay in the profession. This study explored the effects of mentoring on career satisfaction and intent to stay in nursing, and the relationship between career satisfaction and intent to stay in nursing. It was conducted through a mailed survey of RNs 55 years or younger currently in practice, education, administration, or research. Career satisfaction was measured through the use of the newly developed Mariani Nursing Career Satisfaction Scale. Findings revealed no statistically significant effect of mentoring on career satisfaction and intent to stay in nursing. There was a statistically significant relationship between career satisfaction and intent to stay in nursing. The majority of nurses reported participating in a mentoring relationship. Although the findings related to mentoring, career satisfaction, and intent to stay were not statistically significant, there was a prevalence of mentoring in nursing, thus suggesting the need for future research to identify outcomes of mentoring. In addition, the study contributed a newly developed instrument to measure the concept of career satisfaction in nursing.
Mariani, Bette
2012-01-01
Mentoring is important in the career development of novice and experienced nurses. With the anticipated shortage in nursing, it is important to explore factors such as mentoring that may contribute to career satisfaction and intent to stay in the profession. This study explored the effects of mentoring on career satisfaction and intent to stay in nursing, and the relationship between career satisfaction and intent to stay in nursing. It was conducted through a mailed survey of RNs 55 years or younger currently in practice, education, administration, or research. Career satisfaction was measured through the use of the newly developed Mariani Nursing Career Satisfaction Scale. Findings revealed no statistically significant effect of mentoring on career satisfaction and intent to stay in nursing. There was a statistically significant relationship between career satisfaction and intent to stay in nursing. The majority of nurses reported participating in a mentoring relationship. Although the findings related to mentoring, career satisfaction, and intent to stay were not statistically significant, there was a prevalence of mentoring in nursing, thus suggesting the need for future research to identify outcomes of mentoring. In addition, the study contributed a newly developed instrument to measure the concept of career satisfaction in nursing. PMID:22645673
Conroy, M.J.; Samuel, M.D.; White, Joanne C.
1995-01-01
Statistical power (and conversely, Type II error) is often ignored by biologists. Power is important to consider in the design of studies, to ensure that sufficient resources are allocated to address a hypothesis under examination. Deter- mining appropriate sample size when designing experiments or calculating power for a statistical test requires an investigator to consider the importance of making incorrect conclusions about the experimental hypothesis and the biological importance of the alternative hypothesis (or the biological effect size researchers are attempting to measure). Poorly designed studies frequently provide results that are at best equivocal, and do little to advance science or assist in decision making. Completed studies that fail to reject Ho should consider power and the related probability of a Type II error in the interpretation of results, particularly when implicit or explicit acceptance of Ho is used to support a biological hypothesis or management decision. Investigators must consider the biological question they wish to answer (Tacha et al. 1982) and assess power on the basis of biologically significant differences (Taylor and Gerrodette 1993). Power calculations are somewhat subjective, because the author must specify either f or the minimum difference that is biologically important. Biologists may have different ideas about what values are appropriate. While determining biological significance is of central importance in power analysis, it is also an issue of importance in wildlife science. Procedures, references, and computer software to compute power are accessible; therefore, authors should consider power. We welcome comments or suggestions on this subject.
Teaching Statistics in Integration with Psychology
ERIC Educational Resources Information Center
Wiberg, Marie
2009-01-01
The aim was to revise a statistics course in order to get the students motivated to learn statistics and to integrate statistics more throughout a psychology course. Further, we wish to make students become more interested in statistics and to help them see the importance of using statistics in psychology research. To achieve this goal, several…
Score As You Lift (SAYL): A Statistical Relational Learning Approach to Uplift Modeling.
Nassif, Houssam; Kuusisto, Finn; Burnside, Elizabeth S; Page, David; Shavlik, Jude; Costa, Vítor Santos
We introduce Score As You Lift (SAYL), a novel Statistical Relational Learning (SRL) algorithm, and apply it to an important task in the diagnosis of breast cancer. SAYL combines SRL with the marketing concept of uplift modeling, uses the area under the uplift curve to direct clause construction and final theory evaluation, integrates rule learning and probability assignment, and conditions the addition of each new theory rule to existing ones. Breast cancer, the most common type of cancer among women, is categorized into two subtypes: an earlier in situ stage where cancer cells are still confined, and a subsequent invasive stage. Currently older women with in situ cancer are treated to prevent cancer progression, regardless of the fact that treatment may generate undesirable side-effects, and the woman may die of other causes. Younger women tend to have more aggressive cancers, while older women tend to have more indolent tumors. Therefore older women whose in situ tumors show significant dissimilarity with in situ cancer in younger women are less likely to progress, and can thus be considered for watchful waiting. Motivated by this important problem, this work makes two main contributions. First, we present the first multi-relational uplift modeling system, and introduce, implement and evaluate a novel method to guide search in an SRL framework. Second, we compare our algorithm to previous approaches, and demonstrate that the system can indeed obtain differential rules of interest to an expert on real data, while significantly improving the data uplift.
Treated cabin acoustic prediction using statistical energy analysis
NASA Technical Reports Server (NTRS)
Yoerkie, Charles A.; Ingraham, Steven T.; Moore, James A.
1987-01-01
The application of statistical energy analysis (SEA) to the modeling and design of helicopter cabin interior noise control treatment is demonstrated. The information presented here is obtained from work sponsored at NASA Langley for the development of analytic modeling techniques and the basic understanding of cabin noise. Utility and executive interior models are developed directly from existing S-76 aircraft designs. The relative importance of panel transmission loss (TL), acoustic leakage, and absorption to the control of cabin noise is shown using the SEA modeling parameters. It is shown that the major cabin noise improvement below 1000 Hz comes from increased panel TL, while above 1000 Hz it comes from reduced acoustic leakage and increased absorption in the cabin and overhead cavities.
Garud, Nandita R; Rosenberg, Noah A
2015-06-01
Soft selective sweeps represent an important form of adaptation in which multiple haplotypes bearing adaptive alleles rise to high frequency. Most statistical methods for detecting selective sweeps from genetic polymorphism data, however, have focused on identifying hard selective sweeps in which a favored allele appears on a single haplotypic background; these methods might be underpowered to detect soft sweeps. Among exceptions is the set of haplotype homozygosity statistics introduced for the detection of soft sweeps by Garud et al. (2015). These statistics, examining frequencies of multiple haplotypes in relation to each other, include H12, a statistic designed to identify both hard and soft selective sweeps, and H2/H1, a statistic that conditional on high H12 values seeks to distinguish between hard and soft sweeps. A challenge in the use of H2/H1 is that its range depends on the associated value of H12, so that equal H2/H1 values might provide different levels of support for a soft sweep model at different values of H12. Here, we enhance the H12 and H2/H1 haplotype homozygosity statistics for selective sweep detection by deriving the upper bound on H2/H1 as a function of H12, thereby generating a statistic that normalizes H2/H1 to lie between 0 and 1. Through a reanalysis of resequencing data from inbred lines of Drosophila, we show that the enhanced statistic both strengthens interpretations obtained with the unnormalized statistic and leads to empirical insights that are less readily apparent without the normalization. Copyright © 2015 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Azila Che Musa, Nor; Mahmud, Zamalia; Baharun, Norhayati
2017-09-01
One of the important skills that is required from any student who are learning statistics is knowing how to solve statistical problems correctly using appropriate statistical methods. This will enable them to arrive at a conclusion and make a significant contribution and decision for the society. In this study, a group of 22 students majoring in statistics at UiTM Shah Alam were given problems relating to topics on testing of hypothesis which require them to solve the problems using confidence interval, traditional and p-value approach. Hypothesis testing is one of the techniques used in solving real problems and it is listed as one of the difficult concepts for students to grasp. The objectives of this study is to explore students’ perceived and actual ability in solving statistical problems and to determine which item in statistical problem solving that students find difficult to grasp. Students’ perceived and actual ability were measured based on the instruments developed from the respective topics. Rasch measurement tools such as Wright map and item measures for fit statistics were used to accomplish the objectives. Data were collected and analysed using Winsteps 3.90 software which is developed based on the Rasch measurement model. The results showed that students’ perceived themselves as moderately competent in solving the statistical problems using confidence interval and p-value approach even though their actual performance showed otherwise. Item measures for fit statistics also showed that the maximum estimated measures were found on two problems. These measures indicate that none of the students have attempted these problems correctly due to reasons which include their lack of understanding in confidence interval and probability values.
Methods for detrending success metrics to account for inflationary and deflationary factors*
NASA Astrophysics Data System (ADS)
Petersen, A. M.; Penner, O.; Stanley, H. E.
2011-01-01
Time-dependent economic, technological, and social factors can artificially inflate or deflate quantitative measures for career success. Here we develop and test a statistical method for normalizing career success metrics across time dependent factors. In particular, this method addresses the long standing question: how do we compare the career achievements of professional athletes from different historical eras? Developing an objective approach will be of particular importance over the next decade as major league baseball (MLB) players from the "steroids era" become eligible for Hall of Fame induction. Some experts are calling for asterisks (*) to be placed next to the career statistics of athletes found guilty of using performance enhancing drugs (PED). Here we address this issue, as well as the general problem of comparing statistics from distinct eras, by detrending the seasonal statistics of professional baseball players. We detrend player statistics by normalizing achievements to seasonal averages, which accounts for changes in relative player ability resulting from a range of factors. Our methods are general, and can be extended to various arenas of competition where time-dependent factors play a key role. For five statistical categories, we compare the probability density function (pdf) of detrended career statistics to the pdf of raw career statistics calculated for all player careers in the 90-year period 1920-2009. We find that the functional form of these pdfs is stationary under detrending. This stationarity implies that the statistical regularity observed in the right-skewed distributions for longevity and success in professional sports arises from both the wide range of intrinsic talent among athletes and the underlying nature of competition. We fit the pdfs for career success by the Gamma distribution in order to calculate objective benchmarks based on extreme statistics which can be used for the identification of extraordinary careers.
Stochastic modeling of sunshine number data
NASA Astrophysics Data System (ADS)
Brabec, Marek; Paulescu, Marius; Badescu, Viorel
2013-11-01
In this paper, we will present a unified statistical modeling framework for estimation and forecasting sunshine number (SSN) data. Sunshine number has been proposed earlier to describe sunshine time series in qualitative terms (Theor Appl Climatol 72 (2002) 127-136) and since then, it was shown to be useful not only for theoretical purposes but also for practical considerations, e.g. those related to the development of photovoltaic energy production. Statistical modeling and prediction of SSN as a binary time series has been challenging problem, however. Our statistical model for SSN time series is based on an underlying stochastic process formulation of Markov chain type. We will show how its transition probabilities can be efficiently estimated within logistic regression framework. In fact, our logistic Markovian model can be relatively easily fitted via maximum likelihood approach. This is optimal in many respects and it also enables us to use formalized statistical inference theory to obtain not only the point estimates of transition probabilities and their functions of interest, but also related uncertainties, as well as to test of various hypotheses of practical interest, etc. It is straightforward to deal with non-homogeneous transition probabilities in this framework. Very importantly from both physical and practical points of view, logistic Markov model class allows us to test hypotheses about how SSN dependents on various external covariates (e.g. elevation angle, solar time, etc.) and about details of the dynamic model (order and functional shape of the Markov kernel, etc.). Therefore, using generalized additive model approach (GAM), we can fit and compare models of various complexity which insist on keeping physical interpretation of the statistical model and its parts. After introducing the Markovian model and general approach for identification of its parameters, we will illustrate its use and performance on high resolution SSN data from the Solar Radiation Monitoring Station of the West University of Timisoara.
Stochastic modeling of sunshine number data
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brabec, Marek, E-mail: mbrabec@cs.cas.cz; Paulescu, Marius; Badescu, Viorel
2013-11-13
In this paper, we will present a unified statistical modeling framework for estimation and forecasting sunshine number (SSN) data. Sunshine number has been proposed earlier to describe sunshine time series in qualitative terms (Theor Appl Climatol 72 (2002) 127-136) and since then, it was shown to be useful not only for theoretical purposes but also for practical considerations, e.g. those related to the development of photovoltaic energy production. Statistical modeling and prediction of SSN as a binary time series has been challenging problem, however. Our statistical model for SSN time series is based on an underlying stochastic process formulation ofmore » Markov chain type. We will show how its transition probabilities can be efficiently estimated within logistic regression framework. In fact, our logistic Markovian model can be relatively easily fitted via maximum likelihood approach. This is optimal in many respects and it also enables us to use formalized statistical inference theory to obtain not only the point estimates of transition probabilities and their functions of interest, but also related uncertainties, as well as to test of various hypotheses of practical interest, etc. It is straightforward to deal with non-homogeneous transition probabilities in this framework. Very importantly from both physical and practical points of view, logistic Markov model class allows us to test hypotheses about how SSN dependents on various external covariates (e.g. elevation angle, solar time, etc.) and about details of the dynamic model (order and functional shape of the Markov kernel, etc.). Therefore, using generalized additive model approach (GAM), we can fit and compare models of various complexity which insist on keeping physical interpretation of the statistical model and its parts. After introducing the Markovian model and general approach for identification of its parameters, we will illustrate its use and performance on high resolution SSN data from the Solar Radiation Monitoring Station of the West University of Timisoara.« less
NASA Astrophysics Data System (ADS)
Castebrunet, H.; Eckert, N.; Giraud, G.; Durand, Y.; Morin, S.
2014-09-01
Projecting changes in snow cover due to climate warming is important for many societal issues, including the adaptation of avalanche risk mitigation strategies. Efficient modelling of future snow cover requires high resolution to properly resolve the topography. Here, we introduce results obtained through statistical downscaling techniques allowing simulations of future snowpack conditions including mechanical stability estimates for the mid and late 21st century in the French Alps under three climate change scenarios. Refined statistical descriptions of snowpack characteristics are provided in comparison to a 1960-1990 reference period, including latitudinal, altitudinal and seasonal gradients. These results are then used to feed a statistical model relating avalanche activity to snow and meteorological conditions, so as to produce the first projection on annual/seasonal timescales of future natural avalanche activity based on past observations. The resulting statistical indicators are fundamental for the mountain economy in terms of anticipation of changes. Whereas precipitation is expected to remain quite stationary, temperature increase interacting with topography will constrain the evolution of snow-related variables on all considered spatio-temporal scales and will, in particular, lead to a reduction of the dry snowpack and an increase of the wet snowpack. Overall, compared to the reference period, changes are strong for the end of the 21st century, but already significant for the mid century. Changes in winter are less important than in spring, but wet-snow conditions are projected to appear at high elevations earlier in the season. At the same altitude, the southern French Alps will not be significantly more affected than the northern French Alps, which means that the snowpack will be preserved for longer in the southern massifs which are higher on average. Regarding avalanche activity, a general decrease in mean (20-30%) and interannual variability is projected. These changes are relatively strong compared to changes in snow and meteorological variables. The decrease is amplified in spring and at low altitude. In contrast, an increase in avalanche activity is expected in winter at high altitude because of conditions favourable to wet-snow avalanches earlier in the season. Comparison with the outputs of the deterministic avalanche hazard model MEPRA (Modèle Expert d'aide à la Prévision du Risque d'Avalanche) shows generally consistent results but suggests that, even if the frequency of winters with high avalanche activity is clearly projected to decrease, the decreasing trend may be less strong and smooth than suggested by the statistical analysis based on changes in snowpack characteristics and their links to avalanches observations in the past. This important point for risk assessment pleads for further work focusing on shorter timescales. Finally, the small differences between different climate change scenarios show the robustness of the predicted avalanche activity changes.
Haebig, Eileen; Saffran, Jenny R; Ellis Weismer, Susan
2017-11-01
Word learning is an important component of language development that influences child outcomes across multiple domains. Despite the importance of word knowledge, word-learning mechanisms are poorly understood in children with specific language impairment (SLI) and children with autism spectrum disorder (ASD). This study examined underlying mechanisms of word learning, specifically, statistical learning and fast-mapping, in school-aged children with typical and atypical development. Statistical learning was assessed through a word segmentation task and fast-mapping was examined in an object-label association task. We also examined children's ability to map meaning onto newly segmented words in a third task that combined exposure to an artificial language and a fast-mapping task. Children with SLI had poorer performance on the word segmentation and fast-mapping tasks relative to the typically developing and ASD groups, who did not differ from one another. However, when children with SLI were exposed to an artificial language with phonemes used in the subsequent fast-mapping task, they successfully learned more words than in the isolated fast-mapping task. There was some evidence that word segmentation abilities are associated with word learning in school-aged children with typical development and ASD, but not SLI. Follow-up analyses also examined performance in children with ASD who did and did not have a language impairment. Children with ASD with language impairment evidenced intact statistical learning abilities, but subtle weaknesses in fast-mapping abilities. As the Procedural Deficit Hypothesis (PDH) predicts, children with SLI have impairments in statistical learning. However, children with SLI also have impairments in fast-mapping. Nonetheless, they are able to take advantage of additional phonological exposure to boost subsequent word-learning performance. In contrast to the PDH, children with ASD appear to have intact statistical learning, regardless of language status; however, fast-mapping abilities differ according to broader language skills. © 2017 Association for Child and Adolescent Mental Health.
AIDS susceptibility in a migrant population: perception and behavior.
McBride, D C; Weatherby, N L; Inciardi, J A; Gillespie, S A
1999-01-01
Within the framework of the Health Belief Model, this paper examines correlates of perception of AIDS susceptibility among 846 drug-using migrant farm workers and their sex partners. Significant but relatively small differences by ethnicity and gender were found. The data showed a consistent significant statistical relationship between frequency of drug use, high-risk sexual behavior, and perception of AIDS susceptibility. Perception of AIDS susceptibility was significantly related to a subsequent reduction in sexual risk behaviors. Consistent with the Health Belief Model, the data suggest that increasing perception of AIDS susceptibility may be an important motivator in reducing high-risk behaviors.
NASA Astrophysics Data System (ADS)
Marras, L.; Fontana, R.; Gambino, M. C.; Greco, M.; Materazzi, M.; Pampaloni, E.; Pezzati, L.; Poggi, P.
The knowledge of the shape of an artwork is an important element for its study and conservation. When dealing with a stone statue, roughness measurement is a very useful contribution to document its surface conditions, to assess either changes due to restoration intervention or surface decays due to weathering agents, and to monitor its time-evolution in terms of shape variations. In this work we present the preliminary results of the statistical analysis carried out on acquired data relative to six areas of the Michelangelo's David marble statue, representative of differently degraded surfaces. Determination of the roughness and its relative characteristic wavelength is shown.
On the game of life: population and its diversity
NASA Astrophysics Data System (ADS)
Sales, T. M.; Garcia, J. B. C.; Jyh, T. I.; Ren, T. I.; Gomes, M. A. F.
1993-08-01
One of the most important features of biological life in all levels is its astounding diversity. In this work we study the well-known game “Life” due to Conway analysing the statistics of cluster population, N( t), and cluster diversity, D( t). We have performed simulations on “Life” for dimensions d = 1 and 2 starting with an uncorrelated distribution of live and dead sites at t = 0. For d = 2 we study the effect of different neighbourhood relations in identifying and counting clusters. An interesting scaling relation connecting the maxima of N( t) and D( t) is found.
Engaging with the Art & Science of Statistics
ERIC Educational Resources Information Center
Peters, Susan A.
2010-01-01
How can statistics clearly be mathematical and yet distinct from mathematics? The answer lies in the reality that statistics is both an art and a science, and both aspects are important for teaching and learning statistics. Statistics is a mathematical science in that it applies mathematical theories and techniques. Mathematics provides the…
Mathematical background and attitudes toward statistics in a sample of Spanish college students.
Carmona, José; Martínez, Rafael J; Sánchez, Manuel
2005-08-01
To examine the relation of mathematical background and initial attitudes toward statistics of Spanish college students in social sciences the Survey of Attitudes Toward Statistics was given to 827 students. Multivariate analyses tested the effects of two indicators of mathematical background (amount of exposure and achievement in previous courses) on the four subscales. Analysis suggested grades in previous courses are more related to initial attitudes toward statistics than the number of mathematics courses taken. Mathematical background was related with students' affective responses to statistics but not with their valuing of statistics. Implications of possible research are discussed.
Sustainable energy policy: the impact of government subsidies on ethanol as a renewable fuel
NASA Astrophysics Data System (ADS)
Osuagwu, Denis Ahamarula
The United States Congress passed the Energy Policy Act of 1978 to promote ethanol production and reduce American dependence on foreign oil. The provision of subsidies in the act is indicative of the importance of energy in the economy. America needs a national energy policy that is economically, socially, and environmentally sustainable. Considering the importance of these needs, this study examines (a) the implementation of the Energy Policy Act of 1978 in regard to government subsidies and its effect on ethanol production, (b) the effect of gasoline consumption and cost on ethanol production, (c) the effect of corn production and price on ethanol fuel, and (d) the role of mandates and global crises on ethanol production. Secondary qualitative and quantitative data collected from various sources in 1978 through 2005 study the effect of ethanol subsidies on ethanol production. An autoregression error model is used to estimate the relevance of the explanatory variables on variations in ethanol production. The following are major study findings: (1) there is a positive correlation between corn production and ethanol production, is statistically significant; (2) government subsidies have a statistically significant positive correlation with ethanol production; (3) oil import has a statistically significant positive correlation with ethanol production, but has not contributed to a reduction the quantity of imported oil; (4) the price of corn has a statistically significant inverse relationship with ethanol production; (5) though not statistically significant, the price per barrel of oil is inversely related to ethanol production; (6) the budget surplus or deficit is associated with ethanol production; and (7) advocacy and lobbying for renewable fuel have encouraged support of ethanol production. The findings also show that global crises in the oil producing regions tend to influence the passage of favorable legislation for ethanol production. Furthermore, the incremental approach to implementation of alternative energy programs has been a failure. In the absence of a national energy regulation policy, oil import is on the increase; exacerbating American dependence on foreign supplies. A sustainable energy policy requires vision and commitment, but policymakers do not seem to command political capital to achieve this objective. Investigation reveals that subsidies have contributed in the production of ethanol. The four billion gallons of ethanol produced in 2005 is significantly higher than the quantity produced in 1978. However, an increase in ethanol production has made no considerable contribution to reducing American dependence on foreign oil. A sustainable energy policy requires a proactive public policy that includes public and private investment in renewable energy and technology, together with a continuance of local oil drilling.
Björnstig, Johanna; Bylund, Per-Olof; Björnstig, Ulf
2017-12-01
A data acquisition from the medical sector may give one important view of the burden on the society caused by vehicle related injuries. The official police-reported statistics may only reflect a part of all vehicle-related injured seeking medical attention. The aim is to provide a comprehensive picture of the burden of vehicle related injuries on the medical sector (2013), and to compare with official police-reported statistics and the development year 2000-2013. The data set includes 1085 injured from the Injury Data Base at Umeå University Hospital's catchment area with 148,500 inhabitants in 2013. Bicyclists were the most frequently injured (54%). One-third had non-minor (MAIS2+) injuries, and bicyclists accounted for 58% of the 1071 hospital bed days for all vehicle-related injuries. Car occupants represented 23% of all injured, and only 9% had MAIS2+ injuries. They accounted for 17% of the hospital bed days. Motorized two wheel vehicle riders represented 11% of the injured and 39% had MAIS2+ injuries and they occupied 11% of the hospital bed days. Of the 1085 medically treated persons, 767 were injured in public traffic areas, and, therefore, should be included in the official police statistics; however, only a third (232) of them were reported by the police. The annual injury rate had not changed during 2000-2013 for bicyclists, motor-cycle riders, pedestrians or snowmobile riders. However, for passenger car occupants a decrease was observed after 2008, and for mopedists the injury rate was halved after 2009 when a licensing regulation was introduced. The Swedish traffic injury reducing strategy Vision Zero, may have contributed to the reduction of injured car occupants and moped riders. The official police-reported statistics was a biased data source for vehicle related injuries and the total number medically treated was in total five times higher. Bicyclists caused the heaviest burden on the medical sector; consequently, they need to be prioritized in future safety work, as recently declared in the Government plan Vision Zero 2.0.
Determination of nursing students' attitudes towards the use of technology.
Terkes, Nurten; Celik, Ferya; Bektas, Hicran
2018-03-11
The use of technology is increasingly important in nursing education and practice. For this reason, it is necessary to determine the attitudes of nursing students towards technology. This study was conducted with 508 nursing students. A personal information form that was prepared by the researchers and the Attitudes Toward Technology Scale were used as the data collection tools. The mean score that was obtained by the nursing students from the Attitudes Toward Technology Scale was 61.53 ± 1.13. The Cronbach's alpha coefficient was found to be 0.90. There was a statistically significant difference between the sexes, using a computer, tablet, or laptop, using technology to reach health-related information, and for professional development, using mobile applications related to drug information. There was also a statistical difference between using the Periscope and Scorpio accounts from social media and using Excel and PowerPoint from Microsoft programs. Nursing students are capable of technology-based teaching, which can be expanded as a result. © 2018 Japan Academy of Nursing Science.
Complex Network Analysis for Characterizing Global Value Chains in Equipment Manufacturing.
Xiao, Hao; Sun, Tianyang; Meng, Bo; Cheng, Lihong
2017-01-01
The rise of global value chains (GVCs) characterized by the so-called "outsourcing", "fragmentation production", and "trade in tasks" has been considered one of the most important phenomena for the 21st century trade. GVCs also can play a decisive role in trade policy making. However, due to the increasing complexity and sophistication of international production networks, especially in the equipment manufacturing industry, conventional trade statistics and the corresponding trade indicators may give us a distorted picture of trade. This paper applies various network analysis tools to the new GVC accounting system proposed by Koopman et al. (2014) and Wang et al. (2013) in which gross exports can be decomposed into value-added terms through various routes along GVCs. This helps to divide the equipment manufacturing-related GVCs into some sub-networks with clear visualization. The empirical results of this paper significantly improve our understanding of the topology of equipment manufacturing-related GVCs as well as the interdependency of countries in these GVCs that is generally invisible from the traditional trade statistics.
2011-05-13
Violence-related firearm deaths remain an important public health concern in the United States. During 2006--2007, a total of 25,423 firearm homicides and 34,235 firearm suicides occurred among U.S. residents. These national totals include 4,166 firearm homicides and 1,446 firearm suicides among youths aged 10--19 years; the rate of firearm homicides among youths slightly exceeded the rate among persons of all ages. This report presents statistics on firearm homicides and firearm suicides for major metropolitan areas and cities, with an emphasis on youths aged 10--19 years in recognition of the importance of early prevention efforts. It integrates analyses conducted by CDC in response to requests for detailed information, arising from a heightened focus on urban violence by the media, the public, and policymakers over the past year. Firearm homicides and suicides and annual rates were tabulated for the 50 largest U.S. metropolitan statistical areas (MSAs) and their central cities for 2006--2007, using data from the National Vital Statistics System and the U.S. Census Bureau. Firearm homicide rates in approximately two thirds of the MSAs exceeded the national rate, and 86% of cities had rates higher than those of their MSAs. The youth firearm homicide rate exceeded the all-ages rate in 80% of the MSAs and in 88% of the cities. Firearm suicide rates in just over half of the MSAs were below the national rate, and 55% of cities had rates below those of their MSAs. Youth firearm suicide rates in the MSAs and cities were collectively low compared with all-ages rates. Such variations in firearm homicide and firearm suicide rates, with respect to both urbanization and age, should be considered in the continuing development of prevention programs directed at reducing firearm violence.
Spectral gene set enrichment (SGSE).
Frost, H Robert; Li, Zhigang; Moore, Jason H
2015-03-03
Gene set testing is typically performed in a supervised context to quantify the association between groups of genes and a clinical phenotype. In many cases, however, a gene set-based interpretation of genomic data is desired in the absence of a phenotype variable. Although methods exist for unsupervised gene set testing, they predominantly compute enrichment relative to clusters of the genomic variables with performance strongly dependent on the clustering algorithm and number of clusters. We propose a novel method, spectral gene set enrichment (SGSE), for unsupervised competitive testing of the association between gene sets and empirical data sources. SGSE first computes the statistical association between gene sets and principal components (PCs) using our principal component gene set enrichment (PCGSE) method. The overall statistical association between each gene set and the spectral structure of the data is then computed by combining the PC-level p-values using the weighted Z-method with weights set to the PC variance scaled by Tracy-Widom test p-values. Using simulated data, we show that the SGSE algorithm can accurately recover spectral features from noisy data. To illustrate the utility of our method on real data, we demonstrate the superior performance of the SGSE method relative to standard cluster-based techniques for testing the association between MSigDB gene sets and the variance structure of microarray gene expression data. Unsupervised gene set testing can provide important information about the biological signal held in high-dimensional genomic data sets. Because it uses the association between gene sets and samples PCs to generate a measure of unsupervised enrichment, the SGSE method is independent of cluster or network creation algorithms and, most importantly, is able to utilize the statistical significance of PC eigenvalues to ignore elements of the data most likely to represent noise.
Neuroimaging study of sex differences in the neuropathology of cocaine abuse.
Li, Chiang-shan Ray; Kemp, Kathleen; Milivojevic, Verica; Sinha, Rajita
2005-09-01
Female and male substance abusers differ in their disease patterns and clinical outcomes. An important question in addiction neuroscience thus concerns the neural substrates underlying these sex differences. This article aims to examine what is known of the neural mechanisms involved in the sex differences between substance abusers. We reviewed neuroimaging studies that addressed sex differences in cerebral perfusion deficits after chronic cocaine use and in regional brain activation during pharmacologic challenge and cue-induced craving. We also present results from a preliminary study in which cocaine-dependent men and women participated in script-guided imagery of stress- and drug cue-related situations while blood oxygenation level-dependent signals of their brain were acquired in a 1.5T scanner. Spatial pre-processing and statistical analysis of brain images were performed. Regional brain activation was compared between stress and drug cue trials in men versus women. The results of our study showed greater activation in the left uncus and right claustrum (both, statistical threshold of P = 0.01, uncorrected; extent = 10 voxels) in men (n = 5) during drug cue trials compared with stress trials. No brain regions showed greater activation during stress trials compared with drug cue trials. In contrast, women (n = 6) showed greater activation in the right medial and superior frontal gyri during stress trials compared with drug cue trials at the same statistical threshold. No brain regions showed more activation during drug cue trials than during stress trials. The studies reviewed underscore the need to consider sex-related factors in examining the neuropathology of cocaine addiction. Our preliminary results also suggest important sex differences in the effect of stress- and drug cue-associated brain activation in individuals with cocaine use disorder.
Sex differences in discriminative power of volleyball game-related statistics.
João, Paulo Vicente; Leite, Nuno; Mesquita, Isabel; Sampaio, Jaime
2010-12-01
To identify sex differences in volleyball game-related statistics, the game-related statistics of several World Championships in 2007 (N=132) were analyzed using the software VIS from the International Volleyball Federation. Discriminant analysis was used to identify the game-related statistics which better discriminated performances by sex. Analysis yielded an emphasis on fault serves (SC = -.40), shot spikes (SC = .40), and reception digs (SC = .31). Specific robust numbers represent that considerable variability was evident in the game-related statistics profile, as men's volleyball games were better associated with terminal actions (errors of service), and women's volleyball games were characterized by continuous actions (in defense and attack). These differences may be related to the anthropometric and physiological differences between women and men and their influence on performance profiles.
Davey, Cynthia S.; Caspi, Caitlin E.; Kubik, Martha Y.; Nanney, Marilyn S.
2016-01-01
Background The promotion of healthy eating and physical activity within school settings is an important component of population-based strategies to prevent obesity; however, adolescents may be vulnerable to weight-related messages as rapid development during this lifestage often leads to preoccupation with body size and shape. Objective This study (1) examines secular trends in secondary school curricula topics relevant to the prevention of unhealthy weight-control behaviors; (2) describes cross-sectional associations between weight-related curricula content and students’ use of weight-control behaviors; and (3) assesses whether implementation of school-based obesity prevention policies/practices are longitudinally related to students’ weight-control behaviors. Design/participants The Minnesota School Health Profiles and Minnesota Student Survey (grades 9 and 12) data were used along with National Center for Education Statistics data to examine secular trends; cross-sectional associations (n=141 schools); and longitudinal associations (n=42 schools). Main Outcome Measures Students self-reported their height and weight along with past-year use of healthy (e.g., exercise), unhealthy (e.g., fasting), and extreme (e.g., use laxatives) weight-control behaviors. Statistical analyses performed Descriptive statistics, generalized estimating equations, and generalized linear regression models accounting for school-level demographics. Results There was no observable pattern over the years 2008 to 2014 in the mean number of curricula topics addressing unhealthy weight-control behaviors despite an increase in the prevalence of curricula addressing acceptance of body size differences. Including three versus fewer weight-control topics and specifically including the topic of eating disorders in the curricula were related to a lower school-level percent of students using any extreme weight-control behaviors. In contrast, an overall measure of implementing school-based obesity prevention policies/practices (e.g., prohibited advertising) was unrelated to use of unhealthy or extreme behaviors. Conclusions Results suggest obesity prevention policies/practices do not have unintended consequences for student weight-control behaviors and support the importance of school-based health education as part of efforts to prevent unhealthy behaviors. PMID:27889315
Forging a link between mentoring and collaboration: a new training model for implementation science.
Luke, Douglas A; Baumann, Ana A; Carothers, Bobbi J; Landsverk, John; Proctor, Enola K
2016-10-13
Training investigators for the rapidly developing field of implementation science requires both mentoring and scientific collaboration. Using social network descriptive analyses, visualization, and modeling, this paper presents results of an evaluation of the mentoring and collaborations fostered over time through the National Institute of Mental Health (NIMH) supported by Implementation Research Institute (IRI). Data were comprised of IRI participant self-reported collaborations and mentoring relationships, measured in three annual surveys from 2012 to 2014. Network descriptive statistics, visualizations, and network statistical modeling were conducted to examine patterns of mentoring and collaboration among IRI participants and to model the relationship between mentoring and subsequent collaboration. Findings suggest that IRI is successful in forming mentoring relationships among its participants, and that these mentoring relationships are related to future scientific collaborations. Exponential random graph network models demonstrated that mentoring received in 2012 was positively and significantly related to the likelihood of having a scientific collaboration 2 years later in 2014 (p = 0.001). More specifically, mentoring was significantly related to future collaborations focusing on new research (p = 0.009), grant submissions (p = 0.003), and publications (p = 0.017). Predictions based on the network model suggest that for every additional mentoring relationships established in 2012, the likelihood of a scientific collaboration 2 years later is increased by almost 7 %. These results support the importance of mentoring in implementation science specifically and team science more generally. Mentoring relationships were established quickly and early by the IRI core faculty. IRI fellows reported increasing scientific collaboration of all types over time, including starting new research, submitting new grants, presenting research results, and publishing peer-reviewed papers. Statistical network models demonstrated that mentoring was strongly and significantly related to subsequent scientific collaboration, which supported a core design principle of the IRI. Future work should establish the link between mentoring and scientific productivity. These results may be of interest to team science, as they suggest the importance of mentoring for future team collaborations, as well as illustrate the utility of network analysis for studying team characteristics and activities.
Statistical Learning is Related to Early Literacy-Related Skills
Spencer, Mercedes; Kaschak, Michael P.; Jones, John L.; Lonigan, Christopher J.
2015-01-01
It has been demonstrated that statistical learning, or the ability to use statistical information to learn the structure of one’s environment, plays a role in young children’s acquisition of linguistic knowledge. Although most research on statistical learning has focused on language acquisition processes, such as the segmentation of words from fluent speech and the learning of syntactic structure, some recent studies have explored the extent to which individual differences in statistical learning are related to literacy-relevant knowledge and skills. The present study extends on this literature by investigating the relations between two measures of statistical learning and multiple measures of skills that are critical to the development of literacy—oral language, vocabulary knowledge, and phonological processing—within a single model. Our sample included a total of 553 typically developing children from prekindergarten through second grade. Structural equation modeling revealed that statistical learning accounted for a unique portion of the variance in these literacy-related skills. Practical implications for instruction and assessment are discussed. PMID:26478658
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gilbert, Richard O.
The application of statistics to environmental pollution monitoring studies requires a knowledge of statistical analysis methods particularly well suited to pollution data. This book fills that need by providing sampling plans, statistical tests, parameter estimation procedure techniques, and references to pertinent publications. Most of the statistical techniques are relatively simple, and examples, exercises, and case studies are provided to illustrate procedures. The book is logically divided into three parts. Chapters 1, 2, and 3 are introductory chapters. Chapters 4 through 10 discuss field sampling designs and Chapters 11 through 18 deal with a broad range of statistical analysis procedures. Somemore » statistical techniques given here are not commonly seen in statistics book. For example, see methods for handling correlated data (Sections 4.5 and 11.12), for detecting hot spots (Chapter 10), and for estimating a confidence interval for the mean of a lognormal distribution (Section 13.2). Also, Appendix B lists a computer code that estimates and tests for trends over time at one or more monitoring stations using nonparametric methods (Chapters 16 and 17). Unfortunately, some important topics could not be included because of their complexity and the need to limit the length of the book. For example, only brief mention could be made of time series analysis using Box-Jenkins methods and of kriging techniques for estimating spatial and spatial-time patterns of pollution, although multiple references on these topics are provided. Also, no discussion of methods for assessing risks from environmental pollution could be included.« less
Comment on "Ducklings imprint on the relational concept of 'same or different'".
Hupé, Jean-Michel
2017-02-24
Martinho and Kacelnik's (Reports, 15 July 2016, p. 286) finding that mallard ducklings can deal with abstract concepts is important for understanding the evolution of cognition. However, a statistically more robust analysis of the data calls their conclusions into question. This example brings to light the risk of drawing too strong an inference by relying solely on P values. Copyright © 2017, American Association for the Advancement of Science.
Department of Defense In-House RDT and E Activities
1972-10-30
test and evaluation (RDTJE). In addition, they briefly described each activ "ty’s mission, current important programs. functions, and equipment...capabilities. The statistical data relate to the past and current fiscal years, i.e. 1972 and 1973. Each in-house RDTfE activity of the Department of Defense...manual correction of typo- graphical errors and clarification of obscure abbreviations and contrac- tions. All current DoD RDT&E activities are
Strong biotic influences on regional patterns of climate regulation services
NASA Astrophysics Data System (ADS)
Serna-Chavez, H. M.; Swenson, N. G.; Weiser, M. D.; van Loon, E. E.; Bouten, W.; Davidson, M. D.; van Bodegom, P. M.
2017-05-01
Climate regulation services from forests are an important leverage in global-change mitigation treaties. Like most ecosystem services, climate regulation is the product of various ecological phenomena with unique spatial features. Elucidating which abiotic and biotic factors relate to spatial patterns of climate regulation services advances our understanding of what underlies climate-mitigation potential and its variation within and across ecosystems. Here we quantify and contrast the statistical relations between climate regulation services (albedo and evapotranspiration, primary productivity, and soil carbon) and abiotic and biotic factors. We focus on 16,955 forest plots in a regional extent across the eastern United States. We find the statistical effects of forest litter and understory carbon on climate regulation services to be as strong as those of temperature-precipitation interactions. These biotic factors likely influence climate regulation through changes in vegetation and canopy density, radiance scattering, and decomposition rates. We also find a moderate relation between leaf nitrogen traits and primary productivity at this regional scale. The statistical relation between climate regulation and temperature-precipitation ranges, seasonality, and climatic thresholds highlights a strong feedback with global climate change. Our assessment suggests the expression of strong biotic influences on climate regulation services at a regional, temperate extent. Biotic homogenization and management practices manipulating forest structure and succession will likely strongly impact climate-mitigation potential. The identity, strength, and direction of primary influences differed for each process involved in climate regulation. Hence, different abiotic and biotic factors are needed to monitor and quantify the full climate-mitigation potential of temperate forest ecosystems.
NASA Technical Reports Server (NTRS)
Margutti, R.; Zaninoni, E.; Bernardini, M. G.; Chincarini, G.; Pasotti, F.; Guidorzi, C.; Angelini, Lorella; Burrows, D. N.; Capalbi, M.; Evans, P. A.;
2012-01-01
We present a comprehensive statistical analysis of Swift X-ray light-curves of Gamma- Ray Bursts (GRBs) collecting data from more than 650 GRBs discovered by Swift and other facilities. The unprecedented sample size allows us to constrain the rest-frame X-ray properties of GRBs from a statistical perspective, with particular reference to intrinsic time scales and the energetics of the different light-curve phases in a common rest-frame 0.3-30 keV energy band. Temporal variability episodes are also studied and their properties constrained. Two fundamental questions drive this effort: i) Does the X-ray emission retain any kind of "memory" of the prompt ?-ray phase? ii) Where is the dividing line between long and short GRB X-ray properties? We show that short GRBs decay faster, are less luminous and less energetic than long GRBs in the X-rays, but are interestingly characterized by similar intrinsic absorption. We furthermore reveal the existence of a number of statistically significant relations that link the X-ray to prompt ?-ray parameters in long GRBs; short GRBs are outliers of the majority of these 2-parameter relations. However and more importantly, we report on the existence of a universal 3-parameter scaling that links the X-ray and the ?-ray energy to the prompt spectral peak energy of both long and short GRBs: E(sub X,iso)? E(sup 1.00+/-0.06)(sub ?,iso) /E(sup 0.60+/-0.10)(sub pk).
NASA Astrophysics Data System (ADS)
Margutti, R.; Zaninoni, E.; Bernardini, M. G.; Chincarini, G.; Pasotti, F.; Guidorzi, C.; Angelini, L.; Burrows, D. N.; Capalbi, M.; Evans, P. A.; Gehrels, N.; Kennea, J.; Mangano, V.; Moretti, A.; Nousek, J.; Osborne, J. P.; Page, K. L.; Perri, M.; Racusin, J.; Romano, P.; Sbarufatti, B.; Stafford, S.; Stamatikos, M.
2013-01-01
We present a comprehensive statistical analysis of Swift X-ray light curves of gamma-ray bursts (GRBs) collecting data from more than 650 GRBs discovered by Swift and other facilities. The unprecedented sample size allows us to constrain the rest-frame X-ray properties of GRBs from a statistical perspective, with particular reference to intrinsic time-scales and the energetics of the different light-curve phases in a common rest-frame 0.3-30 keV energy band. Temporal variability episodes are also studied and their properties constrained. Two fundamental questions drive this effort: (i) Does the X-ray emission retain any kind of `memory' of the prompt γ-ray phase? (ii) Where is the dividing line between long and short GRB X-ray properties? We show that short GRBs decay faster, are less luminous and less energetic than long GRBs in the X-rays, but are interestingly characterized by similar intrinsic absorption. We furthermore reveal the existence of a number of statistically significant relations that link the X-ray to prompt γ-ray parameters in long GRBs; short GRBs are outliers of the majority of these two-parameter relations. However and more importantly, we report on the existence of a universal three-parameter scaling that links the X-ray and the γ-ray energy to the prompt spectral peak energy of both long and short GRBs: EX, iso∝E1.00 ± 0.06γ, iso/E0.60 ± 0.10pk.
Multivariate assessment of event-related potentials with the t-CWT method.
Bostanov, Vladimir
2015-11-05
Event-related brain potentials (ERPs) are usually assessed with univariate statistical tests although they are essentially multivariate objects. Brain-computer interface applications are a notable exception to this practice, because they are based on multivariate classification of single-trial ERPs. Multivariate ERP assessment can be facilitated by feature extraction methods. One such method is t-CWT, a mathematical-statistical algorithm based on the continuous wavelet transform (CWT) and Student's t-test. This article begins with a geometric primer on some basic concepts of multivariate statistics as applied to ERP assessment in general and to the t-CWT method in particular. Further, it presents for the first time a detailed, step-by-step, formal mathematical description of the t-CWT algorithm. A new multivariate outlier rejection procedure based on principal component analysis in the frequency domain is presented as an important pre-processing step. The MATLAB and GNU Octave implementation of t-CWT is also made publicly available for the first time as free and open source code. The method is demonstrated on some example ERP data obtained in a passive oddball paradigm. Finally, some conceptually novel applications of the multivariate approach in general and of the t-CWT method in particular are suggested and discussed. Hopefully, the publication of both the t-CWT source code and its underlying mathematical algorithm along with a didactic geometric introduction to some basic concepts of multivariate statistics would make t-CWT more accessible to both users and developers in the field of neuroscience research.
Rosi, Alessia; Bruine de Bruin, Wändi; Del Missier, Fabio; Cavallini, Elena; Russo, Riccardo
2017-12-28
Older adults perform worse than younger adults when applying decision rules to choose between options that vary along multiple attributes. Although previous studies have shown that general fluid cognitive abilities contribute to the accurate application of decision rules, relatively little is known about which specific cognitive abilities play the most important role. We examined the independent roles of working memory, verbal fluency, semantic knowledge, and components of executive functioning. We found that age-related decline in applying decision rules was statistically mediated by age-related decline in working memory and verbal fluency. Our results have implications for theories of aging and decision-making.
Data Acquisition and Preprocessing in Studies on Humans: What Is Not Taught in Statistics Classes?
Zhu, Yeyi; Hernandez, Ladia M; Mueller, Peter; Dong, Yongquan; Forman, Michele R
2013-01-01
The aim of this paper is to address issues in research that may be missing from statistics classes and important for (bio-)statistics students. In the context of a case study, we discuss data acquisition and preprocessing steps that fill the gap between research questions posed by subject matter scientists and statistical methodology for formal inference. Issues include participant recruitment, data collection training and standardization, variable coding, data review and verification, data cleaning and editing, and documentation. Despite the critical importance of these details in research, most of these issues are rarely discussed in an applied statistics program. One reason for the lack of more formal training is the difficulty in addressing the many challenges that can possibly arise in the course of a study in a systematic way. This article can help to bridge this gap between research questions and formal statistical inference by using an illustrative case study for a discussion. We hope that reading and discussing this paper and practicing data preprocessing exercises will sensitize statistics students to these important issues and achieve optimal conduct, quality control, analysis, and interpretation of a study.
Chiou, Chei-Chang; Wang, Yu-Min; Lee, Li-Tze
2014-08-01
Statistical knowledge is widely used in academia; however, statistics teachers struggle with the issue of how to reduce students' statistics anxiety and enhance students' statistics learning. This study assesses the effectiveness of a "one-minute paper strategy" in reducing students' statistics-related anxiety and in improving students' statistics-related achievement. Participants were 77 undergraduates from two classes enrolled in applied statistics courses. An experiment was implemented according to a pretest/posttest comparison group design. The quasi-experimental design showed that the one-minute paper strategy significantly reduced students' statistics anxiety and improved students' statistics learning achievement. The strategy was a better instructional tool than the textbook exercise for reducing students' statistics anxiety and improving students' statistics achievement.
Stait, Emma; Calnan, Michael
2016-10-18
During the last two decades, differential consumption patterns in health-related behaviours have increasingly been highlighted as playing an important role in explaining persistent and widening health inequalities. This period has also seen government public health policies in England place a greater emphasis on changing 'lifestyle' behaviours, in an attempt to tackle social inequalities in health. The aim of this study was to empirically examine the variation in health-related behaviour in relation to socio-economic position, in the English adult population, to determine the nature of this relationship and whether it has changed over time. The study population was derived from the Health Survey for England between 2001 and 2012 (n = 56,468). The relationships between health-related behaviour (smoking, fruit and vegetable intake, alcohol consumption and physical activity) and three socioeconomic indicators (educational level, occupational social class and equivilised household income) were analysed using log bi-nomial regression. The study found that each of the three socio-economic indicators were statistically related to smoking, fruit and vegetable consumption and alcohol intake, with the strongest relationship found for smoking. For physical activity, no relationship was found in 2003 by education or income and in 2008 by occupation. Statistical analysis showed that the difference between those at the highest and lowest end of the socio-economic indicators had widened in relation to smoking, as measured by educational level, occupation and household income. A similar trend was also found for physical activity as measured by educational level and household income. However, for fruit and vegetable intake and alcohol consumption, the relationship between health-related behaviour and socio-economic position had narrowed over time as measured by education and income. The findings provided only partial support for the thesis that socio-economic variations in health-related behaviours may be significant in explaining widening health inequalities. The significance of socio-economic variations in health-related behaviours might reflect both materialist and cultural explanations for socio-economic inequalities although it was not possible to separate and estimate the relative importance of these effects.
Vieira, Rute; McDonald, Suzanne; Araújo-Soares, Vera; Sniehotta, Falko F; Henderson, Robin
2017-09-01
N-of-1 studies are based on repeated observations within an individual or unit over time and are acknowledged as an important research method for generating scientific evidence about the health or behaviour of an individual. Statistical analyses of n-of-1 data require accurate modelling of the outcome while accounting for its distribution, time-related trend and error structures (e.g., autocorrelation) as well as reporting readily usable contextualised effect sizes for decision-making. A number of statistical approaches have been documented but no consensus exists on which method is most appropriate for which type of n-of-1 design. We discuss the statistical considerations for analysing n-of-1 studies and briefly review some currently used methodologies. We describe dynamic regression modelling as a flexible and powerful approach, adaptable to different types of outcomes and capable of dealing with the different challenges inherent to n-of-1 statistical modelling. Dynamic modelling borrows ideas from longitudinal and event history methodologies which explicitly incorporate the role of time and the influence of past on future. We also present an illustrative example of the use of dynamic regression on monitoring physical activity during the retirement transition. Dynamic modelling has the potential to expand researchers' access to robust and user-friendly statistical methods for individualised studies.
Regional Monitoring of Cervical Cancer.
Crisan-Vida, Mihaela; Lupse, Oana Sorina; Stoicu-Tivadar, Lacramioara; Salvari, Daniela; Catanet, Radu; Bernad, Elena
2017-01-01
Cervical cancer is one of the most important causes of death in women in fertile age in Romania. In order to discover high-risk situations in the first stages of the disease it is important to enhance prevention actions, and ICT, respectively cloud computing and Big Data currently support such activities. The national screening program uses an information system that based on data from different medical units gives feedback related to the women healthcare status and provides statistics and reports. In order to ensure the continuity of care it is updated with HL7 CDA support and cloud computing. The current paper presents the solution and several results.
Relative risk estimates from spatial and space-time scan statistics: Are they biased?
Prates, Marcos O.; Kulldorff, Martin; Assunção, Renato M.
2014-01-01
The purely spatial and space-time scan statistics have been successfully used by many scientists to detect and evaluate geographical disease clusters. Although the scan statistic has high power in correctly identifying a cluster, no study has considered the estimates of the cluster relative risk in the detected cluster. In this paper we evaluate whether there is any bias on these estimated relative risks. Intuitively, one may expect that the estimated relative risks has upward bias, since the scan statistic cherry picks high rate areas to include in the cluster. We show that this intuition is correct for clusters with low statistical power, but with medium to high power the bias becomes negligible. The same behaviour is not observed for the prospective space-time scan statistic, where there is an increasing conservative downward bias of the relative risk as the power to detect the cluster increases. PMID:24639031
Zgheib, Nathalie K; Sleiman, Fatima; Nasreddine, Lara; Nasrallah, Mona; Nakhoul, Nancy; Isma’eel, Hussain; Tamim, Hani
2018-01-01
In Lebanon, data stemming from national cross-sectional surveys indicated significant increasing trends in the prevalence of cardiovascular diseases and associated behavioral and age-related risk factors. To our knowledge, no data are available on relative telomere length (RTL) as a potential biomarker for age-related diseases in a Lebanese population. The aim of this study was to evaluate whether there is an association between RTL and demographic characteristics, lifestyle habits and diseases in the Lebanese. This was a cross-sectional study of 497 Lebanese subjects. Peripheral blood RTL was measured by amplifying telomere and single copy gene using real-time PCR. Mean ± SD RTL was 1.42 ± 0.83, and it was categorized into 3 tertiles. Older age (P=0.002) and wider waist circumference (WC) (P=0.001) were statistically significantly associated with shorter RTL. Multinomial logistic regression showed that subjects who had some level of sleeping difficulty had a statistically significantly shorter RTL when compared to those with no sleeping difficulties at all [OR (95% CI): 2.01 (1.11-3.62) in the first RTL tertile]. Importantly, statistically significantly shorter RTL was found with every additional 10 cm of WC [OR (95% CI): 1.30 (1.11-1.52) for first RTL tertile]. In addition, and after performing the multivariate logistic regression and adjusting for “predictors” of RTL, the odds of having hypertension or being treated for hypertension were higher in patients who had shorter RTL: OR (95% CI): 2.45 (1.36-4.44) and 2.28 (1.22-4.26) in the first RTL tertiles respectively with a similar trend, though not statistically significant, in the second RTL tertiles. This is the first study in Lebanon to show an association between age, central obesity, poor sleep and hypertension and RTL. It is hoped that telomere length measurement be potentially used as a biomarker for biological age and age-related diseases and progression in the Lebanese. PMID:29392083
Recurrence and interoccurrence behavior of self-organized complex phenomena
NASA Astrophysics Data System (ADS)
Abaimov, S. G.; Turcotte, D. L.; Shcherbakov, R.; Rundle, J. B.
2007-08-01
The sandpile, forest-fire and slider-block models are said to exhibit self-organized criticality. Associated natural phenomena include landslides, wildfires, and earthquakes. In all cases the frequency-size distributions are well approximated by power laws (fractals). Another important aspect of both the models and natural phenomena is the statistics of interval times. These statistics are particularly important for earthquakes. For earthquakes it is important to make a distinction between interoccurrence and recurrence times. Interoccurrence times are the interval times between earthquakes on all faults in a region whereas recurrence times are interval times between earthquakes on a single fault or fault segment. In many, but not all cases, interoccurrence time statistics are exponential (Poissonian) and the events occur randomly. However, the distribution of recurrence times are often Weibull to a good approximation. In this paper we study the interval statistics of slip events using a slider-block model. The behavior of this model is sensitive to the stiffness α of the system, α=kC/kL where kC is the spring constant of the connector springs and kL is the spring constant of the loader plate springs. For a soft system (small α) there are no system-wide events and interoccurrence time statistics of the larger events are Poissonian. For a stiff system (large α), system-wide events dominate the energy dissipation and the statistics of the recurrence times between these system-wide events satisfy the Weibull distribution to a good approximation. We argue that this applicability of the Weibull distribution is due to the power-law (scale invariant) behavior of the hazard function, i.e. the probability that the next event will occur at a time t0 after the last event has a power-law dependence on t0. The Weibull distribution is the only distribution that has a scale invariant hazard function. We further show that the onset of system-wide events is a well defined critical point. We find that the number of system-wide events NSWE satisfies the scaling relation NSWE ∝(α-αC)δ where αC is the critical value of the stiffness. The system-wide events represent a new phase for the slider-block system.
Teaching for Statistical Literacy: Utilising Affordances in Real-World Data
ERIC Educational Resources Information Center
Chick, Helen L.; Pierce, Robyn
2012-01-01
It is widely held that context is important in teaching mathematics and statistics. Consideration of context is central to statistical thinking, and any teaching of statistics must incorporate this aspect. Indeed, it has been advocated that real-world data sets can motivate the learning of statistical principles. It is not, however, a…
ERIC Educational Resources Information Center
Levin, Joel R.; And Others
1993-01-01
Journal editors respond to criticisms of reliance on statistical significance in research reporting. Joel R. Levin ("Journal of Educational Psychology") defends its use, whereas William D. Schafer ("Measurement and Evaluation in Counseling and Development") emphasizes the distinction between statistically significant and important. William Asher…
Problematizing Statistical Literacy: An Intersection of Critical and Statistical Literacies
ERIC Educational Resources Information Center
Weiland, Travis
2017-01-01
In this paper, I problematize traditional notions of statistical literacy by juxtaposing it with critical literacy. At the school level statistical literacy is vitally important for students who are preparing to become citizens in modern societies that are increasingly shaped and driven by data based arguments. The teaching of statistics, which is…
Developing Statistical Literacy with Year 9 Students: A Collaborative Research Project
ERIC Educational Resources Information Center
Sharma, Sashi
2013-01-01
Advances in technology and communication have increased the amount of statistical information delivered through everyday media. The importance of statistics in everyday life has led to calls for increased attention to statistical literacy in the mathematics curriculum (Watson 2006). Gal (2004) sees statistical literacy as the need for students to…
On the statistical mechanics of species abundance distributions.
Bowler, Michael G; Kelly, Colleen K
2012-09-01
A central issue in ecology is that of the factors determining the relative abundance of species within a natural community. The proper application of the principles of statistical physics to species abundance distributions (SADs) shows that simple ecological properties could account for the near universal features observed. These properties are (i) a limit on the number of individuals in an ecological guild and (ii) per capita birth and death rates. They underpin the neutral theory of Hubbell (2001), the master equation approach of Volkov et al. (2003, 2005) and the idiosyncratic (extreme niche) theory of Pueyo et al. (2007); they result in an underlying log series SAD, regardless of neutral or niche dynamics. The success of statistical mechanics in this application implies that communities are in dynamic equilibrium and hence that niches must be flexible and that temporal fluctuations on all sorts of scales are likely to be important in community structure. Copyright © 2012 Elsevier Inc. All rights reserved.
Seasonal Drought Prediction: Advances, Challenges, and Future Prospects
NASA Astrophysics Data System (ADS)
Hao, Zengchao; Singh, Vijay P.; Xia, Youlong
2018-03-01
Drought prediction is of critical importance to early warning for drought managements. This review provides a synthesis of drought prediction based on statistical, dynamical, and hybrid methods. Statistical drought prediction is achieved by modeling the relationship between drought indices of interest and a suite of potential predictors, including large-scale climate indices, local climate variables, and land initial conditions. Dynamical meteorological drought prediction relies on seasonal climate forecast from general circulation models (GCMs), which can be employed to drive hydrological models for agricultural and hydrological drought prediction with the predictability determined by both climate forcings and initial conditions. Challenges still exist in drought prediction at long lead time and under a changing environment resulting from natural and anthropogenic factors. Future research prospects to improve drought prediction include, but are not limited to, high-quality data assimilation, improved model development with key processes related to drought occurrence, optimal ensemble forecast to select or weight ensembles, and hybrid drought prediction to merge statistical and dynamical forecasts.
NASA Astrophysics Data System (ADS)
Gaitán Fernández, E.; García Moreno, R.; Pino Otín, M. R.; Ribalaygua Batalla, J.
2012-04-01
Climate and soil are two of the most important limiting factors for agricultural production. Nowadays climate change has been documented in many geographical locations affecting different cropping systems. The General Circulation Models (GCM) has become important tools to simulate the more relevant aspects of the climate expected for the XXI century in the frame of climatic change. These models are able to reproduce the general features of the atmospheric dynamic but their low resolution (about 200 Km) avoids a proper simulation of lower scale meteorological effects. Downscaling techniques allow overcoming this problem by adapting the model outcomes to local scale. In this context, FIC (Fundación para la Investigación del Clima) has developed a statistical downscaling technique based on a two step analogue methods. This methodology has been broadly tested on national and international environments leading to excellent results on future climate models. In a collaboration project, this statistical downscaling technique was applied to predict future scenarios for the grape growing systems in Spain. The application of such model is very important to predict expected climate for the different growing crops, mainly for grape, where the success of different varieties are highly related to climate and soil. The model allowed the implementation of agricultural conservation practices in the crop production, detecting highly sensible areas to negative impacts produced by any modification of climate in the different regions, mainly those protected with protected designation of origin, and the definition of new production areas with optimal edaphoclimatic conditions for the different varieties.
NASA Astrophysics Data System (ADS)
Bergant, Klemen; Kajfež-Bogataj, Lučka; Črepinšek, Zalika
2002-02-01
Phenological observations are a valuable source of information for investigating the relationship between climate variation and plant development. Potential climate change in the future will shift the occurrence of phenological phases. Information about future climate conditions is needed in order to estimate this shift. General circulation models (GCM) provide the best information about future climate change. They are able to simulate reliably the most important mean features on a large scale, but they fail on a regional scale because of their low spatial resolution. A common approach to bridging the scale gap is statistical downscaling, which was used to relate the beginning of flowering of Taraxacum officinale in Slovenia with the monthly mean near-surface air temperature for January, February and March in Central Europe. Statistical models were developed and tested with NCAR/NCEP Reanalysis predictor data and EARS predictand data for the period 1960-1999. Prior to developing statistical models, empirical orthogonal function (EOF) analysis was employed on the predictor data. Multiple linear regression was used to relate the beginning of flowering with expansion coefficients of the first three EOF for the Janauary, Febrauary and March air temperatures, and a strong correlation was found between them. Developed statistical models were employed on the results of two GCM (HadCM3 and ECHAM4/OPYC3) to estimate the potential shifts in the beginning of flowering for the periods 1990-2019 and 2020-2049 in comparison with the period 1960-1989. The HadCM3 model predicts, on average, 4 days earlier occurrence and ECHAM4/OPYC3 5 days earlier occurrence of flowering in the period 1990-2019. The analogous results for the period 2020-2049 are a 10- and 11-day earlier occurrence.
A Probabilistic Approach to Fitting Period–luminosity Relations and Validating Gaia Parallaxes
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sesar, Branimir; Fouesneau, Morgan; Bailer-Jones, Coryn A. L.
Pulsating stars, such as Cepheids, Miras, and RR Lyrae stars, are important distance indicators and calibrators of the “cosmic distance ladder,” and yet their period–luminosity–metallicity (PLZ) relations are still constrained using simple statistical methods that cannot take full advantage of available data. To enable optimal usage of data provided by the Gaia mission, we present a probabilistic approach that simultaneously constrains parameters of PLZ relations and uncertainties in Gaia parallax measurements. We demonstrate this approach by constraining PLZ relations of type ab RR Lyrae stars in near-infrared W 1 and W 2 bands, using Tycho- Gaia Astrometric Solution (TGAS) parallaxmore » measurements for a sample of ≈100 type ab RR Lyrae stars located within 2.5 kpc of the Sun. The fitted PLZ relations are consistent with previous studies, and in combination with other data, deliver distances precise to 6% (once various sources of uncertainty are taken into account). To a precision of 0.05 mas (1 σ ), we do not find a statistically significant offset in TGAS parallaxes for this sample of distant RR Lyrae stars (median parallax of 0.8 mas and distance of 1.4 kpc). With only minor modifications, our probabilistic approach can be used to constrain PLZ relations of other pulsating stars, and we intend to apply it to Cepheid and Mira stars in the near future.« less
Likelihoods for fixed rank nomination networks
HOFF, PETER; FOSDICK, BAILEY; VOLFOVSKY, ALEX; STOVEL, KATHERINE
2014-01-01
Many studies that gather social network data use survey methods that lead to censored, missing, or otherwise incomplete information. For example, the popular fixed rank nomination (FRN) scheme, often used in studies of schools and businesses, asks study participants to nominate and rank at most a small number of contacts or friends, leaving the existence of other relations uncertain. However, most statistical models are formulated in terms of completely observed binary networks. Statistical analyses of FRN data with such models ignore the censored and ranked nature of the data and could potentially result in misleading statistical inference. To investigate this possibility, we compare Bayesian parameter estimates obtained from a likelihood for complete binary networks with those obtained from likelihoods that are derived from the FRN scheme, and therefore accommodate the ranked and censored nature of the data. We show analytically and via simulation that the binary likelihood can provide misleading inference, particularly for certain model parameters that relate network ties to characteristics of individuals and pairs of individuals. We also compare these different likelihoods in a data analysis of several adolescent social networks. For some of these networks, the parameter estimates from the binary and FRN likelihoods lead to different conclusions, indicating the importance of analyzing FRN data with a method that accounts for the FRN survey design. PMID:25110586
Spatial diffusion of influenza outbreak-related climate factors in Chiang Mai Province, Thailand.
Nakapan, Supachai; Tripathi, Nitin Kumar; Tipdecho, Taravudh; Souris, Marc
2012-10-24
Influenza is one of the most important leading causes of respiratory illness in the countries located in the tropical areas of South East Asia and Thailand. In this study the climate factors associated with influenza incidence in Chiang Mai Province, Northern Thailand, were investigated. Identification of factors responsible for influenza outbreaks and the mapping of potential risk areas in Chiang Mai are long overdue. This work examines the association between yearly climate patterns between 2001 and 2008 and influenza outbreaks in the Chiang Mai Province. The climatic factors included the amount of rainfall, percent of rainy days, relative humidity, maximum, minimum temperatures and temperature difference. The study develops a statistical analysis to quantitatively assess the relationship between climate and influenza outbreaks and then evaluate its suitability for predicting influenza outbreaks. A multiple linear regression technique was used to fit the statistical model. The Inverse Distance Weighted (IDW) interpolation and Geographic Information System (GIS) techniques were used in mapping the spatial diffusion of influenza risk zones. The results show that there is a significance correlation between influenza outbreaks and climate factors for the majority of the studied area. A statistical analysis was conducted to assess the validity of the model comparing model outputs and actual outbreaks.
Large and unexpected runup events and their relation to the incident wave field
NASA Astrophysics Data System (ADS)
Li, C.; Ozkan-Haller, H. T.; Garcia-Medina, G.; Holman, R. A.; Ruggiero, P.
2016-12-01
Unusually large runup events are important for the prediction of dune erosion, inundation and coastal flooding during storms and lie at the tail of swash maxima probability distributions. We also distinguish a unique type of large runup event that is sudden and unexpected even if the landward reach of the runup is not a statistical extreme. These unusual runup events are anecdotally reported to be more prevalent on dissipative beaches and are the leading cause of death by drowning along the U.S. Pacific Northwest (northern California, Oregon, and Washington). Herein we examine the environmental conditions that are conducive to large and unexpected runup events and begin to forecast their potential occurrence, validating these predictions with ongoing observations. We explore and compare the statistics of large runup events on two beach types, a dissipative beach at Agate Beach, OR, and the intermediate/reflective site at Duck, NC. Video-based runup observations along with incident wave information from offshore instrumentation are used to assess how frequently large or unexpected runup events occur, how the statistics of these runup events relate to the incident wave characteristics (e.g. height, period, narrow-bandedness), and whether or not these events are indeed more prevalent on dissipative beaches.
NASA Astrophysics Data System (ADS)
Graham, Wendy; Destouni, Georgia; Demmy, George; Foussereau, Xavier
1998-07-01
The methodology developed in Destouni and Graham [Destouni, G., Graham, W.D., 1997. The influence of observation method on local concentration statistics in the subsurface. Water Resour. Res. 33 (4) 663-676.] for predicting locally measured concentration statistics for solute transport in heterogeneous porous media under saturated flow conditions is applied to the prediction of conservative nonreactive solute transport in the vadose zone where observations are obtained by soil coring. Exact analytical solutions are developed for both the mean and variance of solute concentrations measured in discrete soil cores using a simplified physical model for vadose-zone flow and solute transport. Theoretical results show that while the ensemble mean concentration is relatively insensitive to the length-scale of the measurement, predictions of the concentration variance are significantly impacted by the sampling interval. Results also show that accounting for vertical heterogeneity in the soil profile results in significantly less spreading in the mean and variance of the measured solute breakthrough curves, indicating that it is important to account for vertical heterogeneity even for relatively small travel distances. Model predictions for both the mean and variance of locally measured solute concentration, based on independently estimated model parameters, agree well with data from a field tracer test conducted in Manatee County, Florida.
Labots, M; Laarakker, M C; Ohl, F; van Lith, H A
2016-06-29
Selecting chromosome substitution strains (CSSs, also called consomic strains/lines) used in the search for quantitative trait loci (QTLs) consistently requires the identification of the respective phenotypic trait of interest and is simply based on a significant difference between a consomic and host strain. However, statistical significance as represented by P values does not necessarily predicate practical importance. We therefore propose a method that pays attention to both the statistical significance and the actual size of the observed effect. The present paper extends on this approach and describes in more detail the use of effect size measures (Cohen's d, partial eta squared - η p (2) ) together with the P value as statistical selection parameters for the chromosomal assignment of QTLs influencing anxiety-related behavior and locomotion in laboratory mice. The effect size measures were based on integrated behavioral z-scoring and were calculated in three experiments: (A) a complete consomic male mouse panel with A/J as the donor strain and C57BL/6J as the host strain. This panel, including host and donor strains, was analyzed in the modified Hole Board (mHB). The consomic line with chromosome 19 from A/J (CSS-19A) was selected since it showed increased anxiety-related behavior, but similar locomotion compared to its host. (B) Following experiment A, female CSS-19A mice were compared with their C57BL/6J counterparts; however no significant differences and effect sizes close to zero were found. (C) A different consomic mouse strain (CSS-19PWD), with chromosome 19 from PWD/PhJ transferred on the genetic background of C57BL/6J, was compared with its host strain. Here, in contrast with CSS-19A, there was a decreased overall anxiety in CSS-19PWD compared to C57BL/6J males, but not locomotion. This new method shows an improved way to identify CSSs for QTL analysis for anxiety-related behavior using a combination of statistical significance testing and effect sizes. In addition, an intercross between CSS-19A and CSS-19PWD may be of interest for future studies on the genetic background of anxiety-related behavior.
Selimović-Dragaš, Mediha; Hasić-Branković, Lajla; Korać, Fehim; Đapo, Nermin; Huseinbegović, Amina; Kobašlija, Sedin; Lekić, Meliha; Hatibović-Kofman, Šahza
2013-08-01
Fluoride release is important characteristic of glass-ionomer cements. Quantity of fluoride ions released from the glass-ionomer cements has major importance in definition of their biological activity. The objectives of this study were to define the quantity of fluoride ions released from the experimental glass-ionomer cements and to define the effect of fluoride ions released from the experimental glass-ionomer cements on their cytotoxicity. Concentrations of the fluoride ions released in the evaluated glass-ionomer cements were measured indirectly, by the fluoride-selective WTW, F500 electrode potential, combined with reference R503/D electrode. Statistical analyses of F-ion concentrations released by all glass-ionomers evaluated at two time points, after 8 and after 24 hours, show statistically higher fluoride releases from RMGICs: Vitrebond, Fuji II LC and Fuji Plus, when compared to conventional glass-ionomer cements: Fuji Triage, Fuji IX GP Fast and Ketac Silver, both after 8 and after 24 hours. Correlation coefficient between concentrations of fluoride ion released by evaluated glass-ionomer cements and cytotoxic response of UMR-106 osteoblast cell-line are relatively high, but do not reach levels of biological significance. Correlation between concentrations of fluoride ion released and cytotoxic response of NIH3T3 mouse fibroblast cell line after 8 hours is high, positive and statistically significant for conventional GICs, Fuji Triage and Fuji IX GP Fast, and RMGIC, Fuji II LC. Statistically significant Correlation coefficient between concentrations of fluoride ion released and cytotoxic response of NIH3T3 cell line after 24 hours is defined for RMGIC Fuji II LC only.
Historical Data Analysis of Hospital Discharges Related to the Amerithrax Attack in Florida
Burke, Lauralyn K.; Brown, C. Perry; Johnson, Tammie M.
2016-01-01
Interrupted time-series analysis (ITSA) can be used to identify, quantify, and evaluate the magnitude and direction of an event on the basis of time-series data. This study evaluates the impact of the bioterrorist anthrax attacks (“Amerithrax”) on hospital inpatient discharges in the metropolitan statistical area of Palm Beach, Broward, and Miami-Dade counties in the fourth quarter of 2001. Three statistical methods—standardized incidence ratio (SIR), segmented regression, and an autoregressive integrated moving average (ARIMA)—were used to determine whether Amerithrax influenced inpatient utilization. The SIR found a non–statistically significant 2 percent decrease in hospital discharges. Although the segmented regression test found a slight increase in the discharge rate during the fourth quarter, it was also not statistically significant; therefore, it could not be attributed to Amerithrax. Segmented regression diagnostics preparing for ARIMA indicated that the quarterly data time frame was not serially correlated and violated one of the assumptions for the use of the ARIMA method and therefore could not properly evaluate the impact on the time-series data. Lack of data granularity of the time frames hindered the successful evaluation of the impact by the three analytic methods. This study demonstrates that the granularity of the data points is as important as the number of data points in a time series. ITSA is important for the ability to evaluate the impact that any hazard may have on inpatient utilization. Knowledge of hospital utilization patterns during disasters offer healthcare and civic professionals valuable information to plan, respond, mitigate, and evaluate any outcomes stemming from biothreats. PMID:27843420
Enhanced echolocation via robust statistics and super-resolution of sonar images
NASA Astrophysics Data System (ADS)
Kim, Kio
Echolocation is a process in which an animal uses acoustic signals to exchange information with environments. In a recent study, Neretti et al. have shown that the use of robust statistics can significantly improve the resiliency of echolocation against noise and enhance its accuracy by suppressing the development of sidelobes in the processing of an echo signal. In this research, the use of robust statistics is extended to problems in underwater explorations. The dissertation consists of two parts. Part I describes how robust statistics can enhance the identification of target objects, which in this case are cylindrical containers filled with four different liquids. Particularly, this work employs a variation of an existing robust estimator called an L-estimator, which was first suggested by Koenker and Bassett. As pointed out by Au et al.; a 'highlight interval' is an important feature, and it is closely related with many other important features that are known to be crucial for dolphin echolocation. A varied L-estimator described in this text is used to enhance the detection of highlight intervals, which eventually leads to a successful classification of echo signals. Part II extends the problem into 2 dimensions. Thanks to the advances in material and computer technology, various sonar imaging modalities are available on the market. By registering acoustic images from such video sequences, one can extract more information on the region of interest. Computer vision and image processing allowed application of robust statistics to the acoustic images produced by forward looking sonar systems, such as Dual-frequency Identification Sonar and ProViewer. The first use of robust statistics for sonar image enhancement in this text is in image registration. Random Sampling Consensus (RANSAC) is widely used for image registration. The registration algorithm using RANSAC is optimized for sonar image registration, and the performance is studied. The second use of robust statistics is in fusing the images. It is shown that the maximum a posteriori fusion method can be formulated in a Kalman filter-like manner, and also that the resulting expression is identical to a W-estimator with a specific weight function.
Spatiotemporal Analysis of the Malaria Epidemic in Mainland China, 2004-2014.
Huang, Qiang; Hu, Lin; Liao, Qi-Bin; Xia, Jing; Wang, Qian-Ru; Peng, Hong-Juan
2017-08-01
The purpose of this study is to characterize spatiotemporal heterogeneities in malaria distribution at a provincial level and investigate the association between malaria incidence and climate factors from 2004 to 2014 in China to inform current malaria control efforts. National malaria incidence peaked (4.6/100,000) in 2006 and decreased to a very low level (0.21/100,000) in 2014, and the proportion of imported cases increased from 16.2% in 2004 to 98.2% in 2014. Statistical analyses of global and local spatial autocorrelations and purely spatial scan statistics revealed that malaria was localized in Hainan, Anhui, and Yunnan during 2004-2009 and then gradually shifted and clustered in Yunnan after 2010. Purely temporal clusters shortened to less than 5 months during 2012-2014. The two most likely clusters detected using spatiotemporal analysis occurred in Anhui between July 2005 and November 2007 and Yunnan between January 2010 and June 2012. Correlation coefficients for the association between malaria incidence and climate factors sharply decreased after 2010, and there were zero-month lag effects for climate factors during 2010-2014. Overall, the spatiotemporal distribution of malaria in China changed from relatively scattered (2004-2009) to relatively clustered (2010-2014). As the proportion of imported cases increased, the effect of climate factors on malaria incidence has gradually become weaker since 2011. Therefore, new warning systems should be applied to monitor resurgence and outbreaks of malaria in mainland China, and quarantine at borders should be reinforced to control the increasingly trend of imported malaria cases.
MIDAS: Regionally linear multivariate discriminative statistical mapping.
Varol, Erdem; Sotiras, Aristeidis; Davatzikos, Christos
2018-07-01
Statistical parametric maps formed via voxel-wise mass-univariate tests, such as the general linear model, are commonly used to test hypotheses about regionally specific effects in neuroimaging cross-sectional studies where each subject is represented by a single image. Despite being informative, these techniques remain limited as they ignore multivariate relationships in the data. Most importantly, the commonly employed local Gaussian smoothing, which is important for accounting for registration errors and making the data follow Gaussian distributions, is usually chosen in an ad hoc fashion. Thus, it is often suboptimal for the task of detecting group differences and correlations with non-imaging variables. Information mapping techniques, such as searchlight, which use pattern classifiers to exploit multivariate information and obtain more powerful statistical maps, have become increasingly popular in recent years. However, existing methods may lead to important interpretation errors in practice (i.e., misidentifying a cluster as informative, or failing to detect truly informative voxels), while often being computationally expensive. To address these issues, we introduce a novel efficient multivariate statistical framework for cross-sectional studies, termed MIDAS, seeking highly sensitive and specific voxel-wise brain maps, while leveraging the power of regional discriminant analysis. In MIDAS, locally linear discriminative learning is applied to estimate the pattern that best discriminates between two groups, or predicts a variable of interest. This pattern is equivalent to local filtering by an optimal kernel whose coefficients are the weights of the linear discriminant. By composing information from all neighborhoods that contain a given voxel, MIDAS produces a statistic that collectively reflects the contribution of the voxel to the regional classifiers as well as the discriminative power of the classifiers. Critically, MIDAS efficiently assesses the statistical significance of the derived statistic by analytically approximating its null distribution without the need for computationally expensive permutation tests. The proposed framework was extensively validated using simulated atrophy in structural magnetic resonance imaging (MRI) and further tested using data from a task-based functional MRI study as well as a structural MRI study of cognitive performance. The performance of the proposed framework was evaluated against standard voxel-wise general linear models and other information mapping methods. The experimental results showed that MIDAS achieves relatively higher sensitivity and specificity in detecting group differences. Together, our results demonstrate the potential of the proposed approach to efficiently map effects of interest in both structural and functional data. Copyright © 2018. Published by Elsevier Inc.
Ahmadian, Leila; Dorosti, Nafise; Khajouei, Reza; Gohari, Sadrieh Hajesmaeel
2017-06-01
Hospital Information Systems (HIS) are used for easy access to information, improvement of documentation and reducing errors. Nonetheless, using these systems is faced with some barriers and obstacles. This study identifies the challenges and the obstacles of using these systems in the academic and non-academic hospitals in Kerman. This is a cross-sectional study which was carried out in 2015. The statistical population in this study consisted of the nurses who had been working in the academic and non-academic hospitals in Kerman. A questionnaire consisting of two sections was used. The first section consisted of the demographic information of the participants and the second section comprised 34 questions about the challenges of HIS use. Data were analyzed by the descriptive and statistical analysis (t-test, and ANOVA) using SPSS 19 software. The most common and important challenges in the academic hospitals were about human environment factors, particularly "negative attitude of society toward using HIS". In the non-academic hospitals, the most common and important challenges were related to human factors, and among them, "no incentive to use system" was the main factor. The results of the t-test method revealed that there was a significant relationship between gender and the mean score of challenges related to the organizational environment category in the academic hospitals and between familiarity with HIS and mean score of human environment factors (p<0.05). The results of the ANOVA test also revealed that the educational degree and work experience in the healthcare environment (years) in the academic hospitals have a significant relationship with the mean score related to the hardware challenges, as well, experience with HIS has a significant relationship, with the mean score related to the human challenges (p<0.05). The most important challenges in using the information systems are the factors related to the human environment and the human factors. The results of this study can bring a good perspective to the policy makers and the managers regarding obstacles of using HISs from the nurses' perspective, so that they can solve their problems and can successfully implement these systems.
Wästberg, Birgitta A; Erlandsson, Lena-Karin; Eklund, Mona
2016-01-01
The study aimed to investigate (a) if women's perceptions of their work environment changed during a 16-week rehabilitation period and at a 12-month follow-up; (b) whether such changes were related to outcomes in terms of return to work, well-being and valued occupations. Eighty-four gainfully employed women on sick-leave due to stress-related disorders responded to instruments assessing perceptions of the work environment, well-being (self-esteem, self-mastery, quality of life, perceived stress, self-rated health) and perceived occupational value. Data about return to work were collected from registers. Non-parametric statistics were used. The increase in the women's ratings of their work environment was non-significant between baseline and completed rehabilitation but was statistically significant between baseline and the 12-month follow-up. No relationships were found between changes in perceptions of the work environment and outcomes after the rehabilitation. At the follow-up, however, there were associations between perceived work environment changes in a positive direction and return to work; improved self-esteem, self-mastery, quality of life, perceived occupational value and self-rated health; and reduced stress. It seems important to consider the work environment in rehabilitation for stress-related problems, and a follow-up appears warranted to detect changes and associations not visible immediately after rehabilitation. Work environment Perceptions of the work environment seem important for return to work, although other factors are likely to contribute as well. Perceptions of the work environment are associated with several aspects of well-being. When developing rehabilitation interventions a focus on the clients' perceptions of their work environment seems vital.
An empirical-statistical model for laser cladding of Ti-6Al-4V powder on Ti-6Al-4V substrate
NASA Astrophysics Data System (ADS)
Nabhani, Mohammad; Razavi, Reza Shoja; Barekat, Masoud
2018-03-01
In this article, Ti-6Al-4V powder alloy was directly deposited on Ti-6Al-4V substrate using laser cladding process. In this process, some key parameters such as laser power (P), laser scanning rate (V) and powder feeding rate (F) play important roles. Using linear regression analysis, this paper develops the empirical-statistical relation between these key parameters and geometrical characteristics of single clad tracks (i.e. clad height, clad width, penetration depth, wetting angle, and dilution) as a combined parameter (PαVβFγ). The results indicated that the clad width linearly depended on PV-1/3 and powder feeding rate had no effect on it. The dilution controlled by a combined parameter as VF-1/2 and laser power was a dispensable factor. However, laser power was the dominant factor for the clad height, penetration depth, and wetting angle so that they were proportional to PV-1F1/4, PVF-1/8, and P3/4V-1F-1/4, respectively. Based on the results of correlation coefficient (R > 0.9) and analysis of residuals, it was confirmed that these empirical-statistical relations were in good agreement with the measured values of single clad tracks. Finally, these relations led to the design of a processing map that can predict the geometrical characteristics of the single clad tracks based on the key parameters.
Foley, Louise; Jiang, Yannan; Ni Mhurchu, Cliona; Jull, Andrew; Prapavessis, Harry; Rodgers, Anthony; Maddison, Ralph
2014-04-03
The prevention and treatment of childhood obesity is a key public health challenge. However, certain groups within populations have markedly different risk profiles for obesity and related health behaviours. Well-designed subgroup analysis can identify potential differential effects of obesity interventions, which may be important for reducing health inequalities. The study aim was to evaluate the consistency of the effects of active video games across important subgroups in a randomised controlled trial (RCT). A two-arm, parallel RCT was conducted in overweight or obese children (n=322; aged 10-14 years) to determine the effect of active video games on body composition. Statistically significant overall treatment effects favouring the intervention group were found for body mass index, body mass index z-score and percentage body fat at 24 weeks. For these outcomes, pre-specified subgroup analyses were conducted among important baseline demographic (ethnicity, sex) and prognostic (cardiovascular fitness) groups. No statistically significant interaction effects were found between the treatment and subgroup terms in the main regression model (p=0.36 to 0.93), indicating a consistent treatment effect across these groups. Preliminary evidence suggests an active video games intervention had a consistent positive effect on body composition among important subgroups. This may support the use of these games as a pragmatic public health intervention to displace sedentary behaviour with physical activity in young people.
Berge, Jerica M.; Meyer, Craig; MacLehose, Richard F.; Crichlow, Renee; Neumark-Sztainer, Dianne
2015-01-01
Objective To examine whether and how parents’ and adolescent siblings’ weight and weight-related behaviors are correlated. Results will inform which family members may be important to include in adolescent obesity prevention interventions. Design and Methods Data from two linked population-based studies, EAT 2010 and F-EAT, were used for cross-sectional analyses. Parents (n=58; 91% females; mean age=41.7 years) and adolescent siblings (sibling #1 n=58, 50% girls, mean age=14.3 years; sibling #2 n=58, 64% Girls, mean age=14.8) were socioeconomically and racially/ethnically diverse. Results Some weight-related behaviors between adolescent siblings were significantly positively correlated (i.e., fast food consumption, breakfast frequency, sedentary patterns, p<0.05). There were no significant correlations between parent weight and weight-related behaviors and adolescent siblings’ same behaviors. Some of the significant correlations found between adolescent siblings’ weight-related behaviors were statistically different from correlations between parents’ and adolescent siblings’ weight-related behaviors. Conclusions Although not consistently, adolescent siblings’ weight-related behaviors were significantly correlated as compared to parents’ and adolescent siblings’ weight-related behaviors. It may be important to consider including siblings in adolescent obesity prevention interventions or in recommendations healthcare providers give to adolescents regarding their weight and weight-related behaviors. PMID:25820257
Frasquilho, Diana; de Matos, Margarida Gaspar; Marques, Adilson; Neville, Fergus G; Gaspar, Tânia; Caldas-de-Almeida, J M
2016-10-01
We investigated, in a sample of 112 unemployed parents of adolescents aged 10-19 years, the links between parental distress and change in youth emotional problems related to parental unemployment, and the moderation roles of parent-youth relationship and financial deprivation. Data were analyzed using descriptive statistics and correlations. Further, simple moderation, additive moderation, and moderated moderation models of regression were performed to analyze the effects of parental distress, parent-youth relationship and financial deprivation in predicting change in youth emotional problems related to parental unemployment. Results show that parental distress moderated by parent-youth relationship predicted levels of change in youth emotional problems related to parental unemployment. This study provides evidence that during job loss, parental distress is linked to youth emotional well-being and that parent-youth relationships play an important moderation role. This raises the importance of further researching parental distress impacts on youth well-being, especially during periods of high unemployment rates.
Dupuy, Céline; Morignat, Eric; Maugey, Xavier; Vinard, Jean-Luc; Hendrikx, Pascal; Ducrot, Christian; Calavas, Didier; Gay, Emilie
2013-04-30
The slaughterhouse is a central processing point for food animals and thus a source of both demographic data (age, breed, sex) and health-related data (reason for condemnation and condemned portions) that are not available through other sources. Using these data for syndromic surveillance is therefore tempting. However many possible reasons for condemnation and condemned portions exist, making the definition of relevant syndromes challenging.The objective of this study was to determine a typology of cattle with at least one portion of the carcass condemned in order to define syndromes. Multiple factor analysis (MFA) in combination with clustering methods was performed using both health-related data and demographic data. Analyses were performed on 381,186 cattle with at least one portion of the carcass condemned among the 1,937,917 cattle slaughtered in ten French abattoirs. Results of the MFA and clustering methods led to 12 clusters considered as stable according to year of slaughter and slaughterhouse. One cluster was specific to a disease of public health importance (cysticercosis). Two clusters were linked to the slaughtering process (fecal contamination of heart or lungs and deterioration lesions). Two clusters respectively characterized by chronic liver lesions and chronic peritonitis could be linked to diseases of economic importance to farmers. Three clusters could be linked respectively to reticulo-pericarditis, fatty liver syndrome and farmer's lung syndrome, which are related to both diseases of economic importance to farmers and herd management issues. Three clusters respectively characterized by arthritis, myopathy and Dark Firm Dry (DFD) meat could notably be linked to animal welfare issues. Finally, one cluster, characterized by bronchopneumonia, could be linked to both animal health and herd management issues. The statistical approach of combining multiple factor analysis with cluster analysis showed its relevance for the detection of syndromes using available large and complex slaughterhouse data. The advantages of this statistical approach are to i) define groups of reasons for condemnation based on meat inspection data, ii) help grouping reasons for condemnation among a list of various possible reasons for condemnation for which a consensus among experts could be difficult to reach, iii) assign each animal to a single syndrome which allows the detection of changes in trends of syndromes to detect unusual patterns in known diseases and emergence of new diseases.
NASA Astrophysics Data System (ADS)
Funk, C. C.; Shukla, S.; Hoerling, M. P.; Robertson, F. R.; Hoell, A.; Liebmann, B.
2013-12-01
During boreal spring, eastern portions of Kenya and Somalia have experienced more frequent droughts since 1999. Given the region's high levels of food insecurity, better predictions of these droughts could provide substantial humanitarian benefits. We show that dynamical-statistical seasonal climate forecasts, based on the latest generation of coupled atmosphere-ocean and uncoupled atmospheric models, effectively predict boreal spring rainfall in this area. Skill sources are assessed by comparing ensembles driven with full-ocean forcing with ensembles driven with ENSO-only sea surface temperatures (SSTs). Our analysis suggests that both ENSO and non-ENSO Indo-Pacific SST forcing have played an important role in the increase in drought frequencies. Over the past 30 years, La Niña drought teleconnections have strengthened, while non-ENSO Indo-Pacific convection patterns have also supported increased (decreased) Western Pacific (East African) rainfall. To further examine the relative contribution of ENSO, low frequency warming and the Pacific Decadal Oscillation, we present decompositions of ECHAM5, GFS, CAM4 and GMAO AMIP simulations. These decompositions suggest that rapid warming in the western Pacific and steeper western-to-central Pacific SST gradients have likely played an important role in the recent intensification of the Walker circulation, and the associated increase in East African aridity. A linear combination of time series describing the Pacific Decadal Oscillation and the strength of Indo-Pacific warming are shown to track East African rainfall reasonably well. The talk concludes with a few thoughts linking the potentially important interplay of attribution and prediction. At least for recent East African droughts, it appears that a characteristic Indo-Pacific SST and precipitation anomaly pattern can be linked statistically to support forecasts and attribution analyses. The combination of traditional AGCM attribution analyses with simple yet physically plausible statistical estimation procedures may help us better untangle some climate mysteries.
Some Variables in Relation to Students' Anxiety in Learning Statistics.
ERIC Educational Resources Information Center
Sutarso, Toto
The purpose of this study was to investigate some variables that relate to students' anxiety in learning statistics. The variables included sex, class level, students' achievement, school, mathematical background, previous statistics courses, and race. The instrument used was the 24-item Students' Attitudes Toward Statistics (STATS), which was…
Statistical Learning Is Related to Early Literacy-Related Skills
ERIC Educational Resources Information Center
Spencer, Mercedes; Kaschak, Michael P.; Jones, John L.; Lonigan, Christopher J.
2015-01-01
It has been demonstrated that statistical learning, or the ability to use statistical information to learn the structure of one's environment, plays a role in young children's acquisition of linguistic knowledge. Although most research on statistical learning has focused on language acquisition processes, such as the segmentation of words from…
Testing of Hypothesis in Equivalence and Non Inferiority Trials-A Concept.
Juneja, Atul; Aggarwal, Abha R; Adhikari, Tulsi; Pandey, Arvind
2016-04-01
Establishing the appropriate hypothesis is one of the important steps for carrying out the statistical tests/analysis. Its understanding is important for interpreting the results of statistical analysis. The current communication attempts to provide the concept of testing of hypothesis in non inferiority and equivalence trials, where the null hypothesis is just reverse of what is set up for conventional superiority trials. It is similarly looked for rejection for establishing the fact the researcher is intending to prove. It is important to mention that equivalence or non inferiority cannot be proved by accepting the null hypothesis of no difference. Hence, establishing the appropriate statistical hypothesis is extremely important to arrive at meaningful conclusion for the set objectives in research.
Wagner, Tyler; Irwin, Brian J.; James R. Bence,; Daniel B. Hayes,
2016-01-01
Monitoring to detect temporal trends in biological and habitat indices is a critical component of fisheries management. Thus, it is important that management objectives are linked to monitoring objectives. This linkage requires a definition of what constitutes a management-relevant “temporal trend.” It is also important to develop expectations for the amount of time required to detect a trend (i.e., statistical power) and for choosing an appropriate statistical model for analysis. We provide an overview of temporal trends commonly encountered in fisheries management, review published studies that evaluated statistical power of long-term trend detection, and illustrate dynamic linear models in a Bayesian context, as an additional analytical approach focused on shorter term change. We show that monitoring programs generally have low statistical power for detecting linear temporal trends and argue that often management should be focused on different definitions of trends, some of which can be better addressed by alternative analytical approaches.
Ellis, Hugh; Schoenberger, Erica
2017-01-01
According to the most recent estimates, 842,000 deaths in low- to middle-income countries were attributable to inadequate water, sanitation and hygiene in 2012. Despite billions of dollars and decades of effort, we still lack a sound understanding of which kinds of WASH interventions are most effective in improving public health outcomes, and an important corollary-whether the right things are being measured. The World Health Organization (WHO) has made a concerted effort to compile comprehensive data on drinking water quality and sanitation in the developing world. A recent 2014 report provides information on three phenotypes (responses): Unsafe Water Deaths, Unsafe Sanitation Deaths, Unsafe Hygiene Deaths; two grouped phenotypes: Unsafe Water and Sanitation Deaths and Unsafe Water, Sanitation and Hygiene Deaths; and six explanatory variables (predictors): Improved Sanitation, Unimproved Water Source, Piped Water To Premises, Other Improved Water Source, Filtered and Bottled Water in the Household and Handwashing. Regression analyses were performed to identify statistically significant associations between these mortality responses and predictors. Good fitted-model performance required: (1) the use of population-normalized death fractions as opposed to number of deaths; (2) transformed response (logit or power); and (3) square-root predictor transformation. Given the complexity and heterogeneity of the relationships and countries being studied, these models exhibited remarkable performance and explained, for example, about 85% of the observed variance in population-normalized Unsafe Sanitation Death fraction, with a high F-statistic and highly statistically significant predictor p-values. Similar performance was found for all other responses, which was an unexpected result (the expected associations between responses and predictors-i.e., water-related with water-related, etc. did not occur). The set of statistically significant predictors remains the same across all responses. That is, Unsafe Water Source (UWS), Improved Sanitation (IS) and Filtered and Bottled Water in the Household (FBH) were the only statistically significant predictors whether the response was Unsafe Sanitation Death Fraction, Unsafe Hygiene Death Fraction or Unsafe Water Death Fraction. Moreover, the fraction of variance explained for all fitted models remained relatively high (adjusted R2 ranges from 0.7605 to 0.8533). We find that two of the statistically significant predictors-Improved Sanitation and Unimproved Water Sources-are particularly influential. We also find that some predictors (Piped Water to Premises, Other Improved Water Sources) have very little explanatory power for predicting mortality and one (Other Improved Water Sources) has a counterintuitive effect on response (Unsafe Sanitary Death Fraction increases with increases in OIWS) and one predictor (Hand Washing) to have essentially no explanatory usefulness. Our results suggest that a higher priority may need to be given to improved sanitation than has been the case. Nevertheless, while our focus in this paper is mortality, morbidity is a staggering consequence of inadequate water, sanitation and hygiene, and lower impact on mortality may not mean a similarly low impact on morbidity. More specifically, those predictors that we found uninfluential for predicting mortality-related responses may indeed be important when morbidity is the response.
Güldenring, Annette
2015-01-01
After explaining the essential trans* terminology, I offer a short historical overview of the way health care has dealt with the subject of gender, trans* and health in different times. In the third section, I compare the world's most important diagnostic manuals, namely the International statistical classification of diseases and related health problems (ICD) and the Diagnostic and statistical manual of mental disorders (DSM), i.e. their criteria for 'gender identity disorders' (ICD-10) and 'gender dysphoria' (DSM-5). The fourth section branch out the factors which influence every diagnostic conception - of no matter whom - in the health care system. The last section discusses the implications resulting from this diagnostic dilemma for the health situation of gender nonconforming people.
Application of the Teager-Kaiser energy operator in bearing fault diagnosis.
Henríquez Rodríguez, Patricia; Alonso, Jesús B; Ferrer, Miguel A; Travieso, Carlos M
2013-03-01
Condition monitoring of rotating machines is important in the prevention of failures. As most machine malfunctions are related to bearing failures, several bearing diagnosis techniques have been developed. Some of them feature the bearing vibration signal with statistical measures and others extract the bearing fault characteristic frequency from the AM component of the vibration signal. In this paper, we propose to transform the vibration signal to the Teager-Kaiser domain and feature it with statistical and energy-based measures. A bearing database with normal and faulty bearings is used. The diagnosis is performed with two classifiers: a neural network classifier and a LS-SVM classifier. Experiments show that the Teager domain features outperform those based on the temporal or AM signal. Copyright © 2012 ISA. Published by Elsevier Ltd. All rights reserved.
X-ray studies of quasars with the Einstein Observatory. IV - X-ray dependence on radio emission
NASA Technical Reports Server (NTRS)
Worrall, D. M.; Tananbaum, H.; Giommi, P.; Zamorani, G.
1987-01-01
The X-ray properties of a sample of 114 radio-loud quasars observed with the Einstein Observatory are examined, and the results are compared with those obtained from a large sample of radio-quiet quasars. The results of statistical analysis of the dependence of X-ray luminosity on combined functions of optical and radio luminosity show that the dependence on both luminosities is important. However, statistically significant differences are found between subsamples of flat radio spectra quasars and steep radio spectra quasars with regard to dependence of X-ray luminosity on only radio luminosity. The data are consistent with radio-loud quasars having a physical component, not directly related to the optical luminosity, which produces the core radio luminosity plus 'extra' X-ray emission.
Conti, A A; Conti, A; Gensini, G F
2006-09-01
Normality characterises in medicine any possible qualitative or quantitative situation whose absence implies an illness or a state of abnormality. The illness concept was first a philosophical one. But the use of mathematics in the study of biological events, which began with Galton (1822-1911) and with Pearson (1857-1936), changed the frame of reference. In the second part of the 19th century mathematics was used to study the distribution of some biological characteristics in the evolution of the species. Around 1900, statistics became the basis for the study of the diffusion of the illnesses. Half a century later statistics made possible the transition from the description of single cases to groups of cases. Even more important is the concept of "normality" in laboratory medicine. In this field the search for the "perfect norm" was, and possibly still is, under way. The widespread use of statistics in the laboratory has allowed the definition, in a certain sense, of a new normality. This is the reason why the term "reference value" has been introduced. However, even the introduction of this new term has merely shifted the problem, and not resolved it.
Koner, Debasish; Barrios, Lizandra; González-Lezana, Tomás; Panda, Aditya N
2014-09-21
A real wave packet based time-dependent method and a statistical quantum method have been used to study the He + NeH(+) (v, j) reaction with the reactant in various ro-vibrational states, on a recently calculated ab initio ground state potential energy surface. Both the wave packet and statistical quantum calculations were carried out within the centrifugal sudden approximation as well as using the exact Hamiltonian. Quantum reaction probabilities exhibit dense oscillatory pattern for smaller total angular momentum values, which is a signature of resonances in a complex forming mechanism for the title reaction. Significant differences, found between exact and approximate quantum reaction cross sections, highlight the importance of inclusion of Coriolis coupling in the calculations. Statistical results are in fairly good agreement with the exact quantum results, for ground ro-vibrational states of the reactant. Vibrational excitation greatly enhances the reaction cross sections, whereas rotational excitation has relatively small effect on the reaction. The nature of the reaction cross section curves is dependent on the initial vibrational state of the reactant and is typical of a late barrier type potential energy profile.
NASA Astrophysics Data System (ADS)
Jiang, Quan; Zhong, Shan; Cui, Jie; Feng, Xia-Ting; Song, Leibo
2016-12-01
We investigated the statistical characteristics and probability distribution of the mechanical parameters of natural rock using triaxial compression tests. Twenty cores of Jinping marble were tested under each different levels of confining stress (i.e., 5, 10, 20, 30, and 40 MPa). From these full stress-strain data, we summarized the numerical characteristics and determined the probability distribution form of several important mechanical parameters, including deformational parameters, characteristic strength, characteristic strains, and failure angle. The statistical proofs relating to the mechanical parameters of rock presented new information about the marble's probabilistic distribution characteristics. The normal and log-normal distributions were appropriate for describing random strengths of rock; the coefficients of variation of the peak strengths had no relationship to the confining stress; the only acceptable random distribution for both Young's elastic modulus and Poisson's ratio was the log-normal function; and the cohesive strength had a different probability distribution pattern than the frictional angle. The triaxial tests and statistical analysis also provided experimental evidence for deciding the minimum reliable number of experimental sample and for picking appropriate parameter distributions to use in reliability calculations for rock engineering.
Smith, W Brad; Cuenca Lara, Rubí Angélica; Delgado Caballero, Carina Edith; Godínez Valdivia, Carlos Isaías; Kapron, Joseph S; Leyva Reyes, Juan Carlos; Meneses Tovar, Carmen Lourdes; Miles, Patrick D; Oswalt, Sonja N; Ramírez Salgado, Mayra; Song, Xilong Alex; Stinson, Graham; Villela Gaytán, Sergio Armando
2018-05-21
Forests cannot be managed sustainably without reliable data to inform decisions. National Forest Inventories (NFI) tend to report national statistics, with sub-national stratification based on domestic ecological classification systems. It is becoming increasingly important to be able to report statistics on ecosystems that span international borders, as global change and globalization expand stakeholders' spheres of concern. The state of a transnational ecosystem can only be properly assessed by examining the entire ecosystem. In global forest resource assessments, it may be useful to break national statistics down by ecosystem, especially for large countries. The Inventory and Monitoring Working Group (IMWG) of the North American Forest Commission (NAFC) has begun developing a harmonized North American Forest Database (NAFD) for managing forest inventory data, enabling consistent, continental-scale forest assessment supporting ecosystem-level reporting and relational queries. The first iteration of the database contains data describing 1.9 billion ha, including 677.5 million ha of forest. Data harmonization is made challenging by the existence of definitions and methodologies tailored to suit national circumstances, emerging from each country's professional forestry development. This paper reports the methods used to synchronize three national forest inventories, starting with a small suite of variables and attributes.
Variation in reaction norms: Statistical considerations and biological interpretation.
Morrissey, Michael B; Liefting, Maartje
2016-09-01
Analysis of reaction norms, the functions by which the phenotype produced by a given genotype depends on the environment, is critical to studying many aspects of phenotypic evolution. Different techniques are available for quantifying different aspects of reaction norm variation. We examine what biological inferences can be drawn from some of the more readily applicable analyses for studying reaction norms. We adopt a strongly biologically motivated view, but draw on statistical theory to highlight strengths and drawbacks of different techniques. In particular, consideration of some formal statistical theory leads to revision of some recently, and forcefully, advocated opinions on reaction norm analysis. We clarify what simple analysis of the slope between mean phenotype in two environments can tell us about reaction norms, explore the conditions under which polynomial regression can provide robust inferences about reaction norm shape, and explore how different existing approaches may be used to draw inferences about variation in reaction norm shape. We show how mixed model-based approaches can provide more robust inferences than more commonly used multistep statistical approaches, and derive new metrics of the relative importance of variation in reaction norm intercepts, slopes, and curvatures. © 2016 The Author(s). Evolution © 2016 The Society for the Study of Evolution.
An introduction to Bayesian statistics in health psychology.
Depaoli, Sarah; Rus, Holly M; Clifton, James P; van de Schoot, Rens; Tiemensma, Jitske
2017-09-01
The aim of the current article is to provide a brief introduction to Bayesian statistics within the field of health psychology. Bayesian methods are increasing in prevalence in applied fields, and they have been shown in simulation research to improve the estimation accuracy of structural equation models, latent growth curve (and mixture) models, and hierarchical linear models. Likewise, Bayesian methods can be used with small sample sizes since they do not rely on large sample theory. In this article, we discuss several important components of Bayesian statistics as they relate to health-based inquiries. We discuss the incorporation and impact of prior knowledge into the estimation process and the different components of the analysis that should be reported in an article. We present an example implementing Bayesian estimation in the context of blood pressure changes after participants experienced an acute stressor. We conclude with final thoughts on the implementation of Bayesian statistics in health psychology, including suggestions for reviewing Bayesian manuscripts and grant proposals. We have also included an extensive amount of online supplementary material to complement the content presented here, including Bayesian examples using many different software programmes and an extensive sensitivity analysis examining the impact of priors.
Rollins, Derrick K; Teh, Ailing
2010-12-17
Microarray data sets provide relative expression levels for thousands of genes for a small number, in comparison, of different experimental conditions called assays. Data mining techniques are used to extract specific information of genes as they relate to the assays. The multivariate statistical technique of principal component analysis (PCA) has proven useful in providing effective data mining methods. This article extends the PCA approach of Rollins et al. to the development of ranking genes of microarray data sets that express most differently between two biologically different grouping of assays. This method is evaluated on real and simulated data and compared to a current approach on the basis of false discovery rate (FDR) and statistical power (SP) which is the ability to correctly identify important genes. This work developed and evaluated two new test statistics based on PCA and compared them to a popular method that is not PCA based. Both test statistics were found to be effective as evaluated in three case studies: (i) exposing E. coli cells to two different ethanol levels; (ii) application of myostatin to two groups of mice; and (iii) a simulated data study derived from the properties of (ii). The proposed method (PM) effectively identified critical genes in these studies based on comparison with the current method (CM). The simulation study supports higher identification accuracy for PM over CM for both proposed test statistics when the gene variance is constant and for one of the test statistics when the gene variance is non-constant. PM compares quite favorably to CM in terms of lower FDR and much higher SP. Thus, PM can be quite effective in producing accurate signatures from large microarray data sets for differential expression between assays groups identified in a preliminary step of the PCA procedure and is, therefore, recommended for use in these applications.
Biomarkers and surrogate endpoints in kidney disease
2015-01-01
Kidney disease and its related comorbidities impose a large public health burden. Despite this, the number of clinical trials in nephrology lags behind many other fields. An important factor contributing to the relatively slow pace of nephrology trials is that existing clinical endpoints have significant limitations. “Hard” endpoints for chronic kidney disease, such as progression to end-stage renal disease, may not be reached for decades. Traditional biomarkers, such as serum creatinine in acute kidney injury, may lack sensitivity and predictive value. Finding new biomarkers to serve as surrogate endpoints is therefore an important priority in kidney disease research and may help to accelerate nephrology clinical trials. In this paper, I first review key concepts related to the selection of clinical trial endpoints and discuss statistical and regulatory considerations related to the evaluation of biomarkers as surrogate endpoints. This is followed by a discussion of the challenges and opportunities in developing novel biomarkers and surrogate endpoints in three major areas of nephrology research: acute kidney injury, chronic kidney disease, and autosomal dominant polycystic kidney disease. PMID:25980469
GAISE 2016 Promotes Statistical Literacy
ERIC Educational Resources Information Center
Schield, Milo
2017-01-01
In the 2005 Guidelines for Assessment and Instruction in Statistics Education (GAISE), statistical literacy featured as a primary goal. The 2016 revision eliminated statistical literacy as a stated goal. Although this looks like a rejection, this paper argues that by including multivariate thinking and--more importantly--confounding as recommended…
Using Guided Reinvention to Develop Teachers' Understanding of Hypothesis Testing Concepts
ERIC Educational Resources Information Center
Dolor, Jason; Noll, Jennifer
2015-01-01
Statistics education reform efforts emphasize the importance of informal inference in the learning of statistics. Research suggests statistics teachers experience similar difficulties understanding statistical inference concepts as students and how teacher knowledge can impact student learning. This study investigates how teachers reinvented an…
Estimating Wartime Support Resource Requirements. Statistical and Related Policy Issues.
1984-07-01
Requirements, Programs & Studies Group (AF/RDQM) July 1984 Ofc, DCS/R&D and Acquisition 13. NUMEtR OF PAGES Ho UAF Wshinton.D.C. 20330______________ 14.m...personnel responsible for producing requirements estimates. This work was conducted as part of the study effort "The Driving 1" Inputs and Assumptions of...represent another important explanation for high support costs. The Air Force has completed a detailed study of the reasons for growth in spares
Silich, Bert A; Yang, James J
2012-05-01
Measuring workplace performance is important to emergency department management. If an unreliable model is used, the results will be inaccurate. Use of inaccurate results to make decisions, such as how to distribute the incentive pay, will lead to rewarding the wrong people and will potentially demoralize top performers. This article demonstrates a statistical model to reliably measure the work accomplished, which can then be used as a performance measurement.
A biological compression model and its applications.
Cao, Minh Duc; Dix, Trevor I; Allison, Lloyd
2011-01-01
A biological compression model, expert model, is presented which is superior to existing compression algorithms in both compression performance and speed. The model is able to compress whole eukaryotic genomes. Most importantly, the model provides a framework for knowledge discovery from biological data. It can be used for repeat element discovery, sequence alignment and phylogenetic analysis. We demonstrate that the model can handle statistically biased sequences and distantly related sequences where conventional knowledge discovery tools often fail.
2012 Workplace and Gender Relations Survey of Active Duty Members: Nonresponse Bias Analysis Report
2014-01-01
Control and Prevention ), or command climate surveys (e.g., DEOCS). 6 Table 1. Comparison of Trends in WGRA and SOFS-A Response Rates (Shown in...DMDC draws optimized samples to reduce survey burden on members as well as produce high levels of precision for important domain estimates by using...statistical significance at α= .05 Because paygrade is a significant predictor of survey response, we next examined the odds ratio of each paygrade levels
Cosmological consistency tests of gravity theory and cosmic acceleration
NASA Astrophysics Data System (ADS)
Ishak-Boushaki, Mustapha B.
2017-01-01
Testing general relativity at cosmological scales and probing the cause of cosmic acceleration are among the important objectives targeted by incoming and future astronomical surveys and experiments. I present our recent results on consistency tests that can provide insights about the underlying gravity theory and cosmic acceleration using cosmological data sets. We use statistical measures, the rate of cosmic expansion, the growth rate of large scale structure, and the physical consistency of these probes with one another.
Semantic Importance Sampling for Statistical Model Checking
2015-01-16
SMT calls while maintaining correctness. Finally, we implement SIS in a tool called osmosis and use it to verify a number of stochastic systems with...2 surveys related work. Section 3 presents background definitions and concepts. Section 4 presents SIS, and Section 5 presents our tool osmosis . In...which I∗M|=Φ(x) = 1. We do this by first randomly selecting a cube c from C∗ with uniform probability since each cube has equal probability 9 5. OSMOSIS
Recurrence time statistics: versatile tools for genomic DNA sequence analysis.
Cao, Yinhe; Tung, Wen-Wen; Gao, J B
2004-01-01
With the completion of the human and a few model organisms' genomes, and the genomes of many other organisms waiting to be sequenced, it has become increasingly important to develop faster computational tools which are capable of easily identifying the structures and extracting features from DNA sequences. One of the more important structures in a DNA sequence is repeat-related. Often they have to be masked before protein coding regions along a DNA sequence are to be identified or redundant expressed sequence tags (ESTs) are to be sequenced. Here we report a novel recurrence time based method for sequence analysis. The method can conveniently study all kinds of periodicity and exhaustively find all repeat-related features from a genomic DNA sequence. An efficient codon index is also derived from the recurrence time statistics, which has the salient features of being largely species-independent and working well on very short sequences. Efficient codon indices are key elements of successful gene finding algorithms, and are particularly useful for determining whether a suspected EST belongs to a coding or non-coding region. We illustrate the power of the method by studying the genomes of E. coli, the yeast S. cervisivae, the nematode worm C. elegans, and the human, Homo sapiens. Computationally, our method is very efficient. It allows us to carry out analysis of genomes on the whole genomic scale by a PC.
Stage-structured transmission of phocine distemper virus in the Dutch 2002 outbreak
Klepac, Petra; Pomeroy, Laura W.; Bjørnstad, Ottar N.; Kuiken, Thijs; Osterhaus, Albert D.M.E.; Rijks, Jolianne M.
2009-01-01
Heterogeneities in transmission among hosts can be very important in shaping infectious disease dynamics. In mammals with strong social organization, such heterogeneities are often structured by functional stage: juveniles, subadults and adults. We investigate the importance of such stage-related heterogeneities in shaping the 2002 phocine distemper virus (PDV) outbreak in the Dutch Wadden Sea, when more than 40 per cent of the harbour seals were killed. We do this by comparing the statistical fit of a hierarchy of models with varying transmission complexity: homogeneous versus heterogeneous mixing and density- versus frequency-dependent transmission. We use the stranding data as a proxy for incidence and use Poisson likelihoods to estimate the ‘who acquires infection from whom’ (WAIFW) matrix. Statistically, the model with strong heterogeneous mixing and density-dependent transmission was found to best describe the transmission dynamics. However, patterns of incidence support a model of frequency-dependent transmission among adults and juveniles. Based on the maximum-likelihood WAIFW matrix estimates, we use the next-generation formalism to calculate an R0 between 2 and 2.5 for the Dutch 2002 PDV epidemic. PMID:19364743
NASA Technical Reports Server (NTRS)
Kiley, C. M.; Fuelberg, Henry E.; Palmer, P. I.; Allen, D. J.; Carmichael, G. R.; Jacob, D. J.; Mari, C.; Pierce, R. B.; Pickering, K. E.; Tang, Y.
2002-01-01
Four global scale and three regional scale chemical transport models are intercompared and evaluated during NASA's TRACE-P experiment. Model simulated and measured CO are statistically analyzed along aircraft flight tracks. Results for the combination of eleven flights show an overall negative bias in simulated CO. Biases are most pronounced during large CO events. Statistical agreements vary greatly among the individual flights. Those flights with the greatest range of CO values tend to be the worst simulated. However, for each given flight, the models generally provide similar relative results. The models exhibit difficulties simulating intense CO plumes. CO error is found to be greatest in the lower troposphere. Convective mass flux is shown to be very important, particularly near emissions source regions. Occasionally meteorological lift associated with excessive model-calculated mass fluxes leads to an overestimation of mid- and upper- tropospheric mixing ratios. Planetary Boundary Layer (PBL) depth is found to play an important role in simulating intense CO plumes. PBL depth is shown to cap plumes, confining heavy pollution to the very lowest levels.
Gender matters: the relationship between social anxiety and alcohol-related consequences.
Schry, Amie R; Norberg, Melissa M; Maddox, Brenna B; White, Susan W
2014-01-01
Identification of risk factors for alcohol-related consequences is an important public health concern. Both gender and social anxiety have been associated with alcohol-related consequences broadly, but it is unknown whether these variables are differentially related to specific types of alcohol-related consequences for American college students. In the present study, 573 undergraduate students (M(age) = 19.86 years, SD = 1.40; range 18 to 25; 68.9% female) completed an on-line assessment of social anxiety, alcohol use, and four types of alcohol-related consequences (personal, social, physical, and role). Poisson regressions were run to examine social anxiety, gender, and the interaction between social anxiety and gender as predictors of each type of alcohol-related consequences. After controlling for alcohol use, social anxiety was positively associated with all four types of consequences, and females endorsed higher rates of physical, personal, and role consequences. The interaction between social anxiety and gender was statistically significant only for physical consequences, with social anxiety having a stronger effect for males. These findings, which diverge somewhat from those of a prior study with Australian college students, are discussed in the context of a biopsychosocial model of social anxiety and substance use problems. This study highlights the importance of further investigating cultural differences in the relationships among social anxiety, gender, and alcohol-related consequences.
Kicks from the penalty mark in soccer: the roles of stress, skill, and fatigue for kick outcomes.
Jordet, Geir; Hartman, Esther; Visscher, Chris; Lemmink, Koen A P M
2007-01-15
The soccer "penalty shootout" in the knock-out phase of major international tournaments is one of the most dramatic events in international soccer. The outcome of these kicks is typically attributed to factors such as psychology (e.g. coping with stress), skill (e.g. kicking technique), physiology (e.g. overcoming the fatigue of 120 min play), and chance (e.g. in what direction the goalkeeper moves). The purpose of this study was to use internet game record data to estimate the relative importance of these performance components for the outcome of penalty kicks in international matches. Data were collected from soccer statistics internet sites on all 41 penalty shootouts and 409 kicks taken in the World Cup, European Championships, and Copa America between 1976 and 2004. The results showed that the importance of the kicks (indicative of stress) was negatively related to the outcomes of the kicks, whereas skill and fatigue were less, or not, related to outcome. It was concluded that psychological components are most influential for the outcome of penalty kicks. We recommend that practitioners work with players to reduce the perceived importance of each kick.
Statistical physics of media processes: Mediaphysics
NASA Astrophysics Data System (ADS)
Kuznetsov, Dmitri V.; Mandel, Igor
2007-04-01
The processes of mass communications in complicated social or sociobiological systems such as marketing, economics, politics, animal populations, etc. as a subject for the special scientific subbranch-“mediaphysics”-are considered in its relation with sociophysics. A new statistical physics approach to analyze these phenomena is proposed. A keystone of the approach is an analysis of population distribution between two or many alternatives: brands, political affiliations, or opinions. Relative distances between a state of a “person's mind” and the alternatives are measures of propensity to buy (to affiliate, or to have a certain opinion). The distribution of population by those relative distances is time dependent and affected by external (economic, social, marketing, natural) and internal (influential propagation of opinions, “word of mouth”, etc.) factors, considered as fields. Specifically, the interaction and opinion-influence field can be generalized to incorporate important elements of Ising-spin-based sociophysical models and kinetic-equation ones. The distributions were described by a Schrödinger-type equation in terms of Green's functions. The developed approach has been applied to a real mass-media efficiency problem for a large company and generally demonstrated very good results despite low initial correlations of factors and the target variable.
Schenk, Linda; Öberg, Mattias
2018-05-28
Records of injuries and incidents provide an important basis for injury prevention related to hazardous substances at the workplace. The present study aimed to review available data on injuries and incidents involving hazardous substances and investigate how data from the Poisons Information Centre could complement the records of the Swedish Work Environment Authority. We found two major obstacles for using injury/incident data based on employers' mandatory reporting. First, it was not possible to quickly and reliably identify injuries caused by hazardous substances, and second, data identifying substances or products are not systematically included. For two out of five investigated injuries with lost working days likely due to chemical injuries, we could not identify substances and/or products involved. The records based on calls to the Poisons Information Centre allow better understanding of chemical hazards and products. Besides the large share of unidentified chemical hazards in the injury statistics, the most striking difference was found for cleaning agents. Cleaning agents were implicated in one-third of the occupational cases that the consulting Poisons Information Centre expert judged to pose a major risk and in need of immediate healthcare. Only one in 10 injuries with lost days reported by employers was related to this type of product. The identification of exposures and symptoms by the Poisons Information Centre allow recognition of chemicals with problematic occupational uses. Hence, these records may serve as an important complement to official injury statistics related to incidents with hazardous substances at work.
Barber, Julie A; Thompson, Simon G
1998-01-01
Objective To review critically the statistical methods used for health economic evaluations in randomised controlled trials where an estimate of cost is available for each patient in the study. Design Survey of published randomised trials including an economic evaluation with cost values suitable for statistical analysis; 45 such trials published in 1995 were identified from Medline. Main outcome measures The use of statistical methods for cost data was assessed in terms of the descriptive statistics reported, use of statistical inference, and whether the reported conclusions were justified. Results Although all 45 trials reviewed apparently had cost data for each patient, only 9 (20%) reported adequate measures of variability for these data and only 25 (56%) gave results of statistical tests or a measure of precision for the comparison of costs between the randomised groups. Only 16 (36%) of the articles gave conclusions which were justified on the basis of results presented in the paper. No paper reported sample size calculations for costs. Conclusions The analysis and interpretation of cost data from published trials reveal a lack of statistical awareness. Strong and potentially misleading conclusions about the relative costs of alternative therapies have often been reported in the absence of supporting statistical evidence. Improvements in the analysis and reporting of health economic assessments are urgently required. Health economic guidelines need to be revised to incorporate more detailed statistical advice. Key messagesHealth economic evaluations required for important healthcare policy decisions are often carried out in randomised controlled trialsA review of such published economic evaluations assessed whether statistical methods for cost outcomes have been appropriately used and interpretedFew publications presented adequate descriptive information for costs or performed appropriate statistical analysesIn at least two thirds of the papers, the main conclusions regarding costs were not justifiedThe analysis and reporting of health economic assessments within randomised controlled trials urgently need improving PMID:9794854
Turking Statistics: Student-Generated Surveys Increase Student Engagement and Performance
ERIC Educational Resources Information Center
Whitley, Cameron T.; Dietz, Thomas
2018-01-01
Thirty years ago, Hubert M. Blalock Jr. published an article in "Teaching Sociology" about the importance of teaching statistics. We honor Blalock's legacy by assessing how using Amazon Mechanical Turk (MTurk) in statistics classes can enhance student learning and increase statistical literacy among social science gradaute students. In…
Benchmarking Usage Statistics in Collection Management Decisions for Serials
ERIC Educational Resources Information Center
Tucker, Cory
2009-01-01
Usage statistics are an important metric for making decisions on serials. Although the University of Nevada, Las Vegas (UNLV) Libraries have been collecting usage statistics, the statistics had not frequently been used to make decisions and had not been included in collection development policy. After undergoing a collection assessment, the…
Two-Year College Mathematics Instructors' Conceptions of Variation
ERIC Educational Resources Information Center
Dabos, Monica Graciela Gandhini
2011-01-01
Statistics education researchers are urging teachers of statistics to help students develop a more sophisticated understanding of variation, since variation is the core of statistics. However, little research has been done into the conceptions of variation held by instructors of statistics. This is of particular importance at the community college…
Play It Again: Teaching Statistics with Monte Carlo Simulation
ERIC Educational Resources Information Center
Sigal, Matthew J.; Chalmers, R. Philip
2016-01-01
Monte Carlo simulations (MCSs) provide important information about statistical phenomena that would be impossible to assess otherwise. This article introduces MCS methods and their applications to research and statistical pedagogy using a novel software package for the R Project for Statistical Computing constructed to lessen the often steep…
Aggregation and folding phase transitions of RNA molecules
NASA Astrophysics Data System (ADS)
Bundschuh, Ralf
2007-03-01
RNA is a biomolecule that is involved in nearly all aspects of cellular functions. In order to perform many of these functions, RNA molecules have to fold into specific secondary structures. This folding is driven by the tendency of the bases to form Watson-Crick base pairs. Beyond the biological importance of RNA, the relatively simple rules for structure formation of RNA make it a very interesting system from the statistical physics point of view. We will present examples of phase transitions in RNA secondary structure formation that are amenable to analytical descriptions. A special focus will be on aggregation between several RNA molecules which is important for some regulatory circuits based on RNA structure, triplet repeat diseases like Huntington's, and as a model for prion diseases. We show that depending on the relative strength of the intramolecular and the intermolecular base pairing, RNA molecules undergo a transition into an aggregated phase and quantitatively characterize this transition.
ERIC Educational Resources Information Center
Metz, Mary Louise
2010-01-01
Statistics education has become an increasingly important component of the mathematics education of today's citizens. In part to address the call for a more statistically literate citizenship, The "Guidelines for Assessment and Instruction in Statistics Education (GAISE)" were developed in 2005 by the American Statistical Association. These…
Moradinazar, Mehdi; Kurd, Nematullah; Farhadi, Rozita; Amee, Vahid; Najafi, Farid
2013-10-01
Work-related injuries are the most important cause of work absence, disability, retirement, mutilation, and even mortality. In Iran a great number of work-related injuries are occurred in construction industry. However, less than 12% of total workers are active in the construction sector. This study aimed to determine the incidence rate of work-related injuries, the type of injuries, and its other determinants among the construction workers of Ilam (Iran). The participants were the workers and staffs working in the construction activities of Ilam in Western Iran. All the recorded injuries and deaths related to the construction workers of Ilam from 2006-2009 were collected from the Bureau of Labor and Social Affairs and then analyzed by the statistical package of SPSS (version 19, for Windows). During 2006 - 2009 in Ilam, 387workers encountered the building accidents. Their mean age was 34.3 years (SD = 12.4). The average annual incidence of work-related injuries among the workers was 8.2 per 1000 workers. Fracture with 275 cases (71%) was the most common outcome of injuries, and slipping and falling with 77 cases (36%) were the most important events and exposures. The most important factor related to injuries was the lack of surveillance by employers which was also related with the severity of accident-induced injuries (P < 0.004). Considering the effectiveness of the relevant preventive measures activities such as training the workers as well as using safety tools and more surveillance by employers can decrease the number of work-related injuries among constructive workers.
High-precision surface analysis of the roughness of Michelangelo's David
NASA Astrophysics Data System (ADS)
Fontana, Raffaella; Gambino, Maria Chiara; Greco, Marinella; Marras, Luciano; Materazzi, Marzia; Pampaloni, Enrico; Pezzati, Luca
2003-10-01
The knowledge of the shape of an artwork is an important element for its study and conservation. When dealing with a statue, roughness measurement is a very useful contribution to document its surface conditions, to assess either changes due to restoration intervention or surface decays due to wearing agents, and to monitor its time-evolution in terms of shape variations. In this work we present the preliminary results of the statistical analysis carried out on acquired data relative to six areas of the Michelangelo"s David marble statue, representative of differently degraded surfaces. Determination of the roughness and its relative characteristic wavelength is shown.
Ing, Alex; Schwarzbauer, Christian
2014-01-01
Functional connectivity has become an increasingly important area of research in recent years. At a typical spatial resolution, approximately 300 million connections link each voxel in the brain with every other. This pattern of connectivity is known as the functional connectome. Connectivity is often compared between experimental groups and conditions. Standard methods used to control the type 1 error rate are likely to be insensitive when comparisons are carried out across the whole connectome, due to the huge number of statistical tests involved. To address this problem, two new cluster based methods--the cluster size statistic (CSS) and cluster mass statistic (CMS)--are introduced to control the family wise error rate across all connectivity values. These methods operate within a statistical framework similar to the cluster based methods used in conventional task based fMRI. Both methods are data driven, permutation based and require minimal statistical assumptions. Here, the performance of each procedure is evaluated in a receiver operator characteristic (ROC) analysis, utilising a simulated dataset. The relative sensitivity of each method is also tested on real data: BOLD (blood oxygen level dependent) fMRI scans were carried out on twelve subjects under normal conditions and during the hypercapnic state (induced through the inhalation of 6% CO2 in 21% O2 and 73%N2). Both CSS and CMS detected significant changes in connectivity between normal and hypercapnic states. A family wise error correction carried out at the individual connection level exhibited no significant changes in connectivity.
Ing, Alex; Schwarzbauer, Christian
2014-01-01
Functional connectivity has become an increasingly important area of research in recent years. At a typical spatial resolution, approximately 300 million connections link each voxel in the brain with every other. This pattern of connectivity is known as the functional connectome. Connectivity is often compared between experimental groups and conditions. Standard methods used to control the type 1 error rate are likely to be insensitive when comparisons are carried out across the whole connectome, due to the huge number of statistical tests involved. To address this problem, two new cluster based methods – the cluster size statistic (CSS) and cluster mass statistic (CMS) – are introduced to control the family wise error rate across all connectivity values. These methods operate within a statistical framework similar to the cluster based methods used in conventional task based fMRI. Both methods are data driven, permutation based and require minimal statistical assumptions. Here, the performance of each procedure is evaluated in a receiver operator characteristic (ROC) analysis, utilising a simulated dataset. The relative sensitivity of each method is also tested on real data: BOLD (blood oxygen level dependent) fMRI scans were carried out on twelve subjects under normal conditions and during the hypercapnic state (induced through the inhalation of 6% CO2 in 21% O2 and 73%N2). Both CSS and CMS detected significant changes in connectivity between normal and hypercapnic states. A family wise error correction carried out at the individual connection level exhibited no significant changes in connectivity. PMID:24906136
Matsumura, Mina; Nakayama, Takuto; Sozu, Takashi
2016-01-01
A survey of introductory statistics courses at Japanese medical schools was published as a report in 2014. To obtain a complete understanding of the way in which statistics is taught at the university level in Japan, it is important to extend this survey to related fields, including pharmacy, dentistry, and nursing. The current study investigates the introductory statistics courses offered by faculties of pharmaceutical sciences (six-year programs) at Japanese universities, comparing the features of these courses with those studied in the survey of medical schools. We collected relevant data from the online syllabi of statistics courses published on the websites of 71 universities. The survey items included basic course information (for example, the course names, the targeted student grades, the number of credits, and course classification), textbooks, handouts, the doctoral subject and employment status of each lecturer, and course contents. The period surveyed was July-September 2015. We found that these 71 universities offered a total of 128 statistics courses. There were 67 course names, the most common of which was "biostatistics (iryou toukeigaku)." About half of the courses were designed for first- or second-year students. Students earned fewer than two credits. There were 62 different types of textbooks. The lecturers held doctoral degrees in 18 different subjects, the most common being a doctorate in pharmacy or science. Some course content differed, reflecting the lecturers' academic specialties. The content of introductory statistics courses taught in pharmaceutical science programs also differed slightly from the equivalent content taught in medical schools.
Accounting for measurement error: a critical but often overlooked process.
Harris, Edward F; Smith, Richard N
2009-12-01
Due to instrument imprecision and human inconsistencies, measurements are not free of error. Technical error of measurement (TEM) is the variability encountered between dimensions when the same specimens are measured at multiple sessions. A goal of a data collection regimen is to minimise TEM. The few studies that actually quantify TEM, regardless of discipline, report that it is substantial and can affect results and inferences. This paper reviews some statistical approaches for identifying and controlling TEM. Statistically, TEM is part of the residual ('unexplained') variance in a statistical test, so accounting for TEM, which requires repeated measurements, enhances the chances of finding a statistically significant difference if one exists. The aim of this paper was to review and discuss common statistical designs relating to types of error and statistical approaches to error accountability. This paper addresses issues of landmark location, validity, technical and systematic error, analysis of variance, scaled measures and correlation coefficients in order to guide the reader towards correct identification of true experimental differences. Researchers commonly infer characteristics about populations from comparatively restricted study samples. Most inferences are statistical and, aside from concerns about adequate accounting for known sources of variation with the research design, an important source of variability is measurement error. Variability in locating landmarks that define variables is obvious in odontometrics, cephalometrics and anthropometry, but the same concerns about measurement accuracy and precision extend to all disciplines. With increasing accessibility to computer-assisted methods of data collection, the ease of incorporating repeated measures into statistical designs has improved. Accounting for this technical source of variation increases the chance of finding biologically true differences when they exist.
Development of the Statistical Reasoning in Biology Concept Inventory (SRBCI).
Deane, Thomas; Nomme, Kathy; Jeffery, Erica; Pollock, Carol; Birol, Gülnur
2016-01-01
We followed established best practices in concept inventory design and developed a 12-item inventory to assess student ability in statistical reasoning in biology (Statistical Reasoning in Biology Concept Inventory [SRBCI]). It is important to assess student thinking in this conceptual area, because it is a fundamental requirement of being statistically literate and associated skills are needed in almost all walks of life. Despite this, previous work shows that non-expert-like thinking in statistical reasoning is common, even after instruction. As science educators, our goal should be to move students along a novice-to-expert spectrum, which could be achieved with growing experience in statistical reasoning. We used item response theory analyses (the one-parameter Rasch model and associated analyses) to assess responses gathered from biology students in two populations at a large research university in Canada in order to test SRBCI's robustness and sensitivity in capturing useful data relating to the students' conceptual ability in statistical reasoning. Our analyses indicated that SRBCI is a unidimensional construct, with items that vary widely in difficulty and provide useful information about such student ability. SRBCI should be useful as a diagnostic tool in a variety of biology settings and as a means of measuring the success of teaching interventions designed to improve statistical reasoning skills. © 2016 T. Deane et al. CBE—Life Sciences Education © 2016 The American Society for Cell Biology. This article is distributed by The American Society for Cell Biology under license from the author(s). It is available to the public under an Attribution–Noncommercial–Share Alike 3.0 Unported Creative Commons License (http://creativecommons.org/licenses/by-nc-sa/3.0).
Generating survival times to simulate Cox proportional hazards models with time-varying covariates.
Austin, Peter C
2012-12-20
Simulations and Monte Carlo methods serve an important role in modern statistical research. They allow for an examination of the performance of statistical procedures in settings in which analytic and mathematical derivations may not be feasible. A key element in any statistical simulation is the existence of an appropriate data-generating process: one must be able to simulate data from a specified statistical model. We describe data-generating processes for the Cox proportional hazards model with time-varying covariates when event times follow an exponential, Weibull, or Gompertz distribution. We consider three types of time-varying covariates: first, a dichotomous time-varying covariate that can change at most once from untreated to treated (e.g., organ transplant); second, a continuous time-varying covariate such as cumulative exposure at a constant dose to radiation or to a pharmaceutical agent used for a chronic condition; third, a dichotomous time-varying covariate with a subject being able to move repeatedly between treatment states (e.g., current compliance or use of a medication). In each setting, we derive closed-form expressions that allow one to simulate survival times so that survival times are related to a vector of fixed or time-invariant covariates and to a single time-varying covariate. We illustrate the utility of our closed-form expressions for simulating event times by using Monte Carlo simulations to estimate the statistical power to detect as statistically significant the effect of different types of binary time-varying covariates. This is compared with the statistical power to detect as statistically significant a binary time-invariant covariate. Copyright © 2012 John Wiley & Sons, Ltd.
Statistical process control methods allow the analysis and improvement of anesthesia care.
Fasting, Sigurd; Gisvold, Sven E
2003-10-01
Quality aspects of the anesthetic process are reflected in the rate of intraoperative adverse events. The purpose of this report is to illustrate how the quality of the anesthesia process can be analyzed using statistical process control methods, and exemplify how this analysis can be used for quality improvement. We prospectively recorded anesthesia-related data from all anesthetics for five years. The data included intraoperative adverse events, which were graded into four levels, according to severity. We selected four adverse events, representing important quality and safety aspects, for statistical process control analysis. These were: inadequate regional anesthesia, difficult emergence from general anesthesia, intubation difficulties and drug errors. We analyzed the underlying process using 'p-charts' for statistical process control. In 65,170 anesthetics we recorded adverse events in 18.3%; mostly of lesser severity. Control charts were used to define statistically the predictable normal variation in problem rate, and then used as a basis for analysis of the selected problems with the following results: Inadequate plexus anesthesia: stable process, but unacceptably high failure rate; Difficult emergence: unstable process, because of quality improvement efforts; Intubation difficulties: stable process, rate acceptable; Medication errors: methodology not suited because of low rate of errors. By applying statistical process control methods to the analysis of adverse events, we have exemplified how this allows us to determine if a process is stable, whether an intervention is required, and if quality improvement efforts have the desired effect.
31 CFR 9.5 - Applications for investigation.
Code of Federal Regulations, 2010 CFR
2010-07-01
..., both past and current. (c) Statistical material presented should be on a calendar-year basis for... domestic industry concerned with the article in question. (4) Pertinent statistics showing the quantities... competition created by imports of the article in question. (6) The effect, if any, of imports of the article...
Environmental and Energy Aspects of Construction Industry and Green Buildings
NASA Astrophysics Data System (ADS)
Kauskale, L.; Geipele, I.; Zeltins, N.; Lecis, I.
2017-04-01
Green building is an important component of sustainable real estate market development, and one of the reasons is that the construction industry consumes a high amount of resources. Energy consumption of construction industry results in greenhouse gas emissions, so green buildings, energy systems, building technologies and other aspects play an important role in sustainable development of real estate market, construction and environmental development. The aim of the research is to analyse environmental aspects of sustainable real estate market development, focusing on importance of green buildings at the industry level and related energy aspects. Literature review, historical, statistical data analysis and logical access methods have been used in the research. The conducted research resulted in high environmental rationale and importance of environment-friendly buildings, and there are many green building benefits during the building life cycle. Future research direction is environmental information process and its models.
TURBULENCE-INDUCED RELATIVE VELOCITY OF DUST PARTICLES. IV. THE COLLISION KERNEL
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pan, Liubin; Padoan, Paolo, E-mail: lpan@cfa.harvard.edu, E-mail: ppadoan@icc.ub.edu
Motivated by its importance for modeling dust particle growth in protoplanetary disks, we study turbulence-induced collision statistics of inertial particles as a function of the particle friction time, τ{sub p}. We show that turbulent clustering significantly enhances the collision rate for particles of similar sizes with τ{sub p} corresponding to the inertial range of the flow. If the friction time, τ{sub p,} {sub h}, of the larger particle is in the inertial range, the collision kernel per unit cross section increases with increasing friction time, τ{sub p,} {sub l}, of the smaller particle and reaches the maximum at τ{sub p,}more » {sub l} = τ{sub p,} {sub h}, where the clustering effect peaks. This feature is not captured by the commonly used kernel formula, which neglects the effect of clustering. We argue that turbulent clustering helps alleviate the bouncing barrier problem for planetesimal formation. We also investigate the collision velocity statistics using a collision-rate weighting factor to account for higher collision frequency for particle pairs with larger relative velocity. For τ{sub p,} {sub h} in the inertial range, the rms relative velocity with collision-rate weighting is found to be invariant with τ{sub p,} {sub l} and scales with τ{sub p,} {sub h} roughly as ∝ τ{sub p,h}{sup 1/2}. The weighting factor favors collisions with larger relative velocity, and including it leads to more destructive and less sticking collisions. We compare two collision kernel formulations based on spherical and cylindrical geometries. The two formulations give consistent results for the collision rate and the collision-rate weighted statistics, except that the spherical formulation predicts more head-on collisions than the cylindrical formulation.« less
Zaninovich, Orel A; Ramey, Wyatt L; Walter, Christina M; Dumont, Travis M
2017-10-01
The circle of Willis (CoW) is the foremost anastomosis and blood distribution center of the brain. Its effectiveness depends on its completion and the size and patency of its vessels. Gender-related and age-related anatomic variations in the CoW may play an important role in the pathogenesis of cerebrovascular diseases. In this study, we analyzed computed tomography angiograms (CTAs) to assess for differences in CoW completion related to gender, age, and indication for CTA. A total of 834 CTAs were retrospectively analyzed for all CoW vessels to compare the incidence of complete CoW and variation frequency based on gender, age, and indication. The incidence of complete CoW was 37.1% overall. CoW completion showed a statistically significant decrease with increasing age for all age groups in both men (47.0%, 29.4%, 18.8%) and women (59.1%, 44.2%, 30.9%). Completion was greater in women (43.8%) than in men (31.2%) overall and for all age groups. These gender differences were all statistically significant except for the 18-39 years age group. The most frequent of the 28 CoW variations were absent posterior communicating artery (PCOM) bilaterally (17.1%), right PCOM (15.3%), and left PCOM (10.9%). Ischemic stroke and the 18-39 years age group of hemorrhagic stroke showed a statistically significant reduction in completion relative to trauma. The incidence of complete CoW is likely greater in women for all age groups and likely decreases with age in both genders. The most frequently absent vessel is likely the PCOM, either unilaterally or bilaterally. Completion may play a role in ischemic stroke and a subset of patients with hemorrhagic stroke. Copyright © 2017 Elsevier Inc. All rights reserved.
On the meaning of the weighted alternative free-response operating characteristic figure of merit.
Chakraborty, Dev P; Zhai, Xuetong
2016-05-01
The free-response receiver operating characteristic (FROC) method is being increasingly used to evaluate observer performance in search tasks. Data analysis requires definition of a figure of merit (FOM) quantifying performance. While a number of FOMs have been proposed, the recommended one, namely, the weighted alternative FROC (wAFROC) FOM, is not well understood. The aim of this work is to clarify the meaning of this FOM by relating it to the empirical area under a proposed wAFROC curve. The weighted wAFROC FOM is defined in terms of a quasi-Wilcoxon statistic that involves weights, coding the clinical importance, assigned to each lesion. A new wAFROC curve is proposed, the y-axis of which incorporates the weights, giving more credit for marking clinically important lesions, while the x-axis is identical to that of the AFROC curve. An expression is derived relating the area under the empirical wAFROC curve to the wAFROC FOM. Examples are presented with small numbers of cases showing how AFROC and wAFROC curves are affected by correct and incorrect decisions and how the corresponding FOMs credit or penalize these decisions. The wAFROC, AFROC, and inferred ROC FOMs were applied to three clinical data sets involving multiple reader FROC interpretations in different modalities. It is shown analytically that the area under the empirical wAFROC curve equals the wAFROC FOM. This theorem is the FROC analog of a well-known theorem developed in 1975 for ROC analysis, which gave meaning to a Wilcoxon statistic based ROC FOM. A similar equivalence applies between the area under the empirical AFROC curve and the AFROC FOM. The examples show explicitly that the wAFROC FOM gives equal importance to all diseased cases, regardless of the number of lesions, a desirable statistical property not shared by the AFROC FOM. Applications to the clinical data sets show that the wAFROC FOM yields results comparable to that using the AFROC FOM. The equivalence theorem gives meaning to the weighted AFROC FOM, namely, it is identical to the empirical area under weighted AFROC curve.
Funkenbusch, Paul D; Rotella, Mario; Ercoli, Carlo
2015-04-01
Laboratory studies of tooth preparation are often performed under a limited range of conditions involving single values for all variables other than the 1 being tested. In contrast, in clinical settings not all variables can be tightly controlled. For example, a new dental rotary cutting instrument may be tested in the laboratory by making a specific cut with a fixed force, but in clinical practice, the instrument must make different cuts with individual dentists applying a range of different forces. Therefore, the broad applicability of laboratory results to diverse clinical conditions is uncertain and the comparison of effects across studies is difficult. The purpose of this study was to examine the effect of 9 process variables on dental cutting in a single experiment, allowing each variable to be robustly tested over a range of values for the other 8 and permitting a direct comparison of the relative importance of each on the cutting process. The effects of 9 key process variables on the efficiency of a simulated dental cutting operation were measured. A fractional factorial experiment was conducted by using a computer-controlled, dedicated testing apparatus to simulate dental cutting procedures and Macor blocks as the cutting substrate. Analysis of Variance (ANOVA) was used to judge the statistical significance (α=.05). Five variables consistently produced large, statistically significant effects (target applied load, cut length, starting rpm, diamond grit size, and cut type), while 4 variables produced relatively small, statistically insignificant effects (number of cooling ports, rotary cutting instrument diameter, disposability, and water flow rate). The control exerted by the dentist, simulated in this study by targeting a specific level of applied force, was the single most important factor affecting cutting efficiency. Cutting efficiency was also significantly affected by factors simulating patient/clinical circumstances as well as hardware choices. These results highlight the importance of local clinical conditions (procedure, dentist) in understanding dental cutting procedures and in designing adequate experimental methodologies for future studies. Copyright © 2015 Editorial Council for the Journal of Prosthetic Dentistry. Published by Elsevier Inc. All rights reserved.
Al-Badriyeh, Daoud; Alameri, Marwah; Al-Okka, Randa
2017-01-01
Objective To perform a first-time analysis of the cost-effectiveness (CE) literature on chemotherapies, of all types, in cancer, in terms of trends and change over time, including the influence of industry funding. Design Systematic review. Setting A wide range of cancer-related research settings within healthcare, including health systems, hospitals and medical centres. Participants All literature comparative CE research of drug-based cancer therapies in the period 1986 to 2015. Primary and secondary outcome measures Primary outcomes are the literature trends in relation to journal subject category, authorship, research design, data sources, funds and consultation involvement. An additional outcome measure is the association between industry funding and study outcomes. Analysis Descriptive statistics and the χ2, Fisher exact or Somer's D tests were used to perform non-parametric statistics, with a p value of <0.05 as the statistical significance measure. Results Total 574 publications were analysed. The drug-related CE literature expands over time, with increased publishing in the healthcare sciences and services journal subject category (p<0.001). The retrospective data collection in studies increased over time (p<0.001). The usage of prospective data, however, has been decreasing (p<0.001) in relation to randomised clinical trials (RCTs), but is unchanging for non-RCT studies. The industry-sponsored CE studies have especially been increasing (p<0.001), in contrast to those sponsored by other sources. While paid consultation involvement grew throughout the years, the declaration of funding for this is relatively limited. Importantly, there is evidence that industry funding is associated with favourable result to the sponsor (p<0.001). Conclusions This analysis demonstrates clear trends in how the CE cancer research is presented to the practicing community, including in relation to journals, study designs, authorship and consultation, together with increased financial sponsorship by pharmaceutical industries, which may be more influencing study outcomes than other funding sources. PMID:28131999
3 CFR - Enhanced Collection of Relevant Data and Statistics Relating to Women
Code of Federal Regulations, 2012 CFR
2012-01-01
... 3 The President 1 2012-01-01 2012-01-01 false Enhanced Collection of Relevant Data and Statistics Relating to Women Presidential Documents Other Presidential Documents Memorandum of March 4, 2011 Enhanced Collection of Relevant Data and Statistics Relating to Women Memorandum for the Heads of Executive Departments and Agencies I am proud to work...
Code of Federal Regulations, 2014 CFR
2014-07-01
... relating to a hurricane, earthquake, or other natural occurrence? 250.192 Section 250.192 Mineral Resources... What reports and statistics must I submit relating to a hurricane, earthquake, or other natural..., such as a hurricane, a tropical storm, or an earthquake. Statistics include facilities and rigs...
Code of Federal Regulations, 2012 CFR
2012-07-01
... relating to a hurricane, earthquake, or other natural occurrence? 250.192 Section 250.192 Mineral Resources... What reports and statistics must I submit relating to a hurricane, earthquake, or other natural..., such as a hurricane, a tropical storm, or an earthquake. Statistics include facilities and rigs...
Code of Federal Regulations, 2013 CFR
2013-07-01
... relating to a hurricane, earthquake, or other natural occurrence? 250.192 Section 250.192 Mineral Resources... What reports and statistics must I submit relating to a hurricane, earthquake, or other natural..., such as a hurricane, a tropical storm, or an earthquake. Statistics include facilities and rigs...
Millet, Juan Pablo; Garcia de Olalla, Patricia; Carrillo-Santisteve, Paloma; Gascón, Joaquim; Treviño, Begoña; Muñoz, José; Gómez i Prat, Jordi; Cabezos, Juan; González Cordón, Anna; Caylà, Joan A
2008-01-01
Background International travel and migration have been related with an increase of imported malaria cases. There has been considerable immigration to Barcelona from low-income countries (LIC) in recent years. The objective is to describe the epidemiology and to determine the trends of the disease in Barcelona. Methods Analysis of the cases notified among city residents between 1989 and 2005. Patients were classified as: tourists, voluntary workers, resident immigrants (visiting friends and relatives, VFR) and recently arrived immigrants. An analysis was conducted using the chi2 test and comparison of means. As a measure of association we calculated the Relative Risk (RR) and Odds Ratio (OR) with a Confidence Interval of 95% (CI) and carried out a trends analysis. Results Of the total of 1,579 imported cases notified, 997 (63.1%) lived in Barcelona city, and 55.1% were male. The mean age of patients was 32.7 years. The incidence increased from 2.4 cases/100,000 in 1989 to 3.5 cases/100,000 in 2005 (RR 1.46 CI:1.36–1.55). This increase was not statistically significant (trends analysis, p = 0.36). In terms of reason for travelling, 40.7% were VFR, 33.6% tourists, 12.1% voluntary workers and 13.6% were recently arrived immigrants. The most frequent species found was Plasmodium falciparum (71.3%), mainly in visitors to Africa (OR = 2.3, CI = 1.7–3.2). The vast majority (82.2%) had had some contact with Africa (35.9% with Equatorial Guinea, a Spanish ex-colony) and 96.6% had not completed chemoprophylaxis. Six deaths were observed, all tourists who had travelled to Africa and not taken chemoprophylaxis (3.9% fatality rate). Conclusion Over the period studied there is an increase in malaria incidence, however the trend is not statistically significant. Lack of chemoprophylaxis compliance and the association between Africa and P. falciparum are very clear in the imported cases. Most of the patients with malaria did not take chemoprophylaxis. PMID:18397524
Medical students' attitudes towards science and gross anatomy, and the relationship to personality
Plaisant, Odile; Stephens, Shiby; Apaydin, Nihal; Courtois, Robert; Lignier, Baptiste; Loukas, Marios; Moxham, Bernard
2014-01-01
Assessment of the personalities of medical students can enable medical educators to formulate strategies for the best development of academic and clinical competencies. Previous research has shown that medical students do not share a common personality profile, there being gender differences. We have also shown that, for French medical students, students with personality traits associated with strong competitiveness are selected for admission to medical school. In this study, we further show that the medical students have different personality profiles compared with other student groups (psychology and business studies). The main purpose of the present investigation was to assess attitudes to science and gross anatomy, and to relate these to the students' personalities. Questionnaires (including Thurstone and Chave analyses) were employed to measure attitudes, and personality was assessed using the Big Five Inventory (BFI). Data for attitudes were obtained for students at medical schools in Cardiff (UK), Paris, Descartes/Sorbonne (France), St George's University (Grenada) and Ankara (Turkey). Data obtained from personality tests were available for analysis from the Parisian cohort of students. Although the medical students were found to have strongly supportive views concerning the importance of science in medicine, their knowledge of the scientific method/philosophy of science was poor. Following analyses of the BFI in the French students, ‘openness’ and ‘conscientiousness’ were linked statistically with a positive attitude towards science. For anatomy, again strongly supportive views concerning the subject's importance in medicine were discerned. Analyses of the BFI in the French students did not show links statistically between personality profiles and attitudes towards gross anatomy, except male students with ‘negative affectivity’ showed less appreciation of the importance of anatomy. This contrasts with our earlier studies that showed that there is a relationship between the BF dimensions of personality traits and anxiety towards the dissection room experience (at the start of the course, ‘negative emotionality’ was related to an increased level of anxiety). We conclude that medical students agree on the importance to their studies of both science in general and gross anatomy in particular, and that some personality traits relate to their attitudes that could affect clinical competence. PMID:23594196
Medical students' attitudes towards science and gross anatomy, and the relationship to personality.
Plaisant, Odile; Stephens, Shiby; Apaydin, Nihal; Courtois, Robert; Lignier, Baptiste; Loukas, Marios; Moxham, Bernard
2014-03-01
Assessment of the personalities of medical students can enable medical educators to formulate strategies for the best development of academic and clinical competencies. Previous research has shown that medical students do not share a common personality profile, there being gender differences. We have also shown that, for French medical students, students with personality traits associated with strong competitiveness are selected for admission to medical school. In this study, we further show that the medical students have different personality profiles compared with other student groups (psychology and business studies). The main purpose of the present investigation was to assess attitudes to science and gross anatomy, and to relate these to the students' personalities. Questionnaires (including Thurstone and Chave analyses) were employed to measure attitudes, and personality was assessed using the Big Five Inventory (BFI). Data for attitudes were obtained for students at medical schools in Cardiff (UK), Paris, Descartes/Sorbonne (France), St George's University (Grenada) and Ankara (Turkey). Data obtained from personality tests were available for analysis from the Parisian cohort of students. Although the medical students were found to have strongly supportive views concerning the importance of science in medicine, their knowledge of the scientific method/philosophy of science was poor. Following analyses of the BFI in the French students, 'openness' and 'conscientiousness' were linked statistically with a positive attitude towards science. For anatomy, again strongly supportive views concerning the subject's importance in medicine were discerned. Analyses of the BFI in the French students did not show links statistically between personality profiles and attitudes towards gross anatomy, except male students with 'negative affectivity' showed less appreciation of the importance of anatomy. This contrasts with our earlier studies that showed that there is a relationship between the BF dimensions of personality traits and anxiety towards the dissection room experience (at the start of the course, 'negative emotionality' was related to an increased level of anxiety). We conclude that medical students agree on the importance to their studies of both science in general and gross anatomy in particular, and that some personality traits relate to their attitudes that could affect clinical competence. © 2013 Anatomical Society.
19 CFR 103.31 - Information on vessel manifests and summary statistical reports.
Code of Federal Regulations, 2011 CFR
2011-04-01
... statistical reports. 103.31 Section 103.31 Customs Duties U.S. CUSTOMS AND BORDER PROTECTION, DEPARTMENT OF... Restricted Access § 103.31 Information on vessel manifests and summary statistical reports. (a) Disclosure to... statistical reports of imports and exports and to copy therefrom for publication information and data subject...
ERIC Educational Resources Information Center
Hilton, Sterling C.; Schau, Candace; Olsen, Joseph A.
2004-01-01
In addition to student learning, positive student attitudes have become an important course outcome for many introductory statistics instructors. To adequately assess changes in mean attitudes across introductory statistics courses, the attitude instruments used should be invariant by administration time. Attitudes toward statistics from 4,910…
A Unifying Framework for Teaching Nonparametric Statistical Tests
ERIC Educational Resources Information Center
Bargagliotti, Anna E.; Orrison, Michael E.
2014-01-01
Increased importance is being placed on statistics at both the K-12 and undergraduate level. Research divulging effective methods to teach specific statistical concepts is still widely sought after. In this paper, we focus on best practices for teaching topics in nonparametric statistics at the undergraduate level. To motivate the work, we…
19 CFR 103.31 - Information on vessel manifests and summary statistical reports.
Code of Federal Regulations, 2010 CFR
2010-04-01
... statistical reports. 103.31 Section 103.31 Customs Duties U.S. CUSTOMS AND BORDER PROTECTION, DEPARTMENT OF... Restricted Access § 103.31 Information on vessel manifests and summary statistical reports. (a) Disclosure to... statistical reports of imports and exports and to copy therefrom for publication information and data subject...
Attitudes toward statistics in medical postgraduates: measuring, evaluating and monitoring.
Zhang, Yuhai; Shang, Lei; Wang, Rui; Zhao, Qinbo; Li, Chanjuan; Xu, Yongyong; Su, Haixia
2012-11-23
In medical training, statistics is considered a very difficult course to learn and teach. Current studies have found that students' attitudes toward statistics can influence their learning process. Measuring, evaluating and monitoring the changes of students' attitudes toward statistics are important. Few studies have focused on the attitudes of postgraduates, especially medical postgraduates. Our purpose was to understand current attitudes regarding statistics held by medical postgraduates and explore their effects on students' achievement. We also wanted to explore the influencing factors and the sources of these attitudes and monitor their changes after a systematic statistics course. A total of 539 medical postgraduates enrolled in a systematic statistics course completed the pre-form of the Survey of Attitudes Toward Statistics -28 scale, and 83 postgraduates were selected randomly from among them to complete the post-form scale after the course. Most medical postgraduates held positive attitudes toward statistics, but they thought statistics was a very difficult subject. The attitudes mainly came from experiences in a former statistical or mathematical class. Age, level of statistical education, research experience, specialty and mathematics basis may influence postgraduate attitudes toward statistics. There were significant positive correlations between course achievement and attitudes toward statistics. In general, student attitudes showed negative changes after completing a statistics course. The importance of student attitudes toward statistics must be recognized in medical postgraduate training. To make sure all students have a positive learning environment, statistics teachers should measure their students' attitudes and monitor their change of status during a course. Some necessary assistance should be offered for those students who develop negative attitudes.
NASA Technical Reports Server (NTRS)
Wilson, Robert M.
1998-01-01
On the basis of Trenberth's quantitative definition for marking the occurrence of an El Nino (or La Nina), one can precisely identify by month and year the starts and ends of some 15 El Nino and 10 La Nina events during the interval of 1950-1997, an interval corresponding to the most reliable for cataloging intense hurricane activity in the Atlantic basin (i.e., those of category 3-5 on the Saffir-Simpson hurricane scale). The main purpose of this investigation is primarily two-fold: First, the statistical aspects of these identified extremes and the intervening periods between them (called "interludes") are examined and, second, the statistics of the seasonal frequency of intense hurricanes in comparison to the extremes and interludes are determined. This study clearly demonstrates that of the last 48 hurricane seasons, 20 (42 percent) can be described as being "El Nino-related" (i.e., an El Nino was in progress during all, or part, of the yearly hurricane season--June-November), 13 (27 percent) as "La Nina-related" (i.e., a La Nina was in progress during all, or part, of the yearly hurricane season), and 15 (31 percent) as "interlude-related" (i.e., neither an El Nino nor a La Nina was in progress during any portion of the yearly hurricane season). Combining the latter two subgroups into a single grouping called "non-El Nino-related" seasons, one finds that they have had a mean frequency of intense hurricanes measuring 2.8 events per season, while the El Nino-related seasons have had a mean frequency of intense hurricanes measuring 1.3 events per season, where the observed difference in the means is inferred to be statistically important at the 99.8-percent level of confidence. Therefore, as previously shown more than a decade ago using a different data set, there undeniably exists an El Nino-Atlantic hurricane activity relationship, one which also extends to the class of intense hurricanes. During the interval of 1950-1997, fewer intense hurricanes occurred during El Nino-related seasons (always less than or equal to 3 and usually less than or equal to 2, this latter value having been true for 18 of the 20 El Nino-related seasons), while more usually occurred during non-El Nino-related seasons (typically greater than or equal to 2, having been true for 22 of the 28 non-El Nino-related seasons). Implications for the 1998 and 1999 hurricane seasons are discussed.
Assessing dimensions of nicotine dependence
Piper, Megan E.; McCarthy, Danielle E.; Bolt, Daniel M.; Smith, Stevens S.; Lerman, Caryn; Benowitz, Neal; Fiore, Michael C.; Baker, Timothy B.
2008-01-01
Considerable research, ranging from survey to clinical to genetic, has utilized traditional measures of tobacco dependence, such as the Fagerstrom Test of Nicotine Dependence (FTND) and the Diagnostic and Statistical Manual (DSM-IV) criteria, that focus on endpoint definitions of tobacco dependence such as heavy smoking, time to first cigarette in the morning, and smoking despite consequences. In an effort to better understand possible theories and mechanisms underlying tobacco dependence, which could be used to improve treatment and research, two multidimensional measures of tobacco dependence have been developed: the Nicotine Dependence Syndrome Scale (NDSS) and the Wisconsin Inventory of Smoking Dependence Motives (WISDM). This research used data from three randomized smoking cessation trials to examine the internal consistency and validity (convergent, concurrent and predictive) of these scales, relative to each other and the traditional measures. Results reveal that NDSS and WISDM subscales are related to important dependence criteria, but in a heterogeneous fashion. The data suggest that there are important underlying mechanisms or motives that are significantly related to different important outcomes, such as withdrawal and cessation. The FTND was most strongly related to abstinence at 1 week and 6 months post-quit, whereas the WISDM Tolerance subscale was most strongly related to abstinence at the end of treatment. The NDSS Priority subscale was consistently predictive of outcome at all three follow-up time points. There is also evidence that WISDM subscales are related to a biomarker of the rate of nicotine metabolism. PMID:18584464
Statistical physics of human beings in games: Controlled experiments
NASA Astrophysics Data System (ADS)
Liang, Yuan; Huang, Ji-Ping
2014-07-01
It is important to know whether the laws or phenomena in statistical physics for natural systems with non-adaptive agents still hold for social human systems with adaptive agents, because this implies whether it is possible to study or understand social human systems by using statistical physics originating from natural systems. For this purpose, we review the role of human adaptability in four kinds of specific human behaviors, namely, normal behavior, herd behavior, contrarian behavior, and hedge behavior. The approach is based on controlled experiments in the framework of market-directed resource-allocation games. The role of the controlled experiments could be at least two-fold: adopting the real human decision-making process so that the system under consideration could reflect the performance of genuine human beings; making it possible to obtain macroscopic physical properties of a human system by tuning a particular factor of the system, thus directly revealing cause and effect. As a result, both computer simulations and theoretical analyses help to show a few counterparts of some laws or phenomena in statistical physics for social human systems: two-phase phenomena or phase transitions, entropy-related phenomena, and a non-equilibrium steady state. This review highlights the role of human adaptability in these counterparts, and makes it possible to study or understand some particular social human systems by means of statistical physics coming from natural systems.
[Design and application of user managing system of cardiac remote monitoring network].
Chen, Shouqiang; Zhang, Jianmin; Yuan, Feng; Gao, Haiqing
2007-12-01
According to inpatient records, data managing demand of cardiac remote monitoring network and computer, this software was designed with relative database ACCESS. Its interface, operational button and menu were designed in VBA language assistantly. Its design included collective design, amity, practicability and compatibility. Its function consisted of registering, inquiring, statisticing and printing, et al. It could be used to manage users effectively and could be helpful to exerting important action of cardiac remote monitoring network in preventing cardiac-vascular emergency ulteriorly.
NASA Technical Reports Server (NTRS)
Koch, S. E.; Skillman, W. C.; Kocin, P. J.; Wetzel, P. J.; Brill, K. F.
1985-01-01
The synoptic scale performance characteristics of MASS 2.0 are determined by comparing filtered 12-24 hr model forecasts to same-case forecasts made by the National Meteorological Center's synoptic-scale Limited-area Fine Mesh model. Characteristics of the two systems are contrasted, and the analysis methodology used to determine statistical skill scores and systematic errors is described. The overall relative performance of the two models in the sample is documented, and important systematic errors uncovered are presented.
Campbell, M J
1983-01-01
I describe methods of analysing possible aetiological factors in a follow-up survey, all of which are possible to carry out using the statistical package GENSTAT. A high haemoglobin level carried a significantly increased risk of ischaemic heart disease, and a low one an increased risk of cancer. Smoking was also an important factor. The increased risk was reasonably constant over time. Sugar intake and Quetelet's index did not significantly affect the relative risk.
Research on the Hotel Image Based on the Detail Service
NASA Astrophysics Data System (ADS)
Li, Ban; Shenghua, Zheng; He, Yi
Detail service management, initially developed as marketing programs to enhance customer loyalty, has now become an important part of customer relation strategy. This paper analyzes the critical factors of detail service and its influence on the hotel image. We establish the theoretical model of influencing factors on hotel image and propose corresponding hypotheses. We use applying statistical method to test and verify the above-mentioned hypotheses. This paper provides a foundation for further study of detail service design and planning issues.
Zaki, Rafdzah; Bulgiba, Awang; Ismail, Roshidi; Ismail, Noor Azina
2012-01-01
Accurate values are a must in medicine. An important parameter in determining the quality of a medical instrument is agreement with a gold standard. Various statistical methods have been used to test for agreement. Some of these methods have been shown to be inappropriate. This can result in misleading conclusions about the validity of an instrument. The Bland-Altman method is the most popular method judging by the many citations of the article proposing this method. However, the number of citations does not necessarily mean that this method has been applied in agreement research. No previous study has been conducted to look into this. This is the first systematic review to identify statistical methods used to test for agreement of medical instruments. The proportion of various statistical methods found in this review will also reflect the proportion of medical instruments that have been validated using those particular methods in current clinical practice. Five electronic databases were searched between 2007 and 2009 to look for agreement studies. A total of 3,260 titles were initially identified. Only 412 titles were potentially related, and finally 210 fitted the inclusion criteria. The Bland-Altman method is the most popular method with 178 (85%) studies having used this method, followed by the correlation coefficient (27%) and means comparison (18%). Some of the inappropriate methods highlighted by Altman and Bland since the 1980s are still in use. This study finds that the Bland-Altman method is the most popular method used in agreement research. There are still inappropriate applications of statistical methods in some studies. It is important for a clinician or medical researcher to be aware of this issue because misleading conclusions from inappropriate analyses will jeopardize the quality of the evidence, which in turn will influence quality of care given to patients in the future.
Zaki, Rafdzah; Bulgiba, Awang; Ismail, Roshidi; Ismail, Noor Azina
2012-01-01
Background Accurate values are a must in medicine. An important parameter in determining the quality of a medical instrument is agreement with a gold standard. Various statistical methods have been used to test for agreement. Some of these methods have been shown to be inappropriate. This can result in misleading conclusions about the validity of an instrument. The Bland-Altman method is the most popular method judging by the many citations of the article proposing this method. However, the number of citations does not necessarily mean that this method has been applied in agreement research. No previous study has been conducted to look into this. This is the first systematic review to identify statistical methods used to test for agreement of medical instruments. The proportion of various statistical methods found in this review will also reflect the proportion of medical instruments that have been validated using those particular methods in current clinical practice. Methodology/Findings Five electronic databases were searched between 2007 and 2009 to look for agreement studies. A total of 3,260 titles were initially identified. Only 412 titles were potentially related, and finally 210 fitted the inclusion criteria. The Bland-Altman method is the most popular method with 178 (85%) studies having used this method, followed by the correlation coefficient (27%) and means comparison (18%). Some of the inappropriate methods highlighted by Altman and Bland since the 1980s are still in use. Conclusions This study finds that the Bland-Altman method is the most popular method used in agreement research. There are still inappropriate applications of statistical methods in some studies. It is important for a clinician or medical researcher to be aware of this issue because misleading conclusions from inappropriate analyses will jeopardize the quality of the evidence, which in turn will influence quality of care given to patients in the future. PMID:22662248
Model averaging and muddled multimodel inferences.
Cade, Brian S
2015-09-01
Three flawed practices associated with model averaging coefficients for predictor variables in regression models commonly occur when making multimodel inferences in analyses of ecological data. Model-averaged regression coefficients based on Akaike information criterion (AIC) weights have been recommended for addressing model uncertainty but they are not valid, interpretable estimates of partial effects for individual predictors when there is multicollinearity among the predictor variables. Multicollinearity implies that the scaling of units in the denominators of the regression coefficients may change across models such that neither the parameters nor their estimates have common scales, therefore averaging them makes no sense. The associated sums of AIC model weights recommended to assess relative importance of individual predictors are really a measure of relative importance of models, with little information about contributions by individual predictors compared to other measures of relative importance based on effects size or variance reduction. Sometimes the model-averaged regression coefficients for predictor variables are incorrectly used to make model-averaged predictions of the response variable when the models are not linear in the parameters. I demonstrate the issues with the first two practices using the college grade point average example extensively analyzed by Burnham and Anderson. I show how partial standard deviations of the predictor variables can be used to detect changing scales of their estimates with multicollinearity. Standardizing estimates based on partial standard deviations for their variables can be used to make the scaling of the estimates commensurate across models, a necessary but not sufficient condition for model averaging of the estimates to be sensible. A unimodal distribution of estimates and valid interpretation of individual parameters are additional requisite conditions. The standardized estimates or equivalently the t statistics on unstandardized estimates also can be used to provide more informative measures of relative importance than sums of AIC weights. Finally, I illustrate how seriously compromised statistical interpretations and predictions can be for all three of these flawed practices by critiquing their use in a recent species distribution modeling technique developed for predicting Greater Sage-Grouse (Centrocercus urophasianus) distribution in Colorado, USA. These model averaging issues are common in other ecological literature and ought to be discontinued if we are to make effective scientific contributions to ecological knowledge and conservation of natural resources.
Factors influencing nurse managers' intent to stay or leave: a quantitative analysis.
Hewko, Sarah J; Brown, Pamela; Fraser, Kimberly D; Wong, Carol A; Cummings, Greta G
2015-11-01
To identify and report on the relative importance of factors influencing nurse managers' intentions to stay in or leave their current position. Effective nurse managers play an important role in staff nurse retention and in the quality of patient care. The advancing age of nurse managers, multiple job opportunities within nursing and the generally negative perceptions of the manager role can contribute to difficulties in retaining nurse managers. Ninety-five Canadian nurse managers participated in a web survey. Respondents rated the importance of factors related to their intent to leave or stay in their current position for another 2 years. Descriptive, t-test and mancova statistics were used to assess differences between managers intending to stay or leave. For managers intending to leave (n = 28), the most important factors were work overload, inability to ensure quality patient care, insufficient resources, and lack of empowerment and recognition. Managers intending to leave reported significantly lower job satisfaction, perceptions of their supervisor's resonant leadership and higher burnout levels. Organisations wishing to retain existing nurse managers and to attract front-line staff into leadership positions must create and foster an environment that supports nurse managers. © 2014 John Wiley & Sons Ltd.
Evaluation and application of summary statistic imputation to discover new height-associated loci.
Rüeger, Sina; McDaid, Aaron; Kutalik, Zoltán
2018-05-01
As most of the heritability of complex traits is attributed to common and low frequency genetic variants, imputing them by combining genotyping chips and large sequenced reference panels is the most cost-effective approach to discover the genetic basis of these traits. Association summary statistics from genome-wide meta-analyses are available for hundreds of traits. Updating these to ever-increasing reference panels is very cumbersome as it requires reimputation of the genetic data, rerunning the association scan, and meta-analysing the results. A much more efficient method is to directly impute the summary statistics, termed as summary statistics imputation, which we improved to accommodate variable sample size across SNVs. Its performance relative to genotype imputation and practical utility has not yet been fully investigated. To this end, we compared the two approaches on real (genotyped and imputed) data from 120K samples from the UK Biobank and show that, genotype imputation boasts a 3- to 5-fold lower root-mean-square error, and better distinguishes true associations from null ones: We observed the largest differences in power for variants with low minor allele frequency and low imputation quality. For fixed false positive rates of 0.001, 0.01, 0.05, using summary statistics imputation yielded a decrease in statistical power by 9, 43 and 35%, respectively. To test its capacity to discover novel associations, we applied summary statistics imputation to the GIANT height meta-analysis summary statistics covering HapMap variants, and identified 34 novel loci, 19 of which replicated using data in the UK Biobank. Additionally, we successfully replicated 55 out of the 111 variants published in an exome chip study. Our study demonstrates that summary statistics imputation is a very efficient and cost-effective way to identify and fine-map trait-associated loci. Moreover, the ability to impute summary statistics is important for follow-up analyses, such as Mendelian randomisation or LD-score regression.
Evaluation and application of summary statistic imputation to discover new height-associated loci
2018-01-01
As most of the heritability of complex traits is attributed to common and low frequency genetic variants, imputing them by combining genotyping chips and large sequenced reference panels is the most cost-effective approach to discover the genetic basis of these traits. Association summary statistics from genome-wide meta-analyses are available for hundreds of traits. Updating these to ever-increasing reference panels is very cumbersome as it requires reimputation of the genetic data, rerunning the association scan, and meta-analysing the results. A much more efficient method is to directly impute the summary statistics, termed as summary statistics imputation, which we improved to accommodate variable sample size across SNVs. Its performance relative to genotype imputation and practical utility has not yet been fully investigated. To this end, we compared the two approaches on real (genotyped and imputed) data from 120K samples from the UK Biobank and show that, genotype imputation boasts a 3- to 5-fold lower root-mean-square error, and better distinguishes true associations from null ones: We observed the largest differences in power for variants with low minor allele frequency and low imputation quality. For fixed false positive rates of 0.001, 0.01, 0.05, using summary statistics imputation yielded a decrease in statistical power by 9, 43 and 35%, respectively. To test its capacity to discover novel associations, we applied summary statistics imputation to the GIANT height meta-analysis summary statistics covering HapMap variants, and identified 34 novel loci, 19 of which replicated using data in the UK Biobank. Additionally, we successfully replicated 55 out of the 111 variants published in an exome chip study. Our study demonstrates that summary statistics imputation is a very efficient and cost-effective way to identify and fine-map trait-associated loci. Moreover, the ability to impute summary statistics is important for follow-up analyses, such as Mendelian randomisation or LD-score regression. PMID:29782485
Cued Memory Reactivation During SWS Abolishes the Beneficial Effect of Sleep on Abstraction.
Hennies, Nora; Lambon Ralph, Matthew A; Durrant, Simon J; Cousins, James N; Lewis, Penelope A
2017-08-01
Extracting regularities from stimuli in our environment and generalizing these to new situations are fundamental processes in human cognition. Sleep has been shown to enhance these processes, possibly by facilitating reactivation-triggered memory reorganization. Here, we assessed whether cued reactivation during slow wave sleep (SWS) promotes the beneficial effect of sleep on abstraction of statistical regularities. We used an auditory statistical learning task, in which the benefit of sleep has been firmly established. Participants were exposed to a probabilistically determined sequence of tones and subsequently tested for recognition of novel short sequences adhering to this same statistical pattern in both immediate and delayed recall sessions. In different groups, the exposure stream was replayed during SWS in the night between the recall sessions (SWS-replay group), in wake just before sleep (presleep replay group), or not at all (control group). Surprisingly, participants who received replay in sleep performed worse in the delayed recall session than the control and the presleep replay group. They also failed to show the association between SWS and task performance that has been observed in previous studies and was present in the controls. Importantly, sleep structure and sleep quality did not differ between groups, suggesting that replay during SWS did not impair sleep but rather disrupted or interfered with sleep-dependent mechanisms that underlie the extraction of the statistical pattern. These findings raise important questions about the scope of cued memory reactivation and the mechanisms that underlie sleep-related generalization. © Sleep Research Society 2017. Published by Oxford University Press on behalf of the Sleep Research Society. All rights reserved. For permissions, please e-mail journals.permissions@oup.com.
Statistical Method for Identification of Potential Groundwater Recharge Zone
NASA Astrophysics Data System (ADS)
Banerjee, Pallavi; Singh, V. S.
2010-05-01
The effective development of groundwater resource is essential for a country like India. Artificial recharge is the planned, human activity of augmenting the amount of groundwater available through works designed to increase the natural replenishment or percolation of surface waters into the groundwater aquifers, resulting in a corresponding increase in the amount of groundwater available for abstraction. India receives good amount of average annual rainfall about 114 cm but most of it's part waste through runoff. The imbalance between rainfall and recharge has caused serious shortage of water for drinking, agriculture and industrial purposes. The over exploitation of groundwater due to increasing population is an additional cause of water crisis that resulting in reduction in per capita availability of water in the country. Thus the planning for effective development of groundwater is essential through artificial recharge. Objective of the paper is to identification of artificial recharge zones by arresting runoff through suitable sites to restore groundwater conditions using statistical technique. The water table variation follows a pattern similar to rainfall variation with time delay. The rainfall and its relationship with recharge is a very important process in a shallow aquifer system. Understanding of this process is of critical importance to management of groundwater resource in any terrain. Groundwater system in a top weathered regolith in a balastic terrain forms shallow aquifer is often classified into shallow water table category. In the present study an effort has been made to understand the suitable recharge zone with relation to rainfall and water level by using statistical analysis. Daily time series data of rainfall and borehole water level data are cross correlated to investigate variations in groundwater level response time during the months of monsoon. This measurement facilitate to demarcate favorable areas for Artificial Recharge. KEYWORDS: Water level; Rainfall; Recharge; Statistical analysis; Cross correlation.
Boxwala, Aziz A; Kim, Jihoon; Grillo, Janice M; Ohno-Machado, Lucila
2011-01-01
To determine whether statistical and machine-learning methods, when applied to electronic health record (EHR) access data, could help identify suspicious (ie, potentially inappropriate) access to EHRs. From EHR access logs and other organizational data collected over a 2-month period, the authors extracted 26 features likely to be useful in detecting suspicious accesses. Selected events were marked as either suspicious or appropriate by privacy officers, and served as the gold standard set for model evaluation. The authors trained logistic regression (LR) and support vector machine (SVM) models on 10-fold cross-validation sets of 1291 labeled events. The authors evaluated the sensitivity of final models on an external set of 58 events that were identified as truly inappropriate and investigated independently from this study using standard operating procedures. The area under the receiver operating characteristic curve of the models on the whole data set of 1291 events was 0.91 for LR, and 0.95 for SVM. The sensitivity of the baseline model on this set was 0.8. When the final models were evaluated on the set of 58 investigated events, all of which were determined as truly inappropriate, the sensitivity was 0 for the baseline method, 0.76 for LR, and 0.79 for SVM. The LR and SVM models may not generalize because of interinstitutional differences in organizational structures, applications, and workflows. Nevertheless, our approach for constructing the models using statistical and machine-learning techniques can be generalized. An important limitation is the relatively small sample used for the training set due to the effort required for its construction. The results suggest that statistical and machine-learning methods can play an important role in helping privacy officers detect suspicious accesses to EHRs.
Kim, Jihoon; Grillo, Janice M; Ohno-Machado, Lucila
2011-01-01
Objective To determine whether statistical and machine-learning methods, when applied to electronic health record (EHR) access data, could help identify suspicious (ie, potentially inappropriate) access to EHRs. Methods From EHR access logs and other organizational data collected over a 2-month period, the authors extracted 26 features likely to be useful in detecting suspicious accesses. Selected events were marked as either suspicious or appropriate by privacy officers, and served as the gold standard set for model evaluation. The authors trained logistic regression (LR) and support vector machine (SVM) models on 10-fold cross-validation sets of 1291 labeled events. The authors evaluated the sensitivity of final models on an external set of 58 events that were identified as truly inappropriate and investigated independently from this study using standard operating procedures. Results The area under the receiver operating characteristic curve of the models on the whole data set of 1291 events was 0.91 for LR, and 0.95 for SVM. The sensitivity of the baseline model on this set was 0.8. When the final models were evaluated on the set of 58 investigated events, all of which were determined as truly inappropriate, the sensitivity was 0 for the baseline method, 0.76 for LR, and 0.79 for SVM. Limitations The LR and SVM models may not generalize because of interinstitutional differences in organizational structures, applications, and workflows. Nevertheless, our approach for constructing the models using statistical and machine-learning techniques can be generalized. An important limitation is the relatively small sample used for the training set due to the effort required for its construction. Conclusion The results suggest that statistical and machine-learning methods can play an important role in helping privacy officers detect suspicious accesses to EHRs. PMID:21672912
Selimović-Dragaš, Mediha; Hasić-Branković, Lajla; Korać, Fehim; Đapo, Nermin; Huseinbegović, Amina; Kobašlija, Sedin; Lekić, Meliha; Hatibović-Kofman, Šahza
2013-01-01
Fluoride release is important characteristic of glass-ionomer cements. Quantity of fluoride ions released from the glass-ionomer cements has major importance in definition of their biological activity. The objectives of this study were to define the quantity of fluoride ions released from the experimental glass-ionomer cements and to define the effect of fluoride ions released from the experimental glass-ionomer cements on their cytotoxicity. Concentrations of the fluoride ions released in the evaluated glass-ionomer cements were measured indirectly, by the fluoride-selective WTW, F500 electrode potential, combined with reference R503/D electrode. Statistical analyses of F-ion concentrations released by all glass-ionomers evaluated at two time points, after 8 and after 24 hours, show statistically higher fluoride releases from RMGICs: Vitrebond, Fuji II LC and Fuji Plus, when compared to conventional glass-ionomer cements: Fuji Triage, Fuji IX GP Fast and Ketac Silver, both after 8 and after 24 hours. Correlation coefficient between concentrations of fluoride ion released by evaluated glass-ionomer cements and cytotoxic response of UMR-106 osteoblast cell-line are relatively high, but do not reach levels of biological significance. Correlation between concentrations of fluoride ion released and cytotoxic response of NIH3T3 mouse fibroblast cell line after 8 hours is high, positive and statistically significant for conventional GICs, Fuji Triage and Fuji IX GP Fast, and RMGIC, Fuji II LC. Statistically significant Correlation coefficient between concentrations of fluoride ion released and cytotoxic response of NIH3T3 cell line after 24 hours is defined for RMGIC Fuji II LC only. PMID:23988173
Statistical issues in the design, conduct and analysis of two large safety studies.
Gaffney, Michael
2016-10-01
The emergence, post approval, of serious medical events, which may be associated with the use of a particular drug or class of drugs, is an important public health and regulatory issue. The best method to address this issue is through a large, rigorously designed safety study. Therefore, it is important to elucidate the statistical issues involved in these large safety studies. Two such studies are PRECISION and EAGLES. PRECISION is the primary focus of this article. PRECISION is a non-inferiority design with a clinically relevant non-inferiority margin. Statistical issues in the design, conduct and analysis of PRECISION are discussed. Quantitative and clinical aspects of the selection of the composite primary endpoint, the determination and role of the non-inferiority margin in a large safety study and the intent-to-treat and modified intent-to-treat analyses in a non-inferiority safety study are shown. Protocol changes that were necessary during the conduct of PRECISION are discussed from a statistical perspective. Issues regarding the complex analysis and interpretation of the results of PRECISION are outlined. EAGLES is presented as a large, rigorously designed safety study when a non-inferiority margin was not able to be determined by a strong clinical/scientific method. In general, when a non-inferiority margin is not able to be determined, the width of the 95% confidence interval is a way to size the study and to assess the cost-benefit of relative trial size. A non-inferiority margin, when able to be determined by a strong scientific method, should be included in a large safety study. Although these studies could not be called "pragmatic," they are examples of best real-world designs to address safety and regulatory concerns. © The Author(s) 2016.
Shi, Jie; Collignon, Olivier; Xu, Liang; Wang, Gang; Kang, Yue; Leporé, Franco; Lao, Yi; Joshi, Anand A; Leporé, Natasha; Wang, Yalin
2015-07-01
Blindness represents a unique model to study how visual experience may shape the development of brain organization. Exploring how the structure of the corpus callosum (CC) reorganizes ensuing visual deprivation is of particular interest due to its important functional implication in vision (e.g., via the splenium of the CC). Moreover, comparing early versus late visually deprived individuals has the potential to unravel the existence of a sensitive period for reshaping the CC structure. Here, we develop a novel framework to capture a complete set of shape differences in the CC between congenitally blind (CB), late blind (LB) and sighted control (SC) groups. The CCs were manually segmented from T1-weighted brain MRI and modeled by 3D tetrahedral meshes. We statistically compared the combination of local area and thickness at each point between subject groups. Differences in area are found using surface tensor-based morphometry; thickness is estimated by tracing the streamlines in the volumetric harmonic field. Group differences were assessed on this combined measure using Hotelling's T(2) test. Interestingly, we observed that the total callosal volume did not differ between the groups. However, our fine-grained analysis reveals significant differences mostly localized around the splenium areas between both blind groups and the sighted group (general effects of blindness) and, importantly, specific dissimilarities between the LB and CB groups, illustrating the existence of a sensitive period for reorganization. The new multivariate statistics also gave better effect sizes for detecting morphometric differences, relative to other statistics. They may boost statistical power for CC morphometric analyses.
September Arctic Sea Ice minimum prediction - a new skillful statistical approach
NASA Astrophysics Data System (ADS)
Ionita-Scholz, Monica; Grosfeld, Klaus; Scholz, Patrick; Treffeisen, Renate; Lohmann, Gerrit
2017-04-01
Sea ice in both Polar Regions is an important indicator for the expression of global climate change and its polar amplification. Consequently, a broad interest exists on sea ice, its coverage, variability and long term change. Knowledge on sea ice requires high quality data on ice extent, thickness and its dynamics. However, its predictability is complex and it depends on various climate and oceanic parameters and conditions. In order to provide insights into the potential development of a monthly/seasonal signal of sea ice evolution, we developed a robust statistical model based on ocean heat content, sea surface temperature and different atmospheric variables to calculate an estimate of the September Sea ice extent (SSIE) on monthly time scale. Although previous statistical attempts at monthly/seasonal forecasts of SSIE show a relatively reduced skill, we show here that more than 92% (r = 0.96) of the September sea ice extent can be predicted at the end of May by using previous months' climate and oceanic conditions. The skill of the model increases with a decrease in the time lag used for the forecast. At the end of August, our predictions are even able to explain 99% of the SSIE. Our statistical model captures both the general trend as well as the interannual variability of the SSIE. Moreover, it is able to properly forecast the years with extreme high/low SSIE (e.g. 1996/ 2007, 2012, 2013). Besides its forecast skill for SSIE, the model could provide a valuable tool for identifying relevant regions and climate parameters that are important for the sea ice development in the Arctic and for detecting sensitive and critical regions in global coupled climate models with focus on sea ice formation.
Shi, Jie; Collignon, Olivier; Xu, Liang; Wang, Gang; Kang, Yue; Leporé, Franco; Lao, Yi; Joshi, Anand A.
2015-01-01
Blindness represents a unique model to study how visual experience may shape the development of brain organization. Exploring how the structure of the corpus callosum (CC) reorganizes ensuing visual deprivation is of particular interest due to its important functional implication in vision (e.g. via the splenium of the CC). Moreover, comparing early versus late visually deprived individuals has the potential to unravel the existence of a sensitive period for reshaping the CC structure. Here, we develop a novel framework to capture a complete set of shape differences in the CC between congenitally blind (CB), late blind (LB) and sighted control (SC) groups. The CCs were manually segmented from T1-weighted brain MRI and modeled by 3D tetrahedral meshes. We statistically compared the combination of local area and thickness at each point between subject groups. Differences in area are found using surface tensor-based morphometry; thickness is estimated by tracing the streamlines in the volumetric harmonic field. Group differences were assessed on this combined measure using Hotelling’s T2 test. Interestingly, we observed that the total callosal volume did not differ between the groups. However, our fine-grained analysis reveals significant differences mostly localized around the splenium areas between both blind groups and the sighted group (general effects of blindness) and, importantly, specific dissimilarities between the LB and CB groups, illustrating the existence of a sensitive period for reorganization. The new multivariate statistics also gave better effect sizes for detecting morphometric differences, relative to other statistics. They may boost statistical power for CC morphometric analyses. PMID:25649876
Paechter, Manuela; Macher, Daniel; Martskvishvili, Khatuna; Wimmer, Sigrid; Papousek, Ilona
2017-01-01
In many social science majors, e.g., psychology, students report high levels of statistics anxiety. However, these majors are often chosen by students who are less prone to mathematics and who might have experienced difficulties and unpleasant feelings in their mathematics courses at school. The present study investigates whether statistics anxiety is a genuine form of anxiety that impairs students' achievements or whether learners mainly transfer previous experiences in mathematics and their anxiety in mathematics to statistics. The relationship between mathematics anxiety and statistics anxiety, their relationship to learning behaviors and to performance in a statistics examination were investigated in a sample of 225 undergraduate psychology students (164 women, 61 men). Data were recorded at three points in time: At the beginning of term students' mathematics anxiety, general proneness to anxiety, school grades, and demographic data were assessed; 2 weeks before the end of term, they completed questionnaires on statistics anxiety and their learning behaviors. At the end of term, examination scores were recorded. Mathematics anxiety and statistics anxiety correlated highly but the comparison of different structural equation models showed that they had genuine and even antagonistic contributions to learning behaviors and performance in the examination. Surprisingly, mathematics anxiety was positively related to performance. It might be that students realized over the course of their first term that knowledge and skills in higher secondary education mathematics are not sufficient to be successful in statistics. Part of mathematics anxiety may then have strengthened positive extrinsic effort motivation by the intention to avoid failure and may have led to higher effort for the exam preparation. However, via statistics anxiety mathematics anxiety also had a negative contribution to performance. Statistics anxiety led to higher procrastination in the structural equation model and, therefore, contributed indirectly and negatively to performance. Furthermore, it had a direct negative impact on performance (probably via increased tension and worry in the exam). The results of the study speak for shared but also unique components of statistics anxiety and mathematics anxiety. They are also important for instruction and give recommendations to learners as well as to instructors. PMID:28790938
Attitude of teaching faculty towards statistics at a medical university in Karachi, Pakistan.
Khan, Nazeer; Mumtaz, Yasmin
2009-01-01
Statistics is mainly used in biological research to verify the clinicians and researchers findings and feelings, and gives scientific validity for their inferences. In Pakistan, the educational curriculum is developed in such a way that the students who are interested in entering in the field of biological sciences do not study mathematics after grade 10. Therefore, due to their fragile background of mathematical skills, the Pakistani medical professionals feel that they do not have adequate base to understand the basic concepts of statistical techniques when they try to use it in their research or read a scientific article. The aim of the study was to assess the attitude of medical faculty towards statistics. A questionnaire containing 42 close-ended and 4 open-ended questions, related to the attitude and knowledge of statistics, was distributed among the teaching faculty of Dow University of Health Sciences (DUHS). One hundred and sixty-seven filled questionnaires were returned from 374 faculty members (response rate 44.7%). Forty-three percent of the respondents claimed that they had 'introductive' level of statistics courses, 63% of the respondents strongly agreed that a good researcher must have some training in statistics, 82% of the faculty was in favour (strongly agreed or agreed) that statistics was really useful for research. Only 17% correctly stated that statistics is the science of uncertainty. Half of the respondents accepted that they have problem of writing the statistical section of the article. 64% of the subjects indicated that statistical teaching methods were the main reasons for the impression of its difficulties. 53% of the faculty indicated that the co-authorship of the statistician should depend upon his/her contribution in the study. Gender did not show any significant difference among the responses. However, senior faculty showed higher level of the importance for the use of statistics and difficulties of writing result section of articles as compared to junior faculty. The study showed a low level of knowledge, but high level of the awareness for the use of statistical techniques in research and exhibited a good level of motivation for further training.
Paechter, Manuela; Macher, Daniel; Martskvishvili, Khatuna; Wimmer, Sigrid; Papousek, Ilona
2017-01-01
In many social science majors, e.g., psychology, students report high levels of statistics anxiety. However, these majors are often chosen by students who are less prone to mathematics and who might have experienced difficulties and unpleasant feelings in their mathematics courses at school. The present study investigates whether statistics anxiety is a genuine form of anxiety that impairs students' achievements or whether learners mainly transfer previous experiences in mathematics and their anxiety in mathematics to statistics. The relationship between mathematics anxiety and statistics anxiety, their relationship to learning behaviors and to performance in a statistics examination were investigated in a sample of 225 undergraduate psychology students (164 women, 61 men). Data were recorded at three points in time: At the beginning of term students' mathematics anxiety, general proneness to anxiety, school grades, and demographic data were assessed; 2 weeks before the end of term, they completed questionnaires on statistics anxiety and their learning behaviors. At the end of term, examination scores were recorded. Mathematics anxiety and statistics anxiety correlated highly but the comparison of different structural equation models showed that they had genuine and even antagonistic contributions to learning behaviors and performance in the examination. Surprisingly, mathematics anxiety was positively related to performance. It might be that students realized over the course of their first term that knowledge and skills in higher secondary education mathematics are not sufficient to be successful in statistics. Part of mathematics anxiety may then have strengthened positive extrinsic effort motivation by the intention to avoid failure and may have led to higher effort for the exam preparation. However, via statistics anxiety mathematics anxiety also had a negative contribution to performance. Statistics anxiety led to higher procrastination in the structural equation model and, therefore, contributed indirectly and negatively to performance. Furthermore, it had a direct negative impact on performance (probably via increased tension and worry in the exam). The results of the study speak for shared but also unique components of statistics anxiety and mathematics anxiety. They are also important for instruction and give recommendations to learners as well as to instructors.
2018-04-30
2017 Workplace and Gender Relations Survey of Reserve Component Members Statistical Methodology Report Additional copies of this report...Survey of Reserve Component Members Statistical Methodology Report Office of People Analytics (OPA) 4800 Mark Center Drive, Suite...RESERVE COMPONENT MEMBERS STATISTICAL METHODOLOGY REPORT Introduction The Office of People Analytics’ Center for Health and Resilience (OPA[H&R
Code of Federal Regulations, 2010 CFR
2010-07-01
... 30 Mineral Resources 2 2010-07-01 2010-07-01 false What reports and statistics must I submit... statistics must I submit relating to a hurricane, earthquake, or other natural occurrence? (a) You must submit evacuation statistics to the Regional Supervisor for a natural occurrence, such as a hurricane, a...
Code of Federal Regulations, 2011 CFR
2011-07-01
... relating to a hurricane, earthquake, or other natural occurrence? 250.192 Section 250.192 Mineral Resources... Requirements § 250.192 What reports and statistics must I submit relating to a hurricane, earthquake, or other... occurrence, such as a hurricane, a tropical storm, or an earthquake. Statistics include facilities and rigs...
Galloway, Joel M.
2014-01-01
The Red River of the North (hereafter referred to as “Red River”) Basin is an important hydrologic region where water is a valuable resource for the region’s economy. Continuous water-quality monitors have been operated by the U.S. Geological Survey, in cooperation with the North Dakota Department of Health, Minnesota Pollution Control Agency, City of Fargo, City of Moorhead, City of Grand Forks, and City of East Grand Forks at the Red River at Fargo, North Dakota, from 2003 through 2012 and at Grand Forks, N.Dak., from 2007 through 2012. The purpose of the monitoring was to provide a better understanding of the water-quality dynamics of the Red River and provide a way to track changes in water quality. Regression equations were developed that can be used to estimate concentrations and loads for dissolved solids, sulfate, chloride, nitrate plus nitrite, total phosphorus, and suspended sediment using explanatory variables such as streamflow, specific conductance, and turbidity. Specific conductance was determined to be a significant explanatory variable for estimating dissolved solids concentrations at the Red River at Fargo and Grand Forks. The regression equations provided good relations between dissolved solid concentrations and specific conductance for the Red River at Fargo and at Grand Forks, with adjusted coefficients of determination of 0.99 and 0.98, respectively. Specific conductance, log-transformed streamflow, and a seasonal component were statistically significant explanatory variables for estimating sulfate in the Red River at Fargo and Grand Forks. Regression equations provided good relations between sulfate concentrations and the explanatory variables, with adjusted coefficients of determination of 0.94 and 0.89, respectively. For the Red River at Fargo and Grand Forks, specific conductance, streamflow, and a seasonal component were statistically significant explanatory variables for estimating chloride. For the Red River at Grand Forks, a time component also was a statistically significant explanatory variable for estimating chloride. The regression equations for chloride at the Red River at Fargo provided a fair relation between chloride concentrations and the explanatory variables, with an adjusted coefficient of determination of 0.66 and the equation for the Red River at Grand Forks provided a relatively good relation between chloride concentrations and the explanatory variables, with an adjusted coefficient of determination of 0.77. Turbidity and streamflow were statistically significant explanatory variables for estimating nitrate plus nitrite concentrations at the Red River at Fargo and turbidity was the only statistically significant explanatory variable for estimating nitrate plus nitrite concentrations at Grand Forks. The regression equation for the Red River at Fargo provided a relatively poor relation between nitrate plus nitrite concentrations, turbidity, and streamflow, with an adjusted coefficient of determination of 0.46. The regression equation for the Red River at Grand Forks provided a fair relation between nitrate plus nitrite concentrations and turbidity, with an adjusted coefficient of determination of 0.73. Some of the variability that was not explained by the equations might be attributed to different sources contributing nitrates to the stream at different times. Turbidity, streamflow, and a seasonal component were statistically significant explanatory variables for estimating total phosphorus at the Red River at Fargo and Grand Forks. The regression equation for the Red River at Fargo provided a relatively fair relation between total phosphorus concentrations, turbidity, streamflow, and season, with an adjusted coefficient of determination of 0.74. The regression equation for the Red River at Grand Forks provided a good relation between total phosphorus concentrations, turbidity, streamflow, and season, with an adjusted coefficient of determination of 0.87. For the Red River at Fargo, turbidity and streamflow were statistically significant explanatory variables for estimating suspended-sediment concentrations. For the Red River at Grand Forks, turbidity was the only statistically significant explanatory variable for estimating suspended-sediment concentration. The regression equation at the Red River at Fargo provided a good relation between suspended-sediment concentration, turbidity, and streamflow, with an adjusted coefficient of determination of 0.95. The regression equation for the Red River at Grand Forks provided a good relation between suspended-sediment concentration and turbidity, with an adjusted coefficient of determination of 0.96.
Öztin, Hasan; Çağıltay, Eylem; Çağlayan, Sinan; Kaplan, Mustafa; Akpak, Yaşam Kemal; Karaca, Nilay; Tığlıoğlu, Mesut
2016-12-01
Male hypogonadism is defined as the deficiency of testosterone or sperm production synthesized by testicles or the deficiency of both. The reasons for hypogonadism may be primary, meaning testicular or secondary, meaning hypothalamohypophyseal. In hypogonadotropic hypogonadism (HH), there is indeficiency in gonadotropic hormones due to hypothalamic or hypophyseal reasons. Gonadotropin-releasing hormone (GnRH) is an important stimulant in releasing follicular stimulant hormone (FSH), mainly luteinizing hormone (LH). GnRH omitted is under the effect of many hormonal or stimulating factors. Kisspeptin is present in many places of the body, mostly in hypothalamic anteroventral periventricular nucleus and arcuate nucleus. Kisspeptin has a suppressor effect on the metastasis of many tumors such as breast cancer and malign melanoma metastases, and is called "metastin" for this reason. Kisspeptin is a strong stimulant of GnRH. In idiopathic hypogonadotropic hypogonadism (IHH) etiology, there is gonadotropic hormone release indeficiency which cannot be clearly described. A total of 30 male hypogonatropic hypogonadism diagnosed patients over 30 years of age who have applied to Haydarpasa Education Hospital Endocrinology and Metabolic Diseases Service were included in the study. Compared to the control group, the effect of kisspeptin on male patients with hypogonatropic hypogonadism and on insulin resistance developing in hypogonadism patients was investigated in our study. A statistically significant difference was detected between average kisspeptin measurements of the groups (p < 0.01). Kisspeptin measurement of the cases in the patient group were detected significantly high. No statistically significant relation was detected among kisspeptin and LH/FSH levels. Although a positive low relation was detected between kisspeptin measurements of patient group cases and homeostasis model assessment of insulin resistance (HOMA-IR) measurements, this relation was statistically insignificant. When the patient and control groups were compared for HOMA-IR, no statistically significant difference was detected. The reason for high kisspeptin levels in the patient group compared to the control group makes us consider that there may be a GPR54 resistance or GnRH neuronal transfer pathway defect. When patients and control groups were compared for HOMA-IR, the difference was not statistically significant. It is considered that kisspeptin is one of the reasons for hypogonatropic hypogonadism and has less effect on insulin resistance.
Federal Register 2010, 2011, 2012, 2013, 2014
2010-12-28
... Statistics (IMF Statistics) for SRF. SRF received exemptions from import duties and central sales taxes (CST... Territory Sales Tax Exemption program, we have used SRF's total sales of subject merchandise as the... were certified by an accountant. The total values of the GOI redemption document reflect the import and...
Decisions that Make a Difference in Detecting Differential Item Functioning
ERIC Educational Resources Information Center
Sireci, Stephen G.; Rios, Joseph A.
2013-01-01
There are numerous statistical procedures for detecting items that function differently across subgroups of examinees that take a test or survey. However, in endeavouring to detect items that may function differentially, selection of the statistical method is only one of many important decisions. In this article, we discuss the important decisions…
Using Microsoft Excel[R] to Calculate Descriptive Statistics and Create Graphs
ERIC Educational Resources Information Center
Carr, Nathan T.
2008-01-01
Descriptive statistics and appropriate visual representations of scores are important for all test developers, whether they are experienced testers working on large-scale projects, or novices working on small-scale local tests. Many teachers put in charge of testing projects do not know "why" they are important, however, and are utterly convinced…
Counting on COUNTER: The Current State of E-Resource Usage Data in Libraries
ERIC Educational Resources Information Center
Welker, Josh
2012-01-01
Any librarian who has managed electronic resources has experienced the--for want of words--"joy" of gathering and analyzing usage statistics. Such statistics are important for evaluating the effectiveness of resources and for making important budgeting decisions. Unfortunately, the data are usually tedious to collect, inconsistently organized, of…
Comparing the Lifetimes of Two Brands of Batteries
ERIC Educational Resources Information Center
Dunn, Peter K.
2013-01-01
In this paper, we report a case study that illustrates the importance in interpreting the results from statistical tests, and shows the difference between practical importance and statistical significance. This case study presents three sets of data concerning the performance of two brands of batteries. The data are easy to describe and…
Simple Data Sets for Distinct Basic Summary Statistics
ERIC Educational Resources Information Center
Lesser, Lawrence M.
2011-01-01
It is important to avoid ambiguity with numbers because unfortunate choices of numbers can inadvertently make it possible for students to form misconceptions or make it difficult for teachers to tell if students obtained the right answer for the right reason. Therefore, it is important to make sure when introducing basic summary statistics that…
Eng, John; Wilson, Renee F; Subramaniam, Rathan M; Zhang, Allen; Suarez-Cuervo, Catalina; Turban, Sharon; Choi, Michael J; Sherrod, Cheryl; Hutfless, Susan; Iyoha, Emmanuel E; Bass, Eric B
2016-03-15
Iodine contrast media are essential components of many imaging procedures. An important potential side effect is contrast-induced nephropathy (CIN). To compare CIN risk for contrast media within and between osmolality classes in patients receiving diagnostic or therapeutic imaging procedures. PubMed, EMBASE, Cochrane Library, Clinical Trials.gov, and Scopus through June 2015. Randomized, controlled trials that reported CIN-related outcomes in patients receiving low-osmolar contrast media (LOCM) or iso-osmolar contrast media for imaging. Independent study selection and quality assessment by 2 reviewers and dual extraction of study characteristics and results. None of the 5 studies that compared types of LOCM reported a statistically significant or clinically important difference among study groups, but the strength of evidence was low. Twenty-five randomized, controlled trials found a slight reduction in CIN risk with the iso-osmolar contrast media agent iodixanol compared with a diverse group of LOCM that just reached statistical significance in a meta-analysis (pooled relative risk, 0.80 [95% CI, 0.65 to 0.99]; P = 0.045). This comparison's strength of evidence was moderate. In a meta regression of randomized, controlled trials of iodixanol, no relationship was found between route of administration and comparative CIN risk. Few studies compared LOCM. Procedural details about contrast administration were not uniformly reported. Few studies specified clinical indications or severity of baseline renal impairment. No differences were found in CIN risk among types of LOCM. Iodixanol had a slightly lower risk for CIN than LOCM, but the lower risk did not exceed a criterion for clinical importance. Agency for Healthcare Research and Quality.
Wood, Molly S.; Fosness, Ryan L.
2013-01-01
The U.S. Geological Survey, in cooperation with the Bureau of Land Management (BLM), collected streamflow data in 2012 and estimated streamflow statistics for stream segments designated "Wild," "Scenic," or "Recreational" under the National Wild and Scenic Rivers System in the Owyhee Canyonlands Wilderness in southwestern Idaho. The streamflow statistics were used by BLM to develop and file a draft, federal reserved water right claim in autumn 2012 to protect federally designated "outstanding remarkable values" in the stream segments. BLM determined that the daily mean streamflow equaled or exceeded 20 and 80 percent of the time during bimonthly periods (two periods per month) and the bankfull streamflow are important streamflow thresholds for maintaining outstanding remarkable values. Prior to this study, streamflow statistics estimated using available datasets and tools for the Owyhee Canyonlands Wilderness were inaccurate for use in the water rights claim. Streamflow measurements were made at varying intervals during February–September 2012 at 14 monitoring sites; 2 of the monitoring sites were equipped with telemetered streamgaging equipment. Synthetic streamflow records were created for 11 of the 14 monitoring sites using a partial‑record method or a drainage-area-ratio method. Streamflow records were obtained directly from an operating, long-term streamgage at one monitoring site, and from discontinued streamgages at two monitoring sites. For 10 sites analyzed using the partial-record method, discrete measurements were related to daily mean streamflow at a nearby, telemetered “index” streamgage. Resulting regression equations were used to estimate daily mean and annual peak streamflow at the monitoring sites during the full period of record for the index sites. A synthetic streamflow record for Sheep Creek was developed using a drainage-area-ratio method, because measured streamflows did not relate well to any index site to allow use of the partial-record method. The synthetic and actual daily mean streamflow records were used to estimate daily mean streamflow that was exceeded 80, 50, and 20 percent of the time (80-, 50-, and 20-percent exceedances) for bimonthly and annual periods. Bankfull streamflow statistics were calculated by fitting the synthetic and actual annual peak streamflow records to a log Pearson Type III distribution using Bulletin 17B guidelines in the U.S. Geological Survey PeakFQ program. The coefficients of determination (R2) for the regressions between the monitoring and index sites ranged from 0.74 for Wickahoney Creek to 0.98 for the West Fork Bruneau River and Deep Creek. Confidence in computed streamflow statistics is highest among other sites for the East Fork Owyhee River and the West Fork Bruneau River on the basis of regression statistics, visual fit of the related data, and the range and number of streamflow measurements. Streamflow statistics for sites with the greatest uncertainty included Big Jacks, Little Jacks, Cottonwood, Wickahoney, and Sheep Creeks. The uncertainty in computed streamflow statistics was due to a number of factors which included the distance of index sites relative to monitoring sites, relatively low streamflow conditions that occurred during the study, and the limited number and range of streamflow measurements. However, the computed streamflow statistics are considered the best possible estimates given available datasets in the remote study area. Streamflow measurements over a wider range of hydrologic and climatic conditions would improve the relations between streamflow characteristics at monitoring and index sites. Additionally, field surveys are needed to verify if the streamflows selected for the water rights claims are sufficient for maintaining outstanding remarkable values in the Wild and Scenic rivers included in the study.
Complex Network Analysis for Characterizing Global Value Chains in Equipment Manufacturing
Meng, Bo; Cheng, Lihong
2017-01-01
The rise of global value chains (GVCs) characterized by the so-called “outsourcing”, “fragmentation production”, and “trade in tasks” has been considered one of the most important phenomena for the 21st century trade. GVCs also can play a decisive role in trade policy making. However, due to the increasing complexity and sophistication of international production networks, especially in the equipment manufacturing industry, conventional trade statistics and the corresponding trade indicators may give us a distorted picture of trade. This paper applies various network analysis tools to the new GVC accounting system proposed by Koopman et al. (2014) and Wang et al. (2013) in which gross exports can be decomposed into value-added terms through various routes along GVCs. This helps to divide the equipment manufacturing-related GVCs into some sub-networks with clear visualization. The empirical results of this paper significantly improve our understanding of the topology of equipment manufacturing-related GVCs as well as the interdependency of countries in these GVCs that is generally invisible from the traditional trade statistics. PMID:28081201