NASA Technical Reports Server (NTRS)
1980-01-01
MATHPAC image-analysis library is collection of general-purpose mathematical and statistical routines and special-purpose data-analysis and pattern-recognition routines for image analysis. MATHPAC library consists of Linear Algebra, Optimization, Statistical-Summary, Densities and Distribution, Regression, and Statistical-Test packages.
Policy Safeguards and the Legitimacy of Highway Interdiction
2016-12-01
17 B. BIAS WITHIN LAW ENFORCEMENT ..............................................19 C. STATISTICAL DATA GATHERING...32 3. Controlling Discretion .................................................................36 4. Statistical Data Collection for Traffic Stops...49 A. DESCRIPTION OF STATISTICAL DATA COLLECTED ...............50 B. DATA ORGANIZATION AND ANALYSIS
ERIC Educational Resources Information Center
Lau, Joann M.; Korn, Robert W.
2007-01-01
In this article, the authors present a laboratory exercise in data collection and statistical analysis in biological space using clustered stomates on leaves of "Begonia" plants. The exercise can be done in middle school classes by students making their own slides and seeing imprints of cells, or at the high school level through collecting data of…
Prison Radicalization: The New Extremist Training Grounds?
2007-09-01
distributing and collecting survey data , and the data analysis. The analytical methodology includes descriptive and inferential statistical methods, in... statistical analysis of the responses to identify significant correlations and relationships. B. SURVEY DATA COLLECTION To effectively access a...Q18, Q19, Q20, and Q21. Due to the exploratory nature of this small survey, data analyses were confined mostly to descriptive statistics and
Eagle Plus Air Superiority into the 21st Century
1996-04-01
18 Data Collection Method ....................................................................................... 18 Statistical Trend Analysis...19 Statistical Readiness Analysis.................................................................................... 20 Aging Aircraft...generated by Mr. Jeff Hill served as the foundation of our statistical analysis. Special thanks go out to Mrs. Betsy Mullis, LFLL branch chief, and to
78 FR 34101 - Agency Information Collection Activities: Proposed Collection; Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2013-06-06
... and basic descriptive statistics on the quantity and type of consumer-reported patient safety events... conduct correlations, cross tabulations of responses and other statistical analysis. Estimated Annual...
Statistical summaries of selected Iowa streamflow data through September 2013.
DOT National Transportation Integrated Search
2015-01-01
Statistical summaries of streamflow data collected at : 184 streamgages in Iowa are presented in this report. All : streamgages included for analysis have at least 10 years of : continuous record collected before or through September : 2013. This rep...
Statistical analysis in MSW collection performance assessment.
Teixeira, Carlos Afonso; Avelino, Catarina; Ferreira, Fátima; Bentes, Isabel
2014-09-01
The increase of Municipal Solid Waste (MSW) generated over the last years forces waste managers pursuing more effective collection schemes, technically viable, environmentally effective and economically sustainable. The assessment of MSW services using performance indicators plays a crucial role for improving service quality. In this work, we focus on the relevance of regular system monitoring as a service assessment tool. In particular, we select and test a core-set of MSW collection performance indicators (effective collection distance, effective collection time and effective fuel consumption) that highlights collection system strengths and weaknesses and supports pro-active management decision-making and strategic planning. A statistical analysis was conducted with data collected in mixed collection system of Oporto Municipality, Portugal, during one year, a week per month. This analysis provides collection circuits' operational assessment and supports effective short-term municipality collection strategies at the level of, e.g., collection frequency and timetables, and type of containers. Copyright © 2014 Elsevier Ltd. All rights reserved.
Preparing for the first meeting with a statistician.
De Muth, James E
2008-12-15
Practical statistical issues that should be considered when performing data collection and analysis are reviewed. The meeting with a statistician should take place early in the research development before any study data are collected. The process of statistical analysis involves establishing the research question, formulating a hypothesis, selecting an appropriate test, sampling correctly, collecting data, performing tests, and making decisions. Once the objectives are established, the researcher can determine the characteristics or demographics of the individuals required for the study, how to recruit volunteers, what type of data are needed to answer the research question(s), and the best methods for collecting the required information. There are two general types of statistics: descriptive and inferential. Presenting data in a more palatable format for the reader is called descriptive statistics. Inferential statistics involve making an inference or decision about a population based on results obtained from a sample of that population. In order for the results of a statistical test to be valid, the sample should be representative of the population from which it is drawn. When collecting information about volunteers, researchers should only collect information that is directly related to the study objectives. Important information that a statistician will require first is an understanding of the type of variables involved in the study and which variables can be controlled by researchers and which are beyond their control. Data can be presented in one of four different measurement scales: nominal, ordinal, interval, or ratio. Hypothesis testing involves two mutually exclusive and exhaustive statements related to the research question. Statisticians should not be replaced by computer software, and they should be consulted before any research data are collected. When preparing to meet with a statistician, the pharmacist researcher should be familiar with the steps of statistical analysis and consider several questions related to the study to be conducted.
Statistical analysis of the national crash severity study data
DOT National Transportation Integrated Search
1980-08-01
This is the Final Report on a two-year statistical analysis of the data collected in the National Crash Severity Study (NCSS). The analysis presented is primarily concerned with the relationship between occupant injury severity and the crash conditio...
Using SPSS to Analyze Book Collection Data.
ERIC Educational Resources Information Center
Townley, Charles T.
1981-01-01
Describes and illustrates Statistical Package for the Social Sciences (SPSS) procedures appropriate for book collection data analysis. Several different procedures for univariate, bivariate, and multivariate analysis are discussed, and applications of procedures for book collection studies are presented. Included are 24 tables illustrating output…
The Importance of Statistical Modeling in Data Analysis and Inference
ERIC Educational Resources Information Center
Rollins, Derrick, Sr.
2017-01-01
Statistical inference simply means to draw a conclusion based on information that comes from data. Error bars are the most commonly used tool for data analysis and inference in chemical engineering data studies. This work demonstrates, using common types of data collection studies, the importance of specifying the statistical model for sound…
Davis, J.C.
2000-01-01
Geologists may feel that geological data are not amenable to statistical analysis, or at best require specialized approaches such as nonparametric statistics and geostatistics. However, there are many circumstances, particularly in systematic studies conducted for environmental or regulatory purposes, where traditional parametric statistical procedures can be beneficial. An example is the application of analysis of variance to data collected in an annual program of measuring groundwater levels in Kansas. Influences such as well conditions, operator effects, and use of the water can be assessed and wells that yield less reliable measurements can be identified. Such statistical studies have resulted in yearly improvements in the quality and reliability of the collected hydrologic data. Similar benefits may be achieved in other geological studies by the appropriate use of classical statistical tools.
2000-04-10
interest. These include Statistical Energy Analysis (SEA), fuzzy structure theory, and approaches combining modal analysis and SEA. Non-determinism...34 arising with increasing frequency. This has led to Statistical Energy Analysis , in which a system is modelled as a collection of coupled subsystems...22. IUTAM Symposium on Statistical Energy Analysis . 1999 Ed. F.J. Fahy and W.G. Price. Kluwer Academic Publishing. • 23. R.S. Langley and P
Safety Management Information Statistics (SAMIS) - 1993 Annual Report
DOT National Transportation Integrated Search
1995-05-01
The 1993 Safety Management Information Statistics (SAMIS) report, now in its fourth year of publication, is a compilation and analysis of transit accident and casualty statistics uniformly collected from approximately 400 transit agencies throughout ...
PV System Component Fault and Failure Compilation and Analysis.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Klise, Geoffrey Taylor; Lavrova, Olga; Gooding, Renee Lynne
This report describes data collection and analysis of solar photovoltaic (PV) equipment events, which consist of faults and fa ilures that occur during the normal operation of a distributed PV system or PV power plant. We present summary statistics from locations w here maintenance data is being collected at various intervals, as well as reliability statistics gathered from that da ta, consisting of fault/failure distributions and repair distributions for a wide range of PV equipment types.
Image analysis library software development
NASA Technical Reports Server (NTRS)
Guseman, L. F., Jr.; Bryant, J.
1977-01-01
The Image Analysis Library consists of a collection of general purpose mathematical/statistical routines and special purpose data analysis/pattern recognition routines basic to the development of image analysis techniques for support of current and future Earth Resources Programs. Work was done to provide a collection of computer routines and associated documentation which form a part of the Image Analysis Library.
SPA- STATISTICAL PACKAGE FOR TIME AND FREQUENCY DOMAIN ANALYSIS
NASA Technical Reports Server (NTRS)
Brownlow, J. D.
1994-01-01
The need for statistical analysis often arises when data is in the form of a time series. This type of data is usually a collection of numerical observations made at specified time intervals. Two kinds of analysis may be performed on the data. First, the time series may be treated as a set of independent observations using a time domain analysis to derive the usual statistical properties including the mean, variance, and distribution form. Secondly, the order and time intervals of the observations may be used in a frequency domain analysis to examine the time series for periodicities. In almost all practical applications, the collected data is actually a mixture of the desired signal and a noise signal which is collected over a finite time period with a finite precision. Therefore, any statistical calculations and analyses are actually estimates. The Spectrum Analysis (SPA) program was developed to perform a wide range of statistical estimation functions. SPA can provide the data analyst with a rigorous tool for performing time and frequency domain studies. In a time domain statistical analysis the SPA program will compute the mean variance, standard deviation, mean square, and root mean square. It also lists the data maximum, data minimum, and the number of observations included in the sample. In addition, a histogram of the time domain data is generated, a normal curve is fit to the histogram, and a goodness-of-fit test is performed. These time domain calculations may be performed on both raw and filtered data. For a frequency domain statistical analysis the SPA program computes the power spectrum, cross spectrum, coherence, phase angle, amplitude ratio, and transfer function. The estimates of the frequency domain parameters may be smoothed with the use of Hann-Tukey, Hamming, Barlett, or moving average windows. Various digital filters are available to isolate data frequency components. Frequency components with periods longer than the data collection interval are removed by least-squares detrending. As many as ten channels of data may be analyzed at one time. Both tabular and plotted output may be generated by the SPA program. This program is written in FORTRAN IV and has been implemented on a CDC 6000 series computer with a central memory requirement of approximately 142K (octal) of 60 bit words. This core requirement can be reduced by segmentation of the program. The SPA program was developed in 1978.
Methods for collection and analysis of aquatic biological and microbiological samples
Greeson, Phillip E.; Ehlke, T.A.; Irwin, G.A.; Lium, B.W.; Slack, K.V.
1977-01-01
Chapter A4 contains methods used by the U.S. Geological Survey to collect, preserve, and analyze waters to determine their biological and microbiological properties. Part 1 discusses biological sampling and sampling statistics. The statistical procedures are accompanied by examples. Part 2 consists of detailed descriptions of more than 45 individual methods, including those for bacteria, phytoplankton, zooplankton, seston, periphyton, macrophytes, benthic invertebrates, fish and other vertebrates, cellular contents, productivity, and bioassays. Each method is summarized, and the application, interferences, apparatus, reagents, collection, analysis, calculations, reporting of results, precision and references are given. Part 3 consists of a glossary. Part 4 is a list of taxonomic references.
Tanavalee, Chotetawan; Luksanapruksa, Panya; Singhatanadgige, Weerasak
2016-06-01
Microsoft Excel (MS Excel) is a commonly used program for data collection and statistical analysis in biomedical research. However, this program has many limitations, including fewer functions that can be used for analysis and a limited number of total cells compared with dedicated statistical programs. MS Excel cannot complete analyses with blank cells, and cells must be selected manually for analysis. In addition, it requires multiple steps of data transformation and formulas to plot survival analysis graphs, among others. The Megastat add-on program, which will be supported by MS Excel 2016 soon, would eliminate some limitations of using statistic formulas within MS Excel.
This analysis updates EPA's standard VSL estimate by using a more comprehensive collection of VSL studies that include studies published between 1992 and 2000, as well as applying a more appropriate statistical method. We provide a pooled effect VSL estimate by applying the empi...
Swetha, Jonnalagadda Laxmi; Arpita, Ramisetti; Srikanth, Chintalapani; Nutalapati, Rajasekhar
2014-01-01
Biostatistics is an integral part of research protocols. In any field of inquiry or investigation, data obtained is subsequently classified, analyzed and tested for accuracy by statistical methods. Statistical analysis of collected data, thus, forms the basis for all evidence-based conclusions. The aim of this study is to evaluate the cognition, comprehension and application of biostatistics in research among post graduate students in Periodontics, in India. A total of 391 post graduate students registered for a master's course in periodontics at various dental colleges across India were included in the survey. Data regarding the level of knowledge, understanding and its application in design and conduct of the research protocol was collected using a dichotomous questionnaire. A descriptive statistics was used for data analysis. Nearly 79.2% students were aware of the importance of biostatistics in research, 55-65% were familiar with MS-EXCEL spreadsheet for graphical representation of data and with the statistical softwares available on the internet, 26.0% had biostatistics as mandatory subject in their curriculum, 9.5% tried to perform statistical analysis on their own while 3.0% were successful in performing statistical analysis of their studies on their own. Biostatistics should play a central role in planning, conduct, interim analysis, final analysis and reporting of periodontal research especially by the postgraduate students. Indian postgraduate students in periodontics are aware of the importance of biostatistics in research but the level of understanding and application is still basic and needs to be addressed.
DOT National Transportation Integrated Search
2015-01-01
Statistical summaries of streamflow data collected at 184 streamgages in Iowa are presented in this report. All streamgages included for analysis have at least 10 years of continuous record collected before or through September 2013. This report is a...
The Global Oscillation Network Group site survey. 1: Data collection and analysis methods
NASA Technical Reports Server (NTRS)
Hill, Frank; Fischer, George; Grier, Jennifer; Leibacher, John W.; Jones, Harrison B.; Jones, Patricia P.; Kupke, Renate; Stebbins, Robin T.
1994-01-01
The Global Oscillation Network Group (GONG) Project is planning to place a set of instruments around the world to observe solar oscillations as continuously as possible for at least three years. The Project has now chosen the sites that will comprise the network. This paper describes the methods of data collection and analysis that were used to make this decision. Solar irradiance data were collected with a one-minute cadence at fifteen sites around the world and analyzed to produce statistics of cloud cover, atmospheric extinction, and transparency power spectra at the individual sites. Nearly 200 reasonable six-site networks were assembled from the individual stations, and a set of statistical measures of the performance of the networks was analyzed using a principal component analysis. An accompanying paper presents the results of the survey.
ERIC Educational Resources Information Center
Peterlin, Primoz
2010-01-01
Two methods of data analysis are compared: spreadsheet software and a statistics software suite. Their use is compared analysing data collected in three selected experiments taken from an introductory physics laboratory, which include a linear dependence, a nonlinear dependence and a histogram. The merits of each method are compared. (Contains 7…
[Notes on vital statistics for the study of perinatal health].
Juárez, Sol Pía
2014-01-01
Vital statistics, published by the National Statistics Institute in Spain, are a highly important source for the study of perinatal health nationwide. However, the process of data collection is not well-known and has implications both for the quality and interpretation of the epidemiological results derived from this source. The aim of this study was to present how the information is collected and some of the associated problems. This study is the result of an analysis of the methodological notes from the National Statistics Institute and first-hand information obtained from hospitals, the Central Civil Registry of Madrid, and the Madrid Institute for Statistics. Greater integration between these institutions is required to improve the quality of birth and stillbirth statistics. Copyright © 2014 SESPAS. Published by Elsevier Espana. All rights reserved.
Experimental toxicology: Issues of statistics, experimental design, and replication.
Briner, Wayne; Kirwan, Jeral
2017-01-01
The difficulty of replicating experiments has drawn considerable attention. Issues with replication occur for a variety of reasons ranging from experimental design to laboratory errors to inappropriate statistical analysis. Here we review a variety of guidelines for statistical analysis, design, and execution of experiments in toxicology. In general, replication can be improved by using hypothesis driven experiments with adequate sample sizes, randomization, and blind data collection techniques. Copyright © 2016 Elsevier B.V. All rights reserved.
Statistics for nuclear engineers and scientists. Part 1. Basic statistical inference
DOE Office of Scientific and Technical Information (OSTI.GOV)
Beggs, W.J.
1981-02-01
This report is intended for the use of engineers and scientists working in the nuclear industry, especially at the Bettis Atomic Power Laboratory. It serves as the basis for several Bettis in-house statistics courses. The objectives of the report are to introduce the reader to the language and concepts of statistics and to provide a basic set of techniques to apply to problems of the collection and analysis of data. Part 1 covers subjects of basic inference. The subjects include: descriptive statistics; probability; simple inference for normally distributed populations, and for non-normal populations as well; comparison of two populations; themore » analysis of variance; quality control procedures; and linear regression analysis.« less
Swetha, Jonnalagadda Laxmi; Arpita, Ramisetti; Srikanth, Chintalapani; Nutalapati, Rajasekhar
2014-01-01
Background: Biostatistics is an integral part of research protocols. In any field of inquiry or investigation, data obtained is subsequently classified, analyzed and tested for accuracy by statistical methods. Statistical analysis of collected data, thus, forms the basis for all evidence-based conclusions. Aim: The aim of this study is to evaluate the cognition, comprehension and application of biostatistics in research among post graduate students in Periodontics, in India. Materials and Methods: A total of 391 post graduate students registered for a master's course in periodontics at various dental colleges across India were included in the survey. Data regarding the level of knowledge, understanding and its application in design and conduct of the research protocol was collected using a dichotomous questionnaire. A descriptive statistics was used for data analysis. Results: Nearly 79.2% students were aware of the importance of biostatistics in research, 55-65% were familiar with MS-EXCEL spreadsheet for graphical representation of data and with the statistical softwares available on the internet, 26.0% had biostatistics as mandatory subject in their curriculum, 9.5% tried to perform statistical analysis on their own while 3.0% were successful in performing statistical analysis of their studies on their own. Conclusion: Biostatistics should play a central role in planning, conduct, interim analysis, final analysis and reporting of periodontal research especially by the postgraduate students. Indian postgraduate students in periodontics are aware of the importance of biostatistics in research but the level of understanding and application is still basic and needs to be addressed. PMID:24744547
BrightStat.com: free statistics online.
Stricker, Daniel
2008-10-01
Powerful software for statistical analysis is expensive. Here I present BrightStat, a statistical software running on the Internet which is free of charge. BrightStat's goals, its main capabilities and functionalities are outlined. Three different sample runs, a Friedman test, a chi-square test, and a step-wise multiple regression are presented. The results obtained by BrightStat are compared with results computed by SPSS, one of the global leader in providing statistical software, and VassarStats, a collection of scripts for data analysis running on the Internet. Elementary statistics is an inherent part of academic education and BrightStat is an alternative to commercial products.
Statistical Analysis Tools for Learning in Engineering Laboratories.
ERIC Educational Resources Information Center
Maher, Carolyn A.
1990-01-01
Described are engineering programs that have used automated data acquisition systems to implement data collection and analyze experiments. Applications include a biochemical engineering laboratory, heat transfer performance, engineering materials testing, mechanical system reliability, statistical control laboratory, thermo-fluid laboratory, and a…
Applied statistics in agricultural, biological, and environmental sciences.
USDA-ARS?s Scientific Manuscript database
Agronomic research often involves measurement and collection of multiple response variables in an effort to understand the more complex nature of the system being studied. Multivariate statistical methods encompass the simultaneous analysis of all random variables measured on each experimental or s...
GSuite HyperBrowser: integrative analysis of dataset collections across the genome and epigenome.
Simovski, Boris; Vodák, Daniel; Gundersen, Sveinung; Domanska, Diana; Azab, Abdulrahman; Holden, Lars; Holden, Marit; Grytten, Ivar; Rand, Knut; Drabløs, Finn; Johansen, Morten; Mora, Antonio; Lund-Andersen, Christin; Fromm, Bastian; Eskeland, Ragnhild; Gabrielsen, Odd Stokke; Ferkingstad, Egil; Nakken, Sigve; Bengtsen, Mads; Nederbragt, Alexander Johan; Thorarensen, Hildur Sif; Akse, Johannes Andreas; Glad, Ingrid; Hovig, Eivind; Sandve, Geir Kjetil
2017-07-01
Recent large-scale undertakings such as ENCODE and Roadmap Epigenomics have generated experimental data mapped to the human reference genome (as genomic tracks) representing a variety of functional elements across a large number of cell types. Despite the high potential value of these publicly available data for a broad variety of investigations, little attention has been given to the analytical methodology necessary for their widespread utilisation. We here present a first principled treatment of the analysis of collections of genomic tracks. We have developed novel computational and statistical methodology to permit comparative and confirmatory analyses across multiple and disparate data sources. We delineate a set of generic questions that are useful across a broad range of investigations and discuss the implications of choosing different statistical measures and null models. Examples include contrasting analyses across different tissues or diseases. The methodology has been implemented in a comprehensive open-source software system, the GSuite HyperBrowser. To make the functionality accessible to biologists, and to facilitate reproducible analysis, we have also developed a web-based interface providing an expertly guided and customizable way of utilizing the methodology. With this system, many novel biological questions can flexibly be posed and rapidly answered. Through a combination of streamlined data acquisition, interoperable representation of dataset collections, and customizable statistical analysis with guided setup and interpretation, the GSuite HyperBrowser represents a first comprehensive solution for integrative analysis of track collections across the genome and epigenome. The software is available at: https://hyperbrowser.uio.no. © The Author 2017. Published by Oxford University Press.
Taeyoung Kim; Christian Langpap
2015-01-01
This report provides a statistical analysis of the data collected from two survey regions of the United States, the Pacific Northwest and the Southeast. The survey asked about individual agricultural landownersâ characteristics, characteristics of their land, and the landownersâ willingness to enroll in a tree planting program under incentive payments for carbon...
NASA Astrophysics Data System (ADS)
Mottyll, S.; Skoda, R.
2015-12-01
A compressible inviscid flow solver with barotropic cavitation model is applied to two different ultrasonic horn set-ups and compared to hydrophone, shadowgraphy as well as erosion test data. The statistical analysis of single collapse events in wall-adjacent flow regions allows the determination of the flow aggressiveness via load collectives (cumulative event rate vs collapse pressure), which show an exponential decrease in agreement to studies on hydrodynamic cavitation [1]. A post-processing projection of event rate and collapse pressure on a reference grid reduces the grid dependency significantly. In order to evaluate the erosion-sensitive areas a statistical analysis of transient wall loads is utilised. Predicted erosion sensitive areas as well as temporal pressure and vapour volume evolution are in good agreement to the experimental data.
2005-04-01
the radiography gauging. In addition to the Statistical Energy Analysis (SEA) measurement a small exciter table (BK4810) and impedance head (BK 8000... Statistical Energy Analysis ; 7th Conf. on Vehicle System Dynamics, Identification and Anomalies (VSDIA2000), 6-8 Nov. 2000 Budapest, Proc. pp. 491-493... Energy Analysis (SEA) and Ultrasound Test. (UT) were concurrently applied. These methods collect accessory information on the objects under inspection
Statistical analysis of vehicle crashes in Mississippi based on crash data from 2010 to 2014.
DOT National Transportation Integrated Search
2017-08-15
Traffic crash data from 2010 to 2014 were collected by Mississippi Department of Transportation (MDOT) and extracted for the study. Three tasks were conducted in this study: (1) geographic distribution of crashes; (2) descriptive statistics of crash ...
DOT National Transportation Integrated Search
2013-07-01
This study is designed to assist the Ohio Department of Transportation (ODOT) in determining : whether transitioning from manual to state-of the-practice semi-automated pavement distress : data collection is feasible and recommended. Statistical and ...
DOT National Transportation Integrated Search
2014-12-01
The Bureau of Transportation Statistics (BTS) leads in the collection, analysis, and dissemination of transportation data. The Intermodal Passenger Connectivity Database : (ICPD) is an ongoing data collection that measures the degree of connectivity ...
Luo, Li; Zhu, Yun
2012-01-01
Abstract The genome-wide association studies (GWAS) designed for next-generation sequencing data involve testing association of genomic variants, including common, low frequency, and rare variants. The current strategies for association studies are well developed for identifying association of common variants with the common diseases, but may be ill-suited when large amounts of allelic heterogeneity are present in sequence data. Recently, group tests that analyze their collective frequency differences between cases and controls shift the current variant-by-variant analysis paradigm for GWAS of common variants to the collective test of multiple variants in the association analysis of rare variants. However, group tests ignore differences in genetic effects among SNPs at different genomic locations. As an alternative to group tests, we developed a novel genome-information content-based statistics for testing association of the entire allele frequency spectrum of genomic variation with the diseases. To evaluate the performance of the proposed statistics, we use large-scale simulations based on whole genome low coverage pilot data in the 1000 Genomes Project to calculate the type 1 error rates and power of seven alternative statistics: a genome-information content-based statistic, the generalized T2, collapsing method, multivariate and collapsing (CMC) method, individual χ2 test, weighted-sum statistic, and variable threshold statistic. Finally, we apply the seven statistics to published resequencing dataset from ANGPTL3, ANGPTL4, ANGPTL5, and ANGPTL6 genes in the Dallas Heart Study. We report that the genome-information content-based statistic has significantly improved type 1 error rates and higher power than the other six statistics in both simulated and empirical datasets. PMID:22651812
Luo, Li; Zhu, Yun; Xiong, Momiao
2012-06-01
The genome-wide association studies (GWAS) designed for next-generation sequencing data involve testing association of genomic variants, including common, low frequency, and rare variants. The current strategies for association studies are well developed for identifying association of common variants with the common diseases, but may be ill-suited when large amounts of allelic heterogeneity are present in sequence data. Recently, group tests that analyze their collective frequency differences between cases and controls shift the current variant-by-variant analysis paradigm for GWAS of common variants to the collective test of multiple variants in the association analysis of rare variants. However, group tests ignore differences in genetic effects among SNPs at different genomic locations. As an alternative to group tests, we developed a novel genome-information content-based statistics for testing association of the entire allele frequency spectrum of genomic variation with the diseases. To evaluate the performance of the proposed statistics, we use large-scale simulations based on whole genome low coverage pilot data in the 1000 Genomes Project to calculate the type 1 error rates and power of seven alternative statistics: a genome-information content-based statistic, the generalized T(2), collapsing method, multivariate and collapsing (CMC) method, individual χ(2) test, weighted-sum statistic, and variable threshold statistic. Finally, we apply the seven statistics to published resequencing dataset from ANGPTL3, ANGPTL4, ANGPTL5, and ANGPTL6 genes in the Dallas Heart Study. We report that the genome-information content-based statistic has significantly improved type 1 error rates and higher power than the other six statistics in both simulated and empirical datasets.
Making Decisions with Data: Are We Environmentally Friendly?
ERIC Educational Resources Information Center
English, Lyn; Watson, Jane
2016-01-01
Statistical literacy is a vital component of numeracy. Students need to learn to critically evaluate and interpret statistical information if they are to become informed citizens. This article examines a Year 5 unit of work that uses the data collection and analysis cycle within a sustainability context.
Spriestersbach, Albert; Röhrig, Bernd; du Prel, Jean-Baptist; Gerhold-Ay, Aslihan; Blettner, Maria
2009-09-01
Descriptive statistics are an essential part of biometric analysis and a prerequisite for the understanding of further statistical evaluations, including the drawing of inferences. When data are well presented, it is usually obvious whether the author has collected and evaluated them correctly and in keeping with accepted practice in the field. Statistical variables in medicine may be of either the metric (continuous, quantitative) or categorical (nominal, ordinal) type. Easily understandable examples are given. Basic techniques for the statistical description of collected data are presented and illustrated with examples. The goal of a scientific study must always be clearly defined. The definition of the target value or clinical endpoint determines the level of measurement of the variables in question. Nearly all variables, whatever their level of measurement, can be usefully presented graphically and numerically. The level of measurement determines what types of diagrams and statistical values are appropriate. There are also different ways of presenting combinations of two independent variables graphically and numerically. The description of collected data is indispensable. If the data are of good quality, valid and important conclusions can already be drawn when they are properly described. Furthermore, data description provides a basis for inferential statistics.
ERIC Educational Resources Information Center
Ling, Guo
2017-01-01
The author conducted sampling and statistical analysis of papers on education policy research collected by the China National Knowledge Infrastructure in the period from the years 2004--2013. Under the current state of education policy research in China, the number of papers correlates positively with the year; the papers are concentrated in…
STATISTICAL METHOD FOR DETECTION OF A TREND IN ATMOSPHERIC SULFATE
Daily atmospheric concentrations of sulfate collected in northeastern Pennsylvania are regressed against meteorological factors, ozone, and time in order to determine if a significant trend in sulfate can be detected. he data used in this analysis were collected during the Sulfat...
The Dynamics of Collective Violence: Dissecting Food Riots in Contemporary Argentina
ERIC Educational Resources Information Center
Auyero, Javier; Moran, Timothy Patrick
2007-01-01
This article combines a statistical analysis with qualitative research to investigate the dynamics of collective violence in one of its most recurrent forms--the food riot. Using an original dataset collected by the authors on 289 food riot episodes occurring in Argentina in December 2001, the article argues for the need to dissect the local,…
A note about high blood pressure in childhood
NASA Astrophysics Data System (ADS)
Teodoro, M. Filomena; Simão, Carla
2017-06-01
In medical, behavioral and social sciences it is usual to get a binary outcome. In the present work is collected information where some of the outcomes are binary variables (1='yes'/ 0='no'). In [14] a preliminary study about the caregivers perception of pediatric hypertension was introduced. An experimental questionnaire was designed to be answered by the caregivers of routine pediatric consultation attendees in the Santa Maria's hospital (HSM). The collected data was statistically analyzed, where a descriptive analysis and a predictive model were performed. Significant relations between some socio-demographic variables and the assessed knowledge were obtained. In [14] can be found a statistical data analysis using partial questionnaire's information. The present article completes the statistical approach estimating a model for relevant remaining questions of questionnaire by Generalized Linear Models (GLM). Exploring the binary outcome issue, we intend to extend this approach using Generalized Linear Mixed Models (GLMM), but the process is still ongoing.
Introductory Statistics and Fish Management.
ERIC Educational Resources Information Center
Jardine, Dick
2002-01-01
Describes how fisheries research and management data (available on a website) have been incorporated into an Introductory Statistics course. In addition to the motivation gained from seeing the practical relevance of the course, some students have participated in the data collection and analysis for the New Hampshire Fish and Game Department. (MM)
An Interactive Assessment Framework for Visual Engagement: Statistical Analysis of a TEDx Video
ERIC Educational Resources Information Center
Farhan, Muhammad; Aslam, Muhammad
2017-01-01
This study aims to assess the visual engagement of the video lectures. This analysis can be useful for the presenter and student to find out the overall visual attention of the videos. For this purpose, a new algorithm and data collection module are developed. Videos can be transformed into a dataset with the help of data collection module. The…
Primer of statistics in dental research: part I.
Shintani, Ayumi
2014-01-01
Statistics play essential roles in evidence-based dentistry (EBD) practice and research. It ranges widely from formulating scientific questions, designing studies, collecting and analyzing data to interpreting, reporting, and presenting study findings. Mastering statistical concepts appears to be an unreachable goal among many dental researchers in part due to statistical authorities' limitations of explaining statistical principles to health researchers without elaborating complex mathematical concepts. This series of 2 articles aim to introduce dental researchers to 9 essential topics in statistics to conduct EBD with intuitive examples. The part I of the series includes the first 5 topics (1) statistical graph, (2) how to deal with outliers, (3) p-value and confidence interval, (4) testing equivalence, and (5) multiplicity adjustment. Part II will follow to cover the remaining topics including (6) selecting the proper statistical tests, (7) repeated measures analysis, (8) epidemiological consideration for causal association, and (9) analysis of agreement. Copyright © 2014. Published by Elsevier Ltd.
An Independent Filter for Gene Set Testing Based on Spectral Enrichment.
Frost, H Robert; Li, Zhigang; Asselbergs, Folkert W; Moore, Jason H
2015-01-01
Gene set testing has become an indispensable tool for the analysis of high-dimensional genomic data. An important motivation for testing gene sets, rather than individual genomic variables, is to improve statistical power by reducing the number of tested hypotheses. Given the dramatic growth in common gene set collections, however, testing is often performed with nearly as many gene sets as underlying genomic variables. To address the challenge to statistical power posed by large gene set collections, we have developed spectral gene set filtering (SGSF), a novel technique for independent filtering of gene set collections prior to gene set testing. The SGSF method uses as a filter statistic the p-value measuring the statistical significance of the association between each gene set and the sample principal components (PCs), taking into account the significance of the associated eigenvalues. Because this filter statistic is independent of standard gene set test statistics under the null hypothesis but dependent under the alternative, the proportion of enriched gene sets is increased without impacting the type I error rate. As shown using simulated and real gene expression data, the SGSF algorithm accurately filters gene sets unrelated to the experimental outcome resulting in significantly increased gene set testing power.
Barnes, Stephen; Benton, H. Paul; Casazza, Krista; Cooper, Sara; Cui, Xiangqin; Du, Xiuxia; Engler, Jeffrey; Kabarowski, Janusz H.; Li, Shuzhao; Pathmasiri, Wimal; Prasain, Jeevan K.; Renfrow, Matthew B.; Tiwari, Hemant K.
2017-01-01
Metabolomics, a systems biology discipline representing analysis of known and unknown pathways of metabolism, has grown tremendously over the past 20 years. Because of its comprehensive nature, metabolomics requires careful consideration of the question(s) being asked, the scale needed to answer the question(s), collection and storage of the sample specimens, methods for extraction of the metabolites from biological matrices, the analytical method(s) to be employed and the quality control of the analyses, how collected data are correlated, the statistical methods to determine metabolites undergoing significant change, putative identification of metabolites, and the use of stable isotopes to aid in verifying metabolite identity and establishing pathway connections and fluxes. This second part of a comprehensive description of the methods of metabolomics focuses on data analysis, emerging methods in metabolomics and the future of this discipline. PMID:28239968
A statistical analysis of IUE spectra of dwarf novae and nova-like stars
NASA Technical Reports Server (NTRS)
Ladous, Constanze
1990-01-01
First results of a statistical analysis of the IUE International Ultraviolet Explorer archive on dwarf novae and nova like stars are presented. The archive contains approximately 2000 low resolution spectra of somewhat over 100 dwarf novae and nova like stars. Many of these were looked at individually, but so far the collective information content of this set of data has not been explored. The first results of work are reported.
Building Energy Monitoring and Analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hong, Tianzhen; Feng, Wei; Lu, Alison
This project aimed to develop a standard methodology for building energy data definition, collection, presentation, and analysis; apply the developed methods to a standardized energy monitoring platform, including hardware and software, to collect and analyze building energy use data; and compile offline statistical data and online real-time data in both countries for fully understanding the current status of building energy use. This helps decode the driving forces behind the discrepancy of building energy use between the two countries; identify gaps and deficiencies of current building energy monitoring, data collection, and analysis; and create knowledge and tools to collect and analyzemore » good building energy data to provide valuable and actionable information for key stakeholders.« less
1988-12-09
Measurement of Second Order Statistics .... .............. .54 5.4 Measurement of Triple Products ...... ................. .58 5.6 Uncertainty Analysis...deterministic fluctuations, u/ 2 , were 25 times larger than the mean fluctuations, u, there were no significant variations in the mean statistical ...input signals, the three velocity components are cal- culated, Awn in ,i-;dual phase ensembles are collected for the appropriate statistical 3
75 FR 81999 - Notice of Submission for OMB Review
Federal Register 2010, 2011, 2012, 2013, 2014
2010-12-29
... comments which: (1) Evaluate whether the proposed collection of information is necessary for the proper...) Evaluate the accuracy of the agency's estimate of the burden of the proposed collection of information... study will use descriptive statistics and regression analysis to study how student outcomes and school...
77 FR 17460 - Proposed Collection; Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2012-03-26
..., Associated Form, and OMB Control Number: The 2012 Post- Election Survey of State and Local Election Officials; OMB Control Number 0704-0125. Needs and Uses: The information collection requirement is necessary to.... 1973ff]). UOCAVA requires a statistical analysis report to the President and Congress on the...
78 FR 36160 - Notice of Intent To Request New Information Collection
Federal Register 2010, 2011, 2012, 2013, 2014
2013-06-17
... Economic Research Service's intention to request approval for a new information collection for a Survey on... and confidential. Survey responses will be used for statistical analysis and to produce research... DEPARTMENT OF AGRICULTURE Economic Research Service Notice of Intent To Request New Information...
Introduction to bioinformatics.
Can, Tolga
2014-01-01
Bioinformatics is an interdisciplinary field mainly involving molecular biology and genetics, computer science, mathematics, and statistics. Data intensive, large-scale biological problems are addressed from a computational point of view. The most common problems are modeling biological processes at the molecular level and making inferences from collected data. A bioinformatics solution usually involves the following steps: Collect statistics from biological data. Build a computational model. Solve a computational modeling problem. Test and evaluate a computational algorithm. This chapter gives a brief introduction to bioinformatics by first providing an introduction to biological terminology and then discussing some classical bioinformatics problems organized by the types of data sources. Sequence analysis is the analysis of DNA and protein sequences for clues regarding function and includes subproblems such as identification of homologs, multiple sequence alignment, searching sequence patterns, and evolutionary analyses. Protein structures are three-dimensional data and the associated problems are structure prediction (secondary and tertiary), analysis of protein structures for clues regarding function, and structural alignment. Gene expression data is usually represented as matrices and analysis of microarray data mostly involves statistics analysis, classification, and clustering approaches. Biological networks such as gene regulatory networks, metabolic pathways, and protein-protein interaction networks are usually modeled as graphs and graph theoretic approaches are used to solve associated problems such as construction and analysis of large-scale networks.
Tools for Basic Statistical Analysis
NASA Technical Reports Server (NTRS)
Luz, Paul L.
2005-01-01
Statistical Analysis Toolset is a collection of eight Microsoft Excel spreadsheet programs, each of which performs calculations pertaining to an aspect of statistical analysis. These programs present input and output data in user-friendly, menu-driven formats, with automatic execution. The following types of calculations are performed: Descriptive statistics are computed for a set of data x(i) (i = 1, 2, 3 . . . ) entered by the user. Normal Distribution Estimates will calculate the statistical value that corresponds to cumulative probability values, given a sample mean and standard deviation of the normal distribution. Normal Distribution from two Data Points will extend and generate a cumulative normal distribution for the user, given two data points and their associated probability values. Two programs perform two-way analysis of variance (ANOVA) with no replication or generalized ANOVA for two factors with four levels and three repetitions. Linear Regression-ANOVA will curvefit data to the linear equation y=f(x) and will do an ANOVA to check its significance.
Advances in Statistical Methods for Substance Abuse Prevention Research
MacKinnon, David P.; Lockwood, Chondra M.
2010-01-01
The paper describes advances in statistical methods for prevention research with a particular focus on substance abuse prevention. Standard analysis methods are extended to the typical research designs and characteristics of the data collected in prevention research. Prevention research often includes longitudinal measurement, clustering of data in units such as schools or clinics, missing data, and categorical as well as continuous outcome variables. Statistical methods to handle these features of prevention data are outlined. Developments in mediation, moderation, and implementation analysis allow for the extraction of more detailed information from a prevention study. Advancements in the interpretation of prevention research results include more widespread calculation of effect size and statistical power, the use of confidence intervals as well as hypothesis testing, detailed causal analysis of research findings, and meta-analysis. The increased availability of statistical software has contributed greatly to the use of new methods in prevention research. It is likely that the Internet will continue to stimulate the development and application of new methods. PMID:12940467
The Physical and the Virtual: The Relationship between Library as Place and Electronic Collections
ERIC Educational Resources Information Center
Gerke, Jennifer; Maness, Jack M.
2010-01-01
A statistical analysis of responses to a LibQUAL+™ survey at the University of Colorado at Boulder (UCB) was conducted to investigate factors related to patrons' satisfaction with electronic collections. It was found that a respondent's discipline was not related to his or her satisfaction with the Libraries' electronic collection, nor was the…
Measuring the Gas Constant "R": Propagation of Uncertainty and Statistics
ERIC Educational Resources Information Center
Olsen, Robert J.; Sattar, Simeen
2013-01-01
Determining the gas constant "R" by measuring the properties of hydrogen gas collected in a gas buret is well suited for comparing two approaches to uncertainty analysis using a single data set. The brevity of the experiment permits multiple determinations, allowing for statistical evaluation of the standard uncertainty u[subscript…
Statistical analysis of the calibration procedure for personnel radiation measurement instruments
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bush, W.J.; Bengston, S.J.; Kalbeitzer, F.L.
1980-11-01
Thermoluminescent analyzer (TLA) calibration procedures were used to estimate personnel radiation exposure levels at the Idaho National Engineering Laboratory (INEL). A statistical analysis is presented herein based on data collected over a six month period in 1979 on four TLA's located in the Department of Energy (DOE) Radiological and Environmental Sciences Laboratory at the INEL. The data were collected according to the day-to-day procedure in effect at that time. Both gamma and beta radiation models are developed. Observed TLA readings of thermoluminescent dosimeters are correlated with known radiation levels. This correlation is then used to predict unknown radiation doses frommore » future analyzer readings of personnel thermoluminescent dosimeters. The statistical techniques applied in this analysis include weighted linear regression, estimation of systematic and random error variances, prediction interval estimation using Scheffe's theory of calibration, the estimation of the ratio of the means of two normal bivariate distributed random variables and their corresponding confidence limits according to Kendall and Stuart, tests of normality, experimental design, a comparison between instruments, and quality control.« less
Statistical analysis of CCSN/SS7 traffic data from working CCS subnetworks
NASA Astrophysics Data System (ADS)
Duffy, Diane E.; McIntosh, Allen A.; Rosenstein, Mark; Willinger, Walter
1994-04-01
In this paper, we report on an ongoing statistical analysis of actual CCSN traffic data. The data consist of approximately 170 million signaling messages collected from a variety of different working CCS subnetworks. The key findings from our analysis concern: (1) the characteristics of both the telephone call arrival process and the signaling message arrival process; (2) the tail behavior of the call holding time distribution; and (3) the observed performance of the CCSN with respect to a variety of performance and reliability measurements.
Collected Notes on the Workshop for Pattern Discovery in Large Databases
NASA Technical Reports Server (NTRS)
Buntine, Wray (Editor); Delalto, Martha (Editor)
1991-01-01
These collected notes are a record of material presented at the Workshop. The core data analysis is addressed that have traditionally required statistical or pattern recognition techniques. Some of the core tasks include classification, discrimination, clustering, supervised and unsupervised learning, discovery and diagnosis, i.e., general pattern discovery.
Multivariate analysis: greater insights into complex systems
USDA-ARS?s Scientific Manuscript database
Many agronomic researchers measure and collect multiple response variables in an effort to understand the more complex nature of the system being studied. Multivariate (MV) statistical methods encompass the simultaneous analysis of all random variables (RV) measured on each experimental or sampling ...
Assessment of environmental impacts part one. Intervention analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hipel, Keith William; Lettenmaier, Dennis P.; McLeod, A. Ian
The use of intervention analysis as a statistical method of gauging the effects of environmental changes is discussed. The Box-Jenkins model, serves as the basis for the intervention analysis methodology. Environmental studies of the Aswan Dam, the South Saskatchewan River, and a forest fire near the Pipers Hole River, Canada, are included as case studies in which intervention analysis was employed. Methods of data collection for intervention analysis are found to have a significant impact on model reliability; effective data collection processes for the Box-Jenkins model are provided. (15 graphs, 27 references, 2 tables)
Illustrating the practice of statistics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hamada, Christina A; Hamada, Michael S
2009-01-01
The practice of statistics involves analyzing data and planning data collection schemes to answer scientific questions. Issues often arise with the data that must be dealt with and can lead to new procedures. In analyzing data, these issues can sometimes be addressed through the statistical models that are developed. Simulation can also be helpful in evaluating a new procedure. Moreover, simulation coupled with optimization can be used to plan a data collection scheme. The practice of statistics as just described is much more than just using a statistical package. In analyzing the data, it involves understanding the scientific problem andmore » incorporating the scientist's knowledge. In modeling the data, it involves understanding how the data were collected and accounting for limitations of the data where possible. Moreover, the modeling is likely to be iterative by considering a series of models and evaluating the fit of these models. Designing a data collection scheme involves understanding the scientist's goal and staying within hislher budget in terms of time and the available resources. Consequently, a practicing statistician is faced with such tasks and requires skills and tools to do them quickly. We have written this article for students to provide a glimpse of the practice of statistics. To illustrate the practice of statistics, we consider a problem motivated by some precipitation data that our relative, Masaru Hamada, collected some years ago. We describe his rain gauge observational study in Section 2. We describe modeling and an initial analysis of the precipitation data in Section 3. In Section 4, we consider alternative analyses that address potential issues with the precipitation data. In Section 5, we consider the impact of incorporating additional infonnation. We design a data collection scheme to illustrate the use of simulation and optimization in Section 6. We conclude this article in Section 7 with a discussion.« less
Illinois' Forests, 2005: Statistics, Methods, and Quality Assurance
Susan J. Crocker; Charles J. Barnett; Mark A. Hatfield
2013-01-01
The first full annual inventory of Illinois' forests was completed in 2005. This report contains 1) descriptive information on methods, statistics, and quality assurance of data collection, 2) a glossary of terms, 3) tables that summarize quality assurance, and 4) a core set of tabular estimates for a variety of forest resources. A detailed analysis of inventory...
Statistical Analysis of CFD Solutions From the Fifth AIAA Drag Prediction Workshop
NASA Technical Reports Server (NTRS)
Morrison, Joseph H.
2013-01-01
A graphical framework is used for statistical analysis of the results from an extensive N-version test of a collection of Reynolds-averaged Navier-Stokes computational fluid dynamics codes. The solutions were obtained by code developers and users from North America, Europe, Asia, and South America using a common grid sequence and multiple turbulence models for the June 2012 fifth Drag Prediction Workshop sponsored by the AIAA Applied Aerodynamics Technical Committee. The aerodynamic configuration for this workshop was the Common Research Model subsonic transport wing-body previously used for the 4th Drag Prediction Workshop. This work continues the statistical analysis begun in the earlier workshops and compares the results from the grid convergence study of the most recent workshop with previous workshops.
Development of new on-line statistical program for the Korean Society for Radiation Oncology
Song, Si Yeol; Ahn, Seung Do; Chung, Weon Kuu; Choi, Eun Kyung; Cho, Kwan Ho
2015-01-01
Purpose To develop new on-line statistical program for the Korean Society for Radiation Oncology (KOSRO) to collect and extract medical data in radiation oncology more efficiently. Materials and Methods The statistical program is a web-based program. The directory was placed in a sub-folder of the homepage of KOSRO and its web address is http://www.kosro.or.kr/asda. The operating systems server is Linux and the webserver is the Apache HTTP server. For database (DB) server, MySQL is adopted and dedicated scripting language is the PHP. Each ID and password are controlled independently and all screen pages for data input or analysis are made to be friendly to users. Scroll-down menu is actively used for the convenience of user and the consistence of data analysis. Results Year of data is one of top categories and main topics include human resource, equipment, clinical statistics, specialized treatment and research achievement. Each topic or category has several subcategorized topics. Real-time on-line report of analysis is produced immediately after entering each data and the administrator is able to monitor status of data input of each hospital. Backup of data as spread sheets can be accessed by the administrator and be used for academic works by any members of the KOSRO. Conclusion The new on-line statistical program was developed to collect data from nationwide departments of radiation oncology. Intuitive screen and consistent input structure are expected to promote entering data of member hospitals and annual statistics should be a cornerstone of advance in radiation oncology. PMID:26157684
Development of new on-line statistical program for the Korean Society for Radiation Oncology.
Song, Si Yeol; Ahn, Seung Do; Chung, Weon Kuu; Shin, Kyung Hwan; Choi, Eun Kyung; Cho, Kwan Ho
2015-06-01
To develop new on-line statistical program for the Korean Society for Radiation Oncology (KOSRO) to collect and extract medical data in radiation oncology more efficiently. The statistical program is a web-based program. The directory was placed in a sub-folder of the homepage of KOSRO and its web address is http://www.kosro.or.kr/asda. The operating systems server is Linux and the webserver is the Apache HTTP server. For database (DB) server, MySQL is adopted and dedicated scripting language is the PHP. Each ID and password are controlled independently and all screen pages for data input or analysis are made to be friendly to users. Scroll-down menu is actively used for the convenience of user and the consistence of data analysis. Year of data is one of top categories and main topics include human resource, equipment, clinical statistics, specialized treatment and research achievement. Each topic or category has several subcategorized topics. Real-time on-line report of analysis is produced immediately after entering each data and the administrator is able to monitor status of data input of each hospital. Backup of data as spread sheets can be accessed by the administrator and be used for academic works by any members of the KOSRO. The new on-line statistical program was developed to collect data from nationwide departments of radiation oncology. Intuitive screen and consistent input structure are expected to promote entering data of member hospitals and annual statistics should be a cornerstone of advance in radiation oncology.
The outcome of the seminal fluid parameters collected via coitus interruptus versus masturbation.
Bahyah, M Kamarul; Murad, Z Ahmad; Ghazali, I; Roszaman, R; Noraziana, A W; Mokhtar, A; Omar, M H
2010-03-01
A one year study was carried out to determine the outcome of the seminal fluid parameters collected via masturbation and coitus interruptus in 151 patients who were undergoing intrauterine insemination (IUI) and patients who came for seminal analysis. There were no statistically significant differences in terms of volume, concentration, progressive motility and normal morphology from specimens collected via coitus interruptus compared to specimens collected via masturbation. Pregnancy outcomes were also comparable.
Moving beyond Assumptions: The Use of Virtual Reference Data in an Academic Library
ERIC Educational Resources Information Center
Nolen, David S.; Powers, Amanda Clay; Zhang, Li; Xu, Yue; Cannady, Rachel E.; Li, Judy
2012-01-01
The Mississippi State University Libraries' Virtual Reference Service collected statistics about virtual reference usage. Analysis of the data collected by an entry survey from chat and e-mail transactions provided librarians with concrete information about what patron groups were the highest and lowest users of virtual reference services. These…
Macfarlane, Sarah B.
2005-01-01
Efforts to strengthen health information systems in low- and middle-income countries should include forging links with systems in other social and economic sectors. Governments are seeking comprehensive socioeconomic data on the basis of which to implement strategies for poverty reduction and to monitor achievement of the Millennium Development Goals. The health sector is looking to take action on the social factors that determine health outcomes. But there are duplications and inconsistencies between sectors in the collection, reporting, storage and analysis of socioeconomic data. National offices of statistics give higher priority to collection and analysis of economic than to social statistics. The Report of the Commission for Africa has estimated that an additional US$ 60 million a year is needed to improve systems to collect and analyse statistics in Africa. Some donors recognize that such systems have been weakened by numerous international demands for indicators, and have pledged support for national initiatives to strengthen statistical systems, as well as sectoral information systems such as those in health and education. Many governments are working to coordinate information systems to monitor and evaluate poverty reduction strategies. There is therefore an opportunity for the health sector to collaborate with other sectors to lever international resources to rationalize definition and measurement of indicators common to several sectors; streamline the content, frequency and timing of household surveys; and harmonize national and subnational databases that store socioeconomic data. Without long-term commitment to improve training and build career structures for statisticians and information technicians working in the health and other sectors, improvements in information and statistical systems cannot be sustained. PMID:16184278
Barnes, Stephen; Benton, H Paul; Casazza, Krista; Cooper, Sara J; Cui, Xiangqin; Du, Xiuxia; Engler, Jeffrey; Kabarowski, Janusz H; Li, Shuzhao; Pathmasiri, Wimal; Prasain, Jeevan K; Renfrow, Matthew B; Tiwari, Hemant K
2016-08-01
Metabolomics, a systems biology discipline representing analysis of known and unknown pathways of metabolism, has grown tremendously over the past 20 years. Because of its comprehensive nature, metabolomics requires careful consideration of the question(s) being asked, the scale needed to answer the question(s), collection and storage of the sample specimens, methods for extraction of the metabolites from biological matrices, the analytical method(s) to be employed and the quality control of the analyses, how collected data are correlated, the statistical methods to determine metabolites undergoing significant change, putative identification of metabolites and the use of stable isotopes to aid in verifying metabolite identity and establishing pathway connections and fluxes. This second part of a comprehensive description of the methods of metabolomics focuses on data analysis, emerging methods in metabolomics and the future of this discipline. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
Economic and statistical analysis of time limitations for spotting fluids and fishing operations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Keller, P.S.; Brinkmann, P.E.; Taneja, P.K.
1984-05-01
This paper reviews the statistics of ''Spotting Fluids'' to free stuck drill pipe as well as the economics and statistics of drill string fishing operations. Data were taken from Mobil Oil Exploration and Producing Southeast Inc.'s (MOEPSI) records from 1970-1981. Only those events which occur after a drill string becomes stuck are discussed. The data collected were categorized as Directional Wells and Straight Wells. Bar diagrams are presented to show the Success Ratio vs. Soaking Time for each of the two categories. An analysis was made to identify the elapsed time limit to place the spotting fluid for maximum probabilitymore » of success. Also determined was the statistical minimum soaking time and the maximum soaking time. For determining the time limit for fishing operations, the following criteria were used: 1. The Risked ''Economic Breakeven Analysis'' concept was developed based on the work of Harrison. 2. Statistical Probability of Success based on MOEPSI's records from 1970-1981.« less
Anderson, Craig S; Woodward, Mark; Arima, Hisatomi; Chen, Xiaoying; Lindley, Richard I; Wang, Xia; Chalmers, John
2015-12-01
The ENhanced Control of Hypertension And Thrombolysis strokE stuDy trial is a 2 × 2 quasi-factorial active-comparison, prospective, randomized, open, blinded endpoint clinical trial that is evaluating in thrombolysis-eligible acute ischemic stroke patients whether: (1) low-dose (0·6 mg/kg body weight) intravenous alteplase has noninferior efficacy and lower risk of symptomatic intracerebral hemorrhage compared with standard-dose (0·9 mg/kg body weight) intravenous alteplase; and (2) early intensive blood pressure lowering (systolic target 130-140 mmHg) has superior efficacy and lower risk of any intracerebral hemorrhage compared with guideline-recommended blood pressure control (systolic target <180 mmHg). To outline in detail the predetermined statistical analysis plan for the 'alteplase dose arm' of the study. All data collected by participating researchers will be reviewed and formally assessed. Information pertaining to the baseline characteristics of patients, their process of care, and the delivery of treatments will be classified, and for each item, appropriate descriptive statistical analyses are planned with appropriate comparisons made between randomized groups. For the trial outcomes, the most appropriate statistical comparisons to be made between groups are planned and described. A statistical analysis plan was developed for the results of the alteplase dose arm of the study that is transparent, available to the public, verifiable, and predetermined before completion of data collection. We have developed a predetermined statistical analysis plan for the ENhanced Control of Hypertension And Thrombolysis strokE stuDy alteplase dose arm which is to be followed to avoid analysis bias arising from prior knowledge of the study findings. © 2015 The Authors. International Journal of Stroke published by John Wiley & Sons Ltd on behalf of World Stroke Organization.
Brown, Geoffrey W.; Sandstrom, Mary M.; Preston, Daniel N.; ...
2014-11-17
In this study, the Integrated Data Collection Analysis (IDCA) program has conducted a proficiency test for small-scale safety and thermal (SSST) testing of homemade explosives (HMEs). Described here are statistical analyses of the results from this test for impact, friction, electrostatic discharge, and differential scanning calorimetry analysis of the RDX Class 5 Type II standard. The material was tested as a well-characterized standard several times during the proficiency test to assess differences among participants and the range of results that may arise for well-behaved explosive materials.
The statistical analysis of global climate change studies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hardin, J.W.
1992-01-01
The focus of this work is to contribute to the enhancement of the relationship between climatologists and statisticians. The analysis of global change data has been underway for many years by atmospheric scientists. Much of this analysis includes a heavy reliance on statistics and statistical inference. Some specific climatological analyses are presented and the dependence on statistics is documented before the analysis is undertaken. The first problem presented involves the fluctuation-dissipation theorem and its application to global climate models. This problem has a sound theoretical niche in the literature of both climate modeling and physics, but a statistical analysis inmore » which the data is obtained from the model to show graphically the relationship has not been undertaken. It is under this motivation that the author presents this problem. A second problem concerning the standard errors in estimating global temperatures is purely statistical in nature although very little materials exists for sampling on such a frame. This problem not only has climatological and statistical ramifications, but political ones as well. It is planned to use these results in a further analysis of global warming using actual data collected on the earth. In order to simplify the analysis of these problems, the development of a computer program, MISHA, is presented. This interactive program contains many of the routines, functions, graphics, and map projections needed by the climatologist in order to effectively enter the arena of data visualization.« less
Summary of Hydrologic Conditions in Georgia, 2008
Knaak, Andrew E.; Joiner, John K.; Peck, Michael F.
2009-01-01
The United States Geological Survey (USGS) Georgia Water Science Center (WSC) maintains a long-term hydrologic monitoring network of more than 290 real-time streamgages, more than 170 groundwater wells, and 10 lake and reservoir monitoring stations. One of the many benefits of data collected from this monitoring network is that analysis of the data provides an overview of the hydrologic conditions of rivers, creeks, reservoirs, and aquifers in Georgia. Hydrologic conditions are determined by statistical analysis of data collected during the current water year (WY) and comparison of the results to historical data collected at long-term stations. During the drought that persisted through 2008, the USGS succeeded in verifying and documenting numerous historic low-flow statistics at many streamgages and current water levels in aquifers, lakes, and reservoirs in Georgia. Streamflow data from the 2008 WY indicate that this drought is one of the most severe on record when compared to drought periods of 1950-1957, 1985-1989, and 1999-2002.
Effects of temporal variability in ground data collection on classification accuracy
Hoch, G.A.; Cully, J.F.
1999-01-01
This research tested whether the timing of ground data collection can significantly impact the accuracy of land cover classification. Ft. Riley Military Reservation, Kansas, USA was used to test this hypothesis. The U.S. Army's Land Condition Trend Analysis (LCTA) data annually collected at military bases was used to ground truth disturbance patterns. Ground data collected over an entire growing season and data collected one year after the imagery had a kappa statistic of 0.33. When using ground data from only within two weeks of image acquisition the kappa statistic improved to 0.55. Potential sources of this discrepancy are identified. These data demonstrate that there can be significant amounts of land cover change within a narrow time window on military reservations. To accurately conduct land cover classification at military reservations, ground data need to be collected in as narrow a window of time as possible and be closely synchronized with the date of the satellite imagery.
Statistics: The Shape of the Data. Used Numbers: Real Data in the Classroom. Grades 4-6.
ERIC Educational Resources Information Center
Russell, Susan Jo; Corwin, Rebecca B.
A unit of study that introduces collecting, representing, describing, and interpreting data is presented. Suitable for students in grades 4 through 6, it provides a foundation for further work in statistics and data analysis. The investigations may extend from one to four class sessions and are grouped into three parts: "Introduction to Data…
Effects of Cognitive Load on Trust
2013-10-01
that may be affected by load Build a parsing tool to extract relevant features Statistical analysis of results (by load components) Achieved...for a business application. Participants assessed potential job candidates and reviewed the applicants’ virtual resume which included standard...substantially different from each other that would make any confounding problems or other issues. Some statistics of the Australian data collection are
Vahedi, Shahrum; Farrokhi, Farahman; Gahramani, Farahnaz; Issazadegan, Ali
2012-01-01
Approximately 66-80%of graduate students experience statistics anxiety and some researchers propose that many students identify statistics courses as the most anxiety-inducing courses in their academic curriculums. As such, it is likely that statistics anxiety is, in part, responsible for many students delaying enrollment in these courses for as long as possible. This paper proposes a canonical model by treating academic procrastination (AP), learning strategies (LS) as predictor variables and statistics anxiety (SA) as explained variables. A questionnaire survey was used for data collection and 246-college female student participated in this study. To examine the mutually independent relations between procrastination, learning strategies and statistics anxiety variables, a canonical correlation analysis was computed. Findings show that two canonical functions were statistically significant. The set of variables (metacognitive self-regulation, source management, preparing homework, preparing for test and preparing term papers) helped predict changes of statistics anxiety with respect to fearful behavior, Attitude towards math and class, Performance, but not Anxiety. These findings could be used in educational and psychological interventions in the context of statistics anxiety reduction.
Muhammad, Said; Tahir Shah, M; Khan, Sardar
2010-10-01
The present study was conducted in Kohistan region, where mafic and ultramafic rocks (Kohistan island arc and Indus suture zone) and metasedimentary rocks (Indian plate) are exposed. Water samples were collected from the springs, streams and Indus river and analyzed for physical parameters, anions, cations and arsenic (As(3+), As(5+) and arsenic total). The water quality in Kohistan region was evaluated by comparing the physio-chemical parameters with permissible limits set by Pakistan environmental protection agency and world health organization. Most of the studied parameters were found within their respective permissible limits. However in some samples, the iron and arsenic concentrations exceeded their permissible limits. For health risk assessment of arsenic, the average daily dose, hazards quotient (HQ) and cancer risk were calculated by using statistical formulas. The values of HQ were found >1 in the samples collected from Jabba, Dubair, while HQ values were <1 in rest of the samples. This level of contamination should have low chronic risk and medium cancer risk when compared with US EPA guidelines. Furthermore, the inter-dependence of physio-chemical parameters and pollution load was also calculated by using multivariate statistical techniques like one-way ANOVA, correlation analysis, regression analysis, cluster analysis and principle component analysis. Copyright © 2010 Elsevier Ltd. All rights reserved.
Billot, Laurent; Lindley, Richard I; Harvey, Lisa A; Maulik, Pallab K; Hackett, Maree L; Murthy, Gudlavalleti Vs; Anderson, Craig S; Shamanna, Bindiganavale R; Jan, Stephen; Walker, Marion; Forster, Anne; Langhorne, Peter; Verma, Shweta J; Felix, Cynthia; Alim, Mohammed; Gandhi, Dorcas Bc; Pandian, Jeyaraj Durai
2017-02-01
Background In low- and middle-income countries, few patients receive organized rehabilitation after stroke, yet the burden of chronic diseases such as stroke is increasing in these countries. Affordable models of effective rehabilitation could have a major impact. The ATTEND trial is evaluating a family-led caregiver delivered rehabilitation program after stroke. Objective To publish the detailed statistical analysis plan for the ATTEND trial prior to trial unblinding. Methods Based upon the published registration and protocol, the blinded steering committee and management team, led by the trial statistician, have developed a statistical analysis plan. The plan has been informed by the chosen outcome measures, the data collection forms and knowledge of key baseline data. Results The resulting statistical analysis plan is consistent with best practice and will allow open and transparent reporting. Conclusions Publication of the trial statistical analysis plan reduces potential bias in trial reporting, and clearly outlines pre-specified analyses. Clinical Trial Registrations India CTRI/2013/04/003557; Australian New Zealand Clinical Trials Registry ACTRN1261000078752; Universal Trial Number U1111-1138-6707.
Kanda, Junya
2016-01-01
The Transplant Registry Unified Management Program (TRUMP) made it possible for members of the Japan Society for Hematopoietic Cell Transplantation (JSHCT) to analyze large sets of national registry data on autologous and allogeneic hematopoietic stem cell transplantation. However, as the processes used to collect transplantation information are complex and differed over time, the background of these processes should be understood when using TRUMP data. Previously, information on the HLA locus of patients and donors had been collected using a questionnaire-based free-description method, resulting in some input errors. To correct minor but significant errors and provide accurate HLA matching data, the use of a Stata or EZR/R script offered by the JSHCT is strongly recommended when analyzing HLA data in the TRUMP dataset. The HLA mismatch direction, mismatch counting method, and different impacts of HLA mismatches by stem cell source are other important factors in the analysis of HLA data. Additionally, researchers should understand the statistical analyses specific for hematopoietic stem cell transplantation, such as competing risk, landmark analysis, and time-dependent analysis, to correctly analyze transplant data. The data center of the JSHCT can be contacted if statistical assistance is required.
NASA Astrophysics Data System (ADS)
Roy, P. K.; Pal, S.; Banerjee, G.; Biswas Roy, M.; Ray, D.; Majumder, A.
2014-12-01
River is considered as one of the main sources of freshwater all over the world. Hence analysis and maintenance of this water resource is globally considered a matter of major concern. This paper deals with the assessment of surface water quality of the Ichamati river using multivariate statistical techniques. Eight distinct surface water quality observation stations were located and samples were collected. For the samples collected statistical techniques were applied to the physico-chemical parameters and depth of siltation. In this paper cluster analysis is done to determine the relations between surface water quality and siltation depth of river Ichamati. Multiple regressions and mathematical equation modeling have been done to characterize surface water quality of Ichamati river on the basis of physico-chemical parameters. It was found that surface water quality of the downstream river was different from the water quality of the upstream. The analysis of the water quality parameters of the Ichamati river clearly indicate high pollution load on the river water which can be accounted to agricultural discharge, tidal effect and soil erosion. The results further reveal that with the increase in depth of siltation, water quality degraded.
NASA Technical Reports Server (NTRS)
Hyde, G.
1976-01-01
The 13/18 GHz COMSAT Propagation Experiment (CPE) was performed to measure attenuation caused by hydrometeors along slant paths from transmitting terminals on the ground to the ATS-6 satellite. The effectiveness of site diversity in overcoming this impairment was also studied. Problems encountered in assembling a valid data base of rain induced attenuation data for statistical analysis are considered. The procedures used to obtain the various statistics are then outlined. The graphs and tables of statistical data for the 15 dual frequency (13 and 18 GHz) site diversity locations are discussed. Cumulative rain rate statistics for the Fayetteville and Boston sites based on point rainfall data collected are presented along with extrapolations of the attenuation and point rainfall data.
Guevara-García, José Antonio; Montiel-Corona, Virginia
2012-03-01
A statistical analysis of a used battery collection campaign in the state of Tlaxcala, Mexico, is presented. This included a study of the metal composition of spent batteries from formal and informal markets, and a critical discussion about the management of spent batteries in Mexico with respect to legislation. A six-month collection campaign was statistically analyzed: 77% of the battery types were "AA" and 30% of the batteries were from the informal market. A substantial percentage (36%) of batteries had residual voltage in the range 1.2-1.4 V, and 70% had more than 1.0 V; this may reflect underutilization. Metal content analysis and recovery experiments were performed with the five formal and four more frequent informal trademarks. The analysis of Hg, Cd and Pb showed there is no significant difference in content between formal and informal commercialized batteries. All of the analyzed trademarks were under the permissible limit levels of the proposed Mexican Official Norm (NOM) NMX-AA-104-SCFI-2006 and would be classified as not dangerous residues (can be thrown to the domestic rubbish); however, compared with the EU directive 2006/66/EC, 8 out of 9 of the selected battery trademarks would be rejected, since the Mexican Norm content limit is 20, 7.5 and 5 fold higher in Hg, Cd and Pb, respectively, than the EU directive. These results outline the necessity for better regulatory criteria in the proposed Mexican NOM in order to minimize the impact on human health and the environment of this type of residues. Copyright © 2010 Elsevier Ltd. All rights reserved.
Preliminary Survey of Icing Conditions Measured During Routine Transcontinental Airline Operation
NASA Technical Reports Server (NTRS)
Perkins, Porter J.
1952-01-01
Icing data collected on routine operations by four DC-4-type aircraft equipped with NACA pressure-type icing-rate meters are presented as preliminary information obtained from a statistical icing data program sponsored by the NACA with the cooperation of many airline companies and the United States Air Force. The program is continuing on a much greater scale to provide large quantities of data from many air routes in the United States and overseas. Areas not covered by established air routes are also being included in the survey. The four aircraft which collected the data presented in this report were operated by United Air Lines over a transcontinental route from January through May, 1951. An analysis of the pressure-type icing-rate meter was satisfactory for collecting statistical data during routine operations. Data obtained on routine flight icing encounters from.these four instrumented aircraft, although insufficient for a conclusive statistical analysis, provide a greater quantity and considerably more realistic information than that obtained from random research flights. A summary of statistical data will be published when the information obtained daring the 1951-52 icing season and that to be obtained during the 1952-53 season can be analyzed and assembled. The 1951-52 data already analyzed indicate that the quantity, quality, and range of icing information being provided by this expanded program should afford a sound basis for ice-protection-system design by defining the important meteorological parameters of the icing cloud.
Multivariate analysis: A statistical approach for computations
NASA Astrophysics Data System (ADS)
Michu, Sachin; Kaushik, Vandana
2014-10-01
Multivariate analysis is a type of multivariate statistical approach commonly used in, automotive diagnosis, education evaluating clusters in finance etc and more recently in the health-related professions. The objective of the paper is to provide a detailed exploratory discussion about factor analysis (FA) in image retrieval method and correlation analysis (CA) of network traffic. Image retrieval methods aim to retrieve relevant images from a collected database, based on their content. The problem is made more difficult due to the high dimension of the variable space in which the images are represented. Multivariate correlation analysis proposes an anomaly detection and analysis method based on the correlation coefficient matrix. Anomaly behaviors in the network include the various attacks on the network like DDOs attacks and network scanning.
Matysiak, W; Królikowska-Prasał, I; Staszyc, J; Kifer, E; Romanowska-Sarlej, J
1989-01-01
The studies were performed on 44 white female Wistar rats which were intratracheally administered the suspension of the soil dust and the electro-energetic ashes. The electro-energetic ashes were collected from 6 different local heat and power generating plants while the soil dust from several random places of our country. The statistical analysis of the body and the lung mass of the animals subjected to the single dust and ash insufflation was performed. The applied variants proved the statistically significant differences between the body and the lung mass. The observed differences are connected with the kinds of dust and ash used in the experiment.
50 CFR 600.410 - Collection and maintenance of statistics.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 50 Wildlife and Fisheries 10 2011-10-01 2011-10-01 false Collection and maintenance of statistics... of Statistics § 600.410 Collection and maintenance of statistics. (a) General. (1) All statistics..., the Assistant Administrator will remove all identifying particulars from the statistics if doing so is...
50 CFR 600.410 - Collection and maintenance of statistics.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 50 Wildlife and Fisheries 8 2010-10-01 2010-10-01 false Collection and maintenance of statistics... of Statistics § 600.410 Collection and maintenance of statistics. (a) General. (1) All statistics..., the Assistant Administrator will remove all identifying particulars from the statistics if doing so is...
Alaska national hydrography dataset positional accuracy assessment study
Arundel, Samantha; Yamamoto, Kristina H.; Constance, Eric; Mantey, Kim; Vinyard-Houx, Jeremy
2013-01-01
Initial visual assessments Wide range in the quality of fit between features in NHD and these new image sources. No statistical analysis has been performed to actually quantify accuracy Determining absolute accuracy is cost prohibitive (must collect independent, well defined test points) Quantitative analysis of relative positional error is feasible.
75 FR 61136 - Notice of Proposed Information Collection Requests
Federal Register 2010, 2011, 2012, 2013, 2014
2010-10-04
... EdFacts data as well as data from surveys of school principals and special education designees about their school improvement practices. The study will use descriptive statistics and regression analysis to...
NASA Astrophysics Data System (ADS)
Valder, J.; Kenner, S.; Long, A.
2008-12-01
Portions of the Cheyenne River are characterized as impaired by the U.S. Environmental Protection Agency because of water-quality exceedences. The Cheyenne River watershed includes the Black Hills National Forest and part of the Badlands National Park. Preliminary analysis indicates that the Badlands National Park is a major contributor to the exceedances of the water-quality constituents for total dissolved solids and total suspended solids. Water-quality data have been collected continuously since 2007, and in the second year of collection (2008), monthly grab and passive sediment samplers are being used to collect total suspended sediment and total dissolved solids in both base-flow and runoff-event conditions. In addition, sediment samples from the river channel, including bed, bank, and floodplain, have been collected. These samples are being analyzed at the South Dakota School of Mines and Technology's X-Ray Diffraction Lab to quantify the mineralogy of the sediments. A multivariate statistical approach (including principal components, least squares, and maximum likelihood techniques) is applied to the mineral percentages that were characterized for each site to identify the contributing source areas that are causing exceedances of sediment transport in the Cheyenne River watershed. Results of the multivariate analysis demonstrate the likely sources of solids found in the Cheyenne River samples. A further refinement of the methods is in progress that utilizes a conceptual model which, when applied with the multivariate statistical approach, provides a better estimate for sediment sources.
Benchmarking Usage Statistics in Collection Management Decisions for Serials
ERIC Educational Resources Information Center
Tucker, Cory
2009-01-01
Usage statistics are an important metric for making decisions on serials. Although the University of Nevada, Las Vegas (UNLV) Libraries have been collecting usage statistics, the statistics had not frequently been used to make decisions and had not been included in collection development policy. After undergoing a collection assessment, the…
Almeida, Tiago P; Chu, Gavin S; Li, Xin; Dastagir, Nawshin; Tuan, Jiun H; Stafford, Peter J; Schlindwein, Fernando S; Ng, G André
2017-01-01
Purpose: Complex fractionated atrial electrograms (CFAE)-guided ablation after pulmonary vein isolation (PVI) has been used for persistent atrial fibrillation (persAF) therapy. This strategy has shown suboptimal outcomes due to, among other factors, undetected changes in the atrial tissue following PVI. In the present work, we investigate CFAE distribution before and after PVI in patients with persAF using a multivariate statistical model. Methods: 207 pairs of atrial electrograms (AEGs) were collected before and after PVI respectively, from corresponding LA regions in 18 persAF patients. Twelve attributes were measured from the AEGs, before and after PVI. Statistical models based on multivariate analysis of variance (MANOVA) and linear discriminant analysis (LDA) have been used to characterize the atrial regions and AEGs. Results: PVI significantly reduced CFAEs in the LA (70 vs. 40%; P < 0.0001). Four types of LA regions were identified, based on the AEGs characteristics: (i) fractionated before PVI that remained fractionated after PVI (31% of the collected points); (ii) fractionated that converted to normal (39%); (iii) normal prior to PVI that became fractionated (9%) and; (iv) normal that remained normal (21%). Individually, the attributes failed to distinguish these LA regions, but multivariate statistical models were effective in their discrimination ( P < 0.0001). Conclusion: Our results have unveiled that there are LA regions resistant to PVI, while others are affected by it. Although, traditional methods were unable to identify these different regions, the proposed multivariate statistical model discriminated LA regions resistant to PVI from those affected by it without prior ablation information.
Quantifying Variation in Gait Features from Wearable Inertial Sensors Using Mixed Effects Models
Cresswell, Kellen Garrison; Shin, Yongyun; Chen, Shanshan
2017-01-01
The emerging technology of wearable inertial sensors has shown its advantages in collecting continuous longitudinal gait data outside laboratories. This freedom also presents challenges in collecting high-fidelity gait data. In the free-living environment, without constant supervision from researchers, sensor-based gait features are susceptible to variation from confounding factors such as gait speed and mounting uncertainty, which are challenging to control or estimate. This paper is one of the first attempts in the field to tackle such challenges using statistical modeling. By accepting the uncertainties and variation associated with wearable sensor-based gait data, we shift our efforts from detecting and correcting those variations to modeling them statistically. From gait data collected on one healthy, non-elderly subject during 48 full-factorial trials, we identified four major sources of variation, and quantified their impact on one gait outcome—range per cycle—using a random effects model and a fixed effects model. The methodology developed in this paper lays the groundwork for a statistical framework to account for sources of variation in wearable gait data, thus facilitating informative statistical inference for free-living gait analysis. PMID:28245602
New software for statistical analysis of Cambridge Structural Database data
Sykes, Richard A.; McCabe, Patrick; Allen, Frank H.; Battle, Gary M.; Bruno, Ian J.; Wood, Peter A.
2011-01-01
A collection of new software tools is presented for the analysis of geometrical, chemical and crystallographic data from the Cambridge Structural Database (CSD). This software supersedes the program Vista. The new functionality is integrated into the program Mercury in order to provide statistical, charting and plotting options alongside three-dimensional structural visualization and analysis. The integration also permits immediate access to other information about specific CSD entries through the Mercury framework, a common requirement in CSD data analyses. In addition, the new software includes a range of more advanced features focused towards structural analysis such as principal components analysis, cone-angle correction in hydrogen-bond analyses and the ability to deal with topological symmetry that may be exhibited in molecular search fragments. PMID:22477784
Onisko, Agnieszka; Druzdzel, Marek J; Austin, R Marshall
2016-01-01
Classical statistics is a well-established approach in the analysis of medical data. While the medical community seems to be familiar with the concept of a statistical analysis and its interpretation, the Bayesian approach, argued by many of its proponents to be superior to the classical frequentist approach, is still not well-recognized in the analysis of medical data. The goal of this study is to encourage data analysts to use the Bayesian approach, such as modeling with graphical probabilistic networks, as an insightful alternative to classical statistical analysis of medical data. This paper offers a comparison of two approaches to analysis of medical time series data: (1) classical statistical approach, such as the Kaplan-Meier estimator and the Cox proportional hazards regression model, and (2) dynamic Bayesian network modeling. Our comparison is based on time series cervical cancer screening data collected at Magee-Womens Hospital, University of Pittsburgh Medical Center over 10 years. The main outcomes of our comparison are cervical cancer risk assessments produced by the three approaches. However, our analysis discusses also several aspects of the comparison, such as modeling assumptions, model building, dealing with incomplete data, individualized risk assessment, results interpretation, and model validation. Our study shows that the Bayesian approach is (1) much more flexible in terms of modeling effort, and (2) it offers an individualized risk assessment, which is more cumbersome for classical statistical approaches.
Ferraro Petrillo, Umberto; Roscigno, Gianluca; Cattaneo, Giuseppe; Giancarlo, Raffaele
2018-06-01
Information theoretic and compositional/linguistic analysis of genomes have a central role in bioinformatics, even more so since the associated methodologies are becoming very valuable also for epigenomic and meta-genomic studies. The kernel of those methods is based on the collection of k-mer statistics, i.e. how many times each k-mer in {A,C,G,T}k occurs in a DNA sequence. Although this problem is computationally very simple and efficiently solvable on a conventional computer, the sheer amount of data available now in applications demands to resort to parallel and distributed computing. Indeed, those type of algorithms have been developed to collect k-mer statistics in the realm of genome assembly. However, they are so specialized to this domain that they do not extend easily to the computation of informational and linguistic indices, concurrently on sets of genomes. Following the well-established approach in many disciplines, and with a growing success also in bioinformatics, to resort to MapReduce and Hadoop to deal with 'Big Data' problems, we present KCH, the first set of MapReduce algorithms able to perform concurrently informational and linguistic analysis of large collections of genomic sequences on a Hadoop cluster. The benchmarking of KCH that we provide indicates that it is quite effective and versatile. It is also competitive with respect to the parallel and distributed algorithms highly specialized to k-mer statistics collection for genome assembly problems. In conclusion, KCH is a much needed addition to the growing number of algorithms and tools that use MapReduce for bioinformatics core applications. The software, including instructions for running it over Amazon AWS, as well as the datasets are available at http://www.di-srv.unisa.it/KCH. umberto.ferraro@uniroma1.it. Supplementary data are available at Bioinformatics online.
Modelling the Effects of Land-Use Changes on Climate: a Case Study on Yamula DAM
NASA Astrophysics Data System (ADS)
Köylü, Ü.; Geymen, A.
2016-10-01
Dams block flow of rivers and cause artificial water reservoirs which affect the climate and the land use characteristics of the river basin. In this research, the effect of the huge water body obtained by Yamula Dam in Kızılırmak Basin is analysed over surrounding spatial's land use and climate change. Mann Kendal non-parametrical statistical test, Theil&Sen Slope method, Inverse Distance Weighting (IDW), Soil Conservation Service-Curve Number (SCS-CN) methods are integrated for spatial and temporal analysis of the research area. For this research humidity, temperature, wind speed, precipitation observations which are collected in 16 weather stations nearby Kızılırmak Basin are analyzed. After that these statistical information is combined by GIS data over years. An application is developed for GIS analysis in Python Programming Language and integrated with ArcGIS software. Statistical analysis calculated in the R Project for Statistical Computing and integrated with developed application. According to the statistical analysis of extracted time series of meteorological parameters, statistical significant spatiotemporal trends are observed for climate change and land use characteristics. In this study, we indicated the effect of big dams in local climate on semi-arid Yamula Dam.
Dale D. Gormanson; Scott A. Pugh; Charles J. Barnett; Patrick D. Miles; Randall S. Morin; Paul A. Sowers; Jim Westfall
2017-01-01
The U.S. Forest Service Forest Inventory and Analysis (FIA) program collects sample plot data on all forest ownerships across the United States. FIA's primary objective is to determine the extent, condition, volume, growth, and use of trees on the Nation's forest land through a comprehensive inventory and analysis of the Nation's forest resources. The...
de Oliveira Moraes, Gláucia Sabrine; de Souza, Evaristo Jorge Oliveira; Véras, Antonia Sherlânea Chaves; de Paula Almeida, Marina; da Cunha, Márcio Vieira; Torres, Thaysa Rodrigues; da Silva, Camila Sousa; Pereira, Gerfesson Felipe Cavalcanti
2016-10-01
The objective of the present study to assess the effects of mesquite pod addition replacing corn (0, 250, 500, 750, and 1000 g/kg in the dry matter basis) on nutrient intake, animal performance, feeding behavior, nutrient digestibility, nitrogen balance, and microbial protein synthesis. Twenty-five Holstein-Zebu crossbred dairy steers at 219 ± 22 kg initial body weight and 18 months of age were used. The experiment lasted 84 days, divided into three periods of 28 days. A completely randomized design was used, and data were submitted to analysis using PROC GLM for analysis of variance and PROC REG for regression analysis using the software Statistical Analysis Systems version 9.1. Experimental diets were composed of Tifton 85 hay, soybean meal, ground corn, mesquite pod meal, and mineral salt. Samples of food offered were collected during the last 3 days of each period, and the leftovers were collected daily, with samples bulked per week. At the end of each 28-day period, the remaining animals were weighed to determine total weight gain and average daily gain. The assessment of behavioral patterns was performed through instantaneous scans in 5-min intervals for three consecutive 12-h days. A single urine sample from each animal was collected on the last day of each collection period at about 4 h after the first feeding. The replacement of corn by mesquite pod meal did not significantly influence treatments regarding nutrients intake, animal performance, and feeding behavior. Retained and consumed nitrogen ratio did not statistically differ between replacement levels. Likewise, there were no statistical differences regarding microbial protein synthesis and efficiency between replacement levels. Mesquite pod meal can be used in Holstein-Zebu crossbred dairy steers' diet with total corn replacement.
Skouroliakou, Maria; Soloupis, George; Gounaris, Antonis; Charitou, Antonia; Papasarantopoulos, Petros; Markantonis, Sophia L; Golna, Christina; Souliotis, Kyriakos
2008-07-28
This study assesses the results of implementation of a software program that allows for input of admission/discharge summary data (including cost) in a neonatal intensive care unit (NICU) in Greece, based on the establishment of a baseline statistical database for infants treated in a NICU and the statistical analysis of epidemiological and resource utilization data thus collected. A software tool was designed, developed, and implemented between April 2004 and March 2005 in the NICU of the LITO private maternity hospital in Athens, Greece, to allow for the first time for step-by-step collection and management of summary treatment data. Data collected over this period were subsequently analyzed using defined indicators as a basis to extract results related to treatment options, treatment duration, and relative resource utilization. Data for 499 babies were entered in the tool and processed. Information on medical costs (e.g., mean total cost +/- SD of treatment was euro310.44 +/- 249.17 and euro6704.27 +/- 4079.53 for babies weighing more than 2500 g and 1000-1500 g respectively), incidence of complications or disease (e.g., 4.3 percent and 14.3 percent of study babies weighing 1,000 to 1,500 g suffered from cerebral bleeding [grade I] and bronchopulmonary dysplasia, respectively, while overall 6.0 percent had microbial infections), and medical statistics (e.g., perinatal mortality was 6.8 percent) was obtained in a quick and robust manner. The software tool allowed for collection and analysis of data traditionally maintained in paper medical records in the NICU with greater ease and accuracy. Data codification and analysis led to significant findings at the epidemiological, medical resource utilization, and respective hospital cost levels that allowed comparisons with literature findings for the first time in Greece. The tool thus contributed to a clearer understanding of treatment practices in the NICU and set the baseline for the assessment of the impact of future interventions at the policy or hospital level.
Statistical Analysis of CFD Solutions from the 6th AIAA CFD Drag Prediction Workshop
NASA Technical Reports Server (NTRS)
Derlaga, Joseph M.; Morrison, Joseph H.
2017-01-01
A graphical framework is used for statistical analysis of the results from an extensive N- version test of a collection of Reynolds-averaged Navier-Stokes computational uid dynam- ics codes. The solutions were obtained by code developers and users from North America, Europe, Asia, and South America using both common and custom grid sequencees as well as multiple turbulence models for the June 2016 6th AIAA CFD Drag Prediction Workshop sponsored by the AIAA Applied Aerodynamics Technical Committee. The aerodynamic con guration for this workshop was the Common Research Model subsonic transport wing- body previously used for both the 4th and 5th Drag Prediction Workshops. This work continues the statistical analysis begun in the earlier workshops and compares the results from the grid convergence study of the most recent workshop with previous workshops.
Which statistics should tropical biologists learn?
Loaiza Velásquez, Natalia; González Lutz, María Isabel; Monge-Nájera, Julián
2011-09-01
Tropical biologists study the richest and most endangered biodiversity in the planet, and in these times of climate change and mega-extinctions, the need for efficient, good quality research is more pressing than in the past. However, the statistical component in research published by tropical authors sometimes suffers from poor quality in data collection; mediocre or bad experimental design and a rigid and outdated view of data analysis. To suggest improvements in their statistical education, we listed all the statistical tests and other quantitative analyses used in two leading tropical journals, the Revista de Biología Tropical and Biotropica, during a year. The 12 most frequent tests in the articles were: Analysis of Variance (ANOVA), Chi-Square Test, Student's T Test, Linear Regression, Pearson's Correlation Coefficient, Mann-Whitney U Test, Kruskal-Wallis Test, Shannon's Diversity Index, Tukey's Test, Cluster Analysis, Spearman's Rank Correlation Test and Principal Component Analysis. We conclude that statistical education for tropical biologists must abandon the old syllabus based on the mathematical side of statistics and concentrate on the correct selection of these and other procedures and tests, on their biological interpretation and on the use of reliable and friendly freeware. We think that their time will be better spent understanding and protecting tropical ecosystems than trying to learn the mathematical foundations of statistics: in most cases, a well designed one-semester course should be enough for their basic requirements.
2017-05-25
37 Research Design ... research employed a mixed research methodology – quantitative with descriptive statistical analysis and qualitative with a thematic analysis approach...mixed research methodology – quantitative and qualitative, using interviews to collect the data. The interviews included demographic and open-ended
Using Rasch Analysis to Identify Uncharacteristic Responses to Undergraduate Assessments
ERIC Educational Resources Information Center
Edwards, Antony; Alcock, Lara
2010-01-01
Rasch Analysis is a statistical technique that is commonly used to analyse both test data and Likert survey data, to construct and evaluate question item banks, and to evaluate change in longitudinal studies. In this article, we introduce the dichotomous Rasch model, briefly discussing its assumptions. Then, using data collected in an…
Private forest owners and invasive plants: risk perception and management
A. Paige Fischer; Susan Charnley
2012-01-01
We investigated nonindustrial private forest (NIPF) owners' invasive plant risk perceptions and mitigation practices using statistical analysis of mail survey data and qualitative analysis of interview data collected in Oregon's ponderosa pine zone. We found that 52% of the survey sample was aware of invasive plant species considered problematic by local...
Statistical Analysis of CFD Solutions from the Fourth AIAA Drag Prediction Workshop
NASA Technical Reports Server (NTRS)
Morrison, Joseph H.
2010-01-01
A graphical framework is used for statistical analysis of the results from an extensive N-version test of a collection of Reynolds-averaged Navier-Stokes computational fluid dynamics codes. The solutions were obtained by code developers and users from the U.S., Europe, Asia, and Russia using a variety of grid systems and turbulence models for the June 2009 4th Drag Prediction Workshop sponsored by the AIAA Applied Aerodynamics Technical Committee. The aerodynamic configuration for this workshop was a new subsonic transport model, the Common Research Model, designed using a modern approach for the wing and included a horizontal tail. The fourth workshop focused on the prediction of both absolute and incremental drag levels for wing-body and wing-body-horizontal tail configurations. This work continues the statistical analysis begun in the earlier workshops and compares the results from the grid convergence study of the most recent workshop with earlier workshops using the statistical framework.
Vahedi, Shahrum; Farrokhi, Farahman; Gahramani, Farahnaz; Issazadegan, Ali
2012-01-01
Objective: Approximately 66-80%of graduate students experience statistics anxiety and some researchers propose that many students identify statistics courses as the most anxiety-inducing courses in their academic curriculums. As such, it is likely that statistics anxiety is, in part, responsible for many students delaying enrollment in these courses for as long as possible. This paper proposes a canonical model by treating academic procrastination (AP), learning strategies (LS) as predictor variables and statistics anxiety (SA) as explained variables. Methods: A questionnaire survey was used for data collection and 246-college female student participated in this study. To examine the mutually independent relations between procrastination, learning strategies and statistics anxiety variables, a canonical correlation analysis was computed. Results: Findings show that two canonical functions were statistically significant. The set of variables (metacognitive self-regulation, source management, preparing homework, preparing for test and preparing term papers) helped predict changes of statistics anxiety with respect to fearful behavior, Attitude towards math and class, Performance, but not Anxiety. Conclusion: These findings could be used in educational and psychological interventions in the context of statistics anxiety reduction. PMID:24644468
Environmental Health Practice: Statistically Based Performance Measurement
Enander, Richard T.; Gagnon, Ronald N.; Hanumara, R. Choudary; Park, Eugene; Armstrong, Thomas; Gute, David M.
2007-01-01
Objectives. State environmental and health protection agencies have traditionally relied on a facility-by-facility inspection-enforcement paradigm to achieve compliance with government regulations. We evaluated the effectiveness of a new approach that uses a self-certification random sampling design. Methods. Comprehensive environmental and occupational health data from a 3-year statewide industry self-certification initiative were collected from representative automotive refinishing facilities located in Rhode Island. Statistical comparisons between baseline and postintervention data facilitated a quantitative evaluation of statewide performance. Results. The analysis of field data collected from 82 randomly selected automotive refinishing facilities showed statistically significant improvements (P<.05, Fisher exact test) in 4 major performance categories: occupational health and safety, air pollution control, hazardous waste management, and wastewater discharge. Statistical significance was also shown when a modified Bonferroni adjustment for multiple comparisons was performed. Conclusions. Our findings suggest that the new self-certification approach to environmental and worker protection is effective and can be used as an adjunct to further enhance state and federal enforcement programs. PMID:17267709
Holmes, Susan; Alekseyenko, Alexander; Timme, Alden; Nelson, Tyrrell; Pasricha, Pankaj Jay; Spormann, Alfred
2011-01-01
This article explains the statistical and computational methodology used to analyze species abundances collected using the LNBL Phylochip in a study of Irritable Bowel Syndrome (IBS) in rats. Some tools already available for the analysis of ordinary microarray data are useful in this type of statistical analysis. For instance in correcting for multiple testing we use Family Wise Error rate control and step-down tests (available in the multtest package). Once the most significant species are chosen we use the hypergeometric tests familiar for testing GO categories to test specific phyla and families. We provide examples of normalization, multivariate projections, batch effect detection and integration of phylogenetic covariation, as well as tree equalization and robustification methods.
NASA Astrophysics Data System (ADS)
Dennison, Andrew G.
Classification of the seafloor substrate can be done with a variety of methods. These methods include Visual (dives, drop cameras); mechanical (cores, grab samples); acoustic (statistical analysis of echosounder returns). Acoustic methods offer a more powerful and efficient means of collecting useful information about the bottom type. Due to the nature of an acoustic survey, larger areas can be sampled, and by combining the collected data with visual and mechanical survey methods provide greater confidence in the classification of a mapped region. During a multibeam sonar survey, both bathymetric and backscatter data is collected. It is well documented that the statistical characteristic of a sonar backscatter mosaic is dependent on bottom type. While classifying the bottom-type on the basis on backscatter alone can accurately predict and map bottom-type, i.e a muddy area from a rocky area, it lacks the ability to resolve and capture fine textural details, an important factor in many habitat mapping studies. Statistical processing of high-resolution multibeam data can capture the pertinent details about the bottom-type that are rich in textural information. Further multivariate statistical processing can then isolate characteristic features, and provide the basis for an accurate classification scheme. The development of a new classification method is described here. It is based upon the analysis of textural features in conjunction with ground truth sampling. The processing and classification result of two geologically distinct areas in nearshore regions of Lake Superior; off the Lester River,MN and Amnicon River, WI are presented here, using the Minnesota Supercomputer Institute's Mesabi computing cluster for initial processing. Processed data is then calibrated using ground truth samples to conduct an accuracy assessment of the surveyed areas. From analysis of high-resolution bathymetry data collected at both survey sites is was possible to successfully calculate a series of measures that describe textural information about the lake floor. Further processing suggests that the features calculated capture a significant amount of statistical information about the lake floor terrain as well. Two sources of error, an anomalous heave and refraction error significantly deteriorated the quality of the processed data and resulting validate results. Ground truth samples used to validate the classification methods utilized for both survey sites, however, resulted in accuracy values ranging from 5 -30 percent at the Amnicon River, and between 60-70 percent for the Lester River. The final results suggest that this new processing methodology does adequately capture textural information about the lake floor and does provide an acceptable classification in the absence of significant data quality issues.
Federal Register 2010, 2011, 2012, 2013, 2014
2010-08-19
... of new technologies, communication and travel options, as well as social norms will influence... behavior, perspectives and social norms not covered through the statistical analysis. This is the first...
Mathematical problem solving ability of sport students in the statistical study
NASA Astrophysics Data System (ADS)
Sari, E. F. P.; Zulkardi; Putri, R. I. I.
2017-12-01
This study aims to determine the problem-solving ability of sport students of PGRI Palembang semester V in the statistics course. Subjects in this study were sport students of PGRI Palembang semester V which amounted to 31 people. The research method used is quasi experiment type one case shoot study. Data collection techniques in this study use the test and data analysis used is quantitative descriptive statistics. The conclusion of this study shown that the mathematical problem solving ability of PGRI Palembang sport students of V semester in the statistical course is categorized well with the average of the final test score of 80.3.
Statistical summaries of selected Iowa streamflow data through September 2013
Eash, David A.; O'Shea, Padraic S.; Weber, Jared R.; Nguyen, Kevin T.; Montgomery, Nicholas L.; Simonson, Adrian J.
2016-01-04
Statistical summaries of streamflow data collected at 184 streamgages in Iowa are presented in this report. All streamgages included for analysis have at least 10 years of continuous record collected before or through September 2013. This report is an update to two previously published reports that presented statistical summaries of selected Iowa streamflow data through September 1988 and September 1996. The statistical summaries include (1) monthly and annual flow durations, (2) annual exceedance probabilities of instantaneous peak discharges (flood frequencies), (3) annual exceedance probabilities of high discharges, and (4) annual nonexceedance probabilities of low discharges and seasonal low discharges. Also presented for each streamgage are graphs of the annual mean discharges, mean annual mean discharges, 50-percent annual flow-duration discharges (median flows), harmonic mean flows, mean daily mean discharges, and flow-duration curves. Two sets of statistical summaries are presented for each streamgage, which include (1) long-term statistics for the entire period of streamflow record and (2) recent-term statistics for or during the 30-year period of record from 1984 to 2013. The recent-term statistics are only calculated for streamgages with streamflow records pre-dating the 1984 water year and with at least 10 years of record during 1984–2013. The streamflow statistics in this report are not adjusted for the effects of water use; although some of this water is used consumptively, most of it is returned to the streams.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gilbert, R O; Essington, E H; Brady, D N
Statistical design and analysis activities for the Nevada Applied Ecology Group (NAEG) during 1976 are briefly outlined. This is followed by a description of soil data collected thus far at nuclear study sites. Radionuclide concentrations in surface soil collected along a transect from ground zero (GZ) along the main fallout pattern are given for Nuclear Site (NS) 201. Concentrations in soil collected at 315 locations on a grid system at 200 foot spacings are also given for this site. The /sup 241/Am to /sup 137/Cs ratios change over NS 201 depending on location relative to GZ. They range from lessmore » than one where /sup 241/Am is at low levels, to more than fifty where /sup 241/Am levels are high (near GZ). The estimated median /sup 239/ /sup 240/Pu to /sup 241/Am ratio is 11 and appears to be relatively constant over the area (the 95 percent lower and upper limits on the true median ratio are about 8 and 14).« less
ERIC Educational Resources Information Center
Boyce, Jared; Bowers, Alex J.
2018-01-01
This study investigated the differences between how individual teachers perceive leadership for learning and how teachers collectively perceive leadership for learning, using a large nationally generalizable data-set of 7070 schools from the National Center for Education Statistics 2011-2012 Schools and Staffing Survey. This study used…
An Exploratory Data Analysis System for Support in Medical Decision-Making
Copeland, J. A.; Hamel, B.; Bourne, J. R.
1979-01-01
An experimental system was developed to allow retrieval and analysis of data collected during a study of neurobehavioral correlates of renal disease. After retrieving data organized in a relational data base, simple bivariate statistics of parametric and nonparametric nature could be conducted. An “exploratory” mode in which the system provided guidance in selection of appropriate statistical analyses was also available to the user. The system traversed a decision tree using the inherent qualities of the data (e.g., the identity and number of patients, tests, and time epochs) to search for the appropriate analyses to employ.
Goyal, Ravi; De Gruttola, Victor
2018-01-30
Analysis of sexual history data intended to describe sexual networks presents many challenges arising from the fact that most surveys collect information on only a very small fraction of the population of interest. In addition, partners are rarely identified and responses are subject to reporting biases. Typically, each network statistic of interest, such as mean number of sexual partners for men or women, is estimated independently of other network statistics. There is, however, a complex relationship among networks statistics; and knowledge of these relationships can aid in addressing concerns mentioned earlier. We develop a novel method that constrains a posterior predictive distribution of a collection of network statistics in order to leverage the relationships among network statistics in making inference about network properties of interest. The method ensures that inference on network properties is compatible with an actual network. Through extensive simulation studies, we also demonstrate that use of this method can improve estimates in settings where there is uncertainty that arises both from sampling and from systematic reporting bias compared with currently available approaches to estimation. To illustrate the method, we apply it to estimate network statistics using data from the Chicago Health and Social Life Survey. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.
Brix, Tobias Johannes; Bruland, Philipp; Sarfraz, Saad; Ernsting, Jan; Neuhaus, Philipp; Storck, Michael; Doods, Justin; Ständer, Sonja; Dugas, Martin
2018-01-01
A required step for presenting results of clinical studies is the declaration of participants demographic and baseline characteristics as claimed by the FDAAA 801. The common workflow to accomplish this task is to export the clinical data from the used electronic data capture system and import it into statistical software like SAS software or IBM SPSS. This software requires trained users, who have to implement the analysis individually for each item. These expenditures may become an obstacle for small studies. Objective of this work is to design, implement and evaluate an open source application, called ODM Data Analysis, for the semi-automatic analysis of clinical study data. The system requires clinical data in the CDISC Operational Data Model format. After uploading the file, its syntax and data type conformity of the collected data is validated. The completeness of the study data is determined and basic statistics, including illustrative charts for each item, are generated. Datasets from four clinical studies have been used to evaluate the application's performance and functionality. The system is implemented as an open source web application (available at https://odmanalysis.uni-muenster.de) and also provided as Docker image which enables an easy distribution and installation on local systems. Study data is only stored in the application as long as the calculations are performed which is compliant with data protection endeavors. Analysis times are below half an hour, even for larger studies with over 6000 subjects. Medical experts have ensured the usefulness of this application to grant an overview of their collected study data for monitoring purposes and to generate descriptive statistics without further user interaction. The semi-automatic analysis has its limitations and cannot replace the complex analysis of statisticians, but it can be used as a starting point for their examination and reporting.
ERIC Educational Resources Information Center
Larson, Ray R.
1996-01-01
Examines the bibliometrics of the World Wide Web based on analysis of Web pages collected by the Inktomi "Web Crawler" and on the use of the DEC AltaVista search engine for cocitation analysis of a set of Earth Science related Web sites. Looks at the statistical characteristics of Web documents and their hypertext links, and the…
Synthetic Training Data Generation for Activity Monitoring and Behavior Analysis
NASA Astrophysics Data System (ADS)
Monekosso, Dorothy; Remagnino, Paolo
This paper describes a data generator that produces synthetic data to simulate observations from an array of environment monitoring sensors. The overall goal of our work is to monitor the well-being of one occupant in a home. Sensors are embedded in a smart home to unobtrusively record environmental parameters. Based on the sensor observations, behavior analysis and modeling are performed. However behavior analysis and modeling require large data sets to be collected over long periods of time to achieve the level of accuracy expected. A data generator - was developed based on initial data i.e. data collected over periods lasting weeks to facilitate concurrent data collection and development of algorithms. The data generator is based on statistical inference techniques. Variation is introduced into the data using perturbation models.
[Analysis the epidemiological features of 3,258 patients with allergic rhinitis in Yichang City].
Chen, Bo; Zhang, Zhimao; Pei, Zhi; Chen, Shihan; Du, Zhimei; Lan, Yan; Han, Bei; Qi, Qi
2015-02-01
To investigate the epidemiological features in patients with allergic rhinitis (AR) in Yichang city, and put forward effective prevention and control measures. Collecting the data of allergic rhinitis in city proper from 2010 to 2013, input the data into the database and used statistical analysis. In recent years, the AR patients in this area increased year by year. The spring and the winter were the peak season of onset. The patients was constituted by young men. There was statistically significant difference between the age, the area,and the gender (P < 0.01). The history of allergy and the diseases related to the gender composition had statistical significance difference (P < 0.05). The allergens and the positive degree in gender, age structure had statistically significant difference (P < 0.01). Need to conduct the healthy propaganda and education, optimizing the environment, change the bad habits, timely medical treatment, standard treatment.
Portraits of self-organization in fish schools interacting with robots
NASA Astrophysics Data System (ADS)
Aureli, M.; Fiorilli, F.; Porfiri, M.
2012-05-01
In this paper, we propose an enabling computational and theoretical framework for the analysis of experimental instances of collective behavior in response to external stimuli. In particular, this work addresses the characterization of aggregation and interaction phenomena in robot-animal groups through the exemplary analysis of fish schooling in the vicinity of a biomimetic robot. We adapt global observables from statistical mechanics to capture the main features of the shoal collective motion and its response to the robot from experimental observations. We investigate the shoal behavior by using a diffusion mapping analysis performed on these global observables that also informs the definition of relevant portraits of self-organization.
Cores Of Recurrent Events (CORE) | Informatics Technology for Cancer Research (ITCR)
CORE is a statistically supported computational method for finding recurrently targeted regions in massive collections of genomic intervals, such as those arising from DNA copy number analysis of single tumor cells or bulk tumor tissues.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-12-14
... completed and validated, the hardcopy questionnaires will be discarded. Data will be imported into SPSS (Statistical Package for the Social Sciences) for analysis. The database will be maintained at the respective...
Computer-assisted qualitative data analysis software.
Cope, Diane G
2014-05-01
Advances in technology have provided new approaches for data collection methods and analysis for researchers. Data collection is no longer limited to paper-and-pencil format, and numerous methods are now available through Internet and electronic resources. With these techniques, researchers are not burdened with entering data manually and data analysis is facilitated by software programs. Quantitative research is supported by the use of computer software and provides ease in the management of large data sets and rapid analysis of numeric statistical methods. New technologies are emerging to support qualitative research with the availability of computer-assisted qualitative data analysis software (CAQDAS).CAQDAS will be presented with a discussion of advantages, limitations, controversial issues, and recommendations for this type of software use.
2002-06-01
fits our actual data . To determine the goodness of fit, statisticians typically use the following four measures: R2 Statistic. The R2 statistic...reviewing instruction, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of...mathematical model is developed to better estimate cleanup costs using historical cost data that could be used by the Defense Department prior to placing
ERIC Educational Resources Information Center
Dega, Bekele Gashe; Govender, Nadaraj
2016-01-01
This study compares the scientific and alternative conceptions of energy and momentum of university first-year science students in Ethiopia and the US. Written data were collected using the Energy and Momentum Conceptual Survey developed by Singh and Rosengrant. The Concentration Analysis statistical method was used for analysing the Ethiopian…
Ninety six gasoline samples were collected from around the U.S. in Autumn 2004. A detailed hydrocarbon analysis was performed on each sample resulting in a data set of approximately 300 chemicals per sample. Statistical analyses were performed on the entire suite of reported chem...
Implementation of Head Start Planned Variation: 1970-1971. Part II.
ERIC Educational Resources Information Center
Lukas, Carol Van Deusen; Wohlleb, Cynthia
This volume of appendices is Part II of a study of program implementation in 12 models of Head Start Planned Variation. It presents details of the data analysis, copies of data collection instruments, and additional analyses and statistics. The appendices are: (A) Analysis of Variance Designs, (B) Copies of Instruments, (C) Additional Analyses,…
ERIC Educational Resources Information Center
Torrens, Paul M.; Griffin, William A.
2013-01-01
The authors describe an observational and analytic methodology for recording and interpreting dynamic microprocesses that occur during social interaction, making use of space--time data collection techniques, spatial-statistical analysis, and visualization. The scheme has three investigative foci: Structure, Activity Composition, and Clustering.…
ERIC Educational Resources Information Center
Muslihah, Oleh Eneng
2015-01-01
The research examines the correlation between the understanding of school-based management, emotional intelligences and headmaster performance. Data was collected, using quantitative methods. The statistical analysis used was the Pearson Correlation, and multivariate regression analysis. The results of this research suggest firstly that there is…
Digital Natives, Digital Immigrants: An Analysis of Age and ICT Competency in Teacher Education
ERIC Educational Resources Information Center
Guo, Ruth Xiaoqing; Dobson, Teresa; Petrina, Stephen
2008-01-01
This article examines the intersection of age and ICT (information and communication technology) competency and critiques the "digital natives versus digital immigrants" argument proposed by Prensky (2001a, 2001b). Quantitative analysis was applied to a statistical data set collected in the context of a study with over 2,000 pre-service…
A Content Analysis of Dissertations in the Field of Educational Technology: The Case of Turkey
ERIC Educational Resources Information Center
Durak, Gurhan; Cankaya, Serkan; Yunkul, Eyup; Misirli, Zeynel Abidin
2018-01-01
The present study aimed at conducting content analysis on dissertations carried out so far in the field of Educational Technology in Turkey. A total of 137 dissertations were examined to determine the key words, academic discipline, research areas, theoretical frameworks, research designs and models, statistical analyses, data collection tools,…
Patton, Charles J.; Gilroy, Edward J.
1999-01-01
Data on which this report is based, including nutrient concentrations in synthetic reference samples determined concurrently with those in real samples, are extensive (greater than 20,000 determinations) and have been published separately. In addition to confirming the well-documented instability of nitrite in acidified samples, this study also demonstrates that when biota are removed from samples at collection sites by 0.45-micrometer membrane filtration, subsequent preservation with sulfuric acid or mercury (II) provides no statistically significant improvement in nutrient concentration stability during storage at 4 degrees Celsius for 30 days. Biocide preservation had no statistically significant effect on the 30-day stability of phosphorus concentrations in whole-water splits from any of the 15 stations, but did stabilize Kjeldahl nitrogen concentrations in whole-water splits from three data-collection stations where ammonium accounted for at least half of the measured Kjeldahl nitrogen.
Utah Virtual Lab: JAVA interactivity for teaching science and statistics on line.
Malloy, T E; Jensen, G C
2001-05-01
The Utah on-line Virtual Lab is a JAVA program run dynamically off a database. It is embedded in StatCenter (www.psych.utah.edu/learn/statsampler.html), an on-line collection of tools and text for teaching and learning statistics. Instructors author a statistical virtual reality that simulates theories and data in a specific research focus area by defining independent, predictor, and dependent variables and the relations among them. Students work in an on-line virtual environment to discover the principles of this simulated reality: They go to a library, read theoretical overviews and scientific puzzles, and then go to a lab, design a study, collect and analyze data, and write a report. Each student's design and data analysis decisions are computer-graded and recorded in a database; the written research report can be read by the instructor or by other students in peer groups simulating scientific conventions.
Dodge, Kent A.; Hornberger, Michelle I.; Turner, Matthew A.
2018-03-30
Water, bed sediment, and biota were sampled in selected streams from Butte to near Missoula, Montana, as part of a monitoring program in the upper Clark Fork Basin of western Montana. The sampling program was led by the U.S. Geological Survey, in cooperation with the U.S. Environmental Protection Agency, to characterize aquatic resources in the Clark Fork Basin, with emphasis on trace elements associated with historic mining and smelting activities. Sampling sites were on the Clark Fork and selected tributaries. Water samples were collected periodically at 20 sites from October 2015 through September 2016. Bed-sediment and biota samples were collected once at 13 sites during August 2016.This report presents the analytical results and quality-assurance data for water-quality, bed-sediment, and biota samples collected at sites from October 2015 through September 2016. Water-quality data include concentrations of selected major ions, trace elements, and suspended sediment. Samples for analysis of turbidity were collected at 13 sites, whereas samples for analysis of dissolved organic carbon were collected at 10 sites. In addition, samples for analysis of nitrogen (nitrate plus nitrite) were collected at two sites. Daily values of mean suspended-sediment concentration and suspended-sediment discharge were determined for three sites. Seasonal daily values of turbidity were determined for five sites. Bed-sediment data include trace-element concentrations in the fine-grained (less than 0.063 millimeter) fraction. Biological data include trace-element concentrations in whole-body tissue of aquatic benthic insects. Statistical summaries of water-quality, bed-sediment, and biological data for sites in the upper Clark Fork Basin are provided for the period of record.
Performance Analysis of Garbage Collection and Dynamic Reordering in a Lisp System. Ph.D. Thesis
NASA Technical Reports Server (NTRS)
Llames, Rene Lim
1991-01-01
Generation based garbage collection and dynamic reordering of objects are two techniques for improving the efficiency of memory management in Lisp and similar dynamic language systems. An analysis of the effect of generation configuration is presented, focusing on the effect of a number of generations and generation capabilities. Analytic timing and survival models are used to represent garbage collection runtime and to derive structural results on its behavior. The survival model provides bounds on the age of objects surviving a garbage collection at a particular level. Empirical results show that execution time is most sensitive to the capacity of the youngest generation. A technique called scanning for transport statistics, for evaluating the effectiveness of reordering independent of main memory size, is presented.
2008 Homeland Security S and T Stakeholders Conference West-Volume 3 Tuesday
2008-01-16
Architecture ( PNNL SRS) • Online data collection / entry • Data Warehouse • On Demand Analysis and Reporting Tools • Reports, Charts & Graphs • Visual / Data...Sustainability 2007– 2016 Our region wide investment include all PANYNJ business areas Computer Statistical Analysis COMPSTAT •NYPD 1990’s •Personnel Management...Coast Guard, and public health Expertise, Depth, Agility Staff Degrees 6 Our Value Added Capabilities • Risk Analysis • Operations Analysis
76 FR 18582 - Agency Information Collection Activities: Proposed Collection: Comments Requested
Federal Register 2010, 2011, 2012, 2013, 2014
2011-04-04
... (OJP), Bureau of Justice Statistics (BJS), will be submitting the following information collection... Christine Eith, Bureau of Justice Statistics, 810 Seventh Street, NW., Washington, DC 20531 (phone: 202- 305... applicable component of the Department sponsoring the collection: PPCS-1. Bureau of Justice Statistics...
Tannery, Nancy Hrinya; Silverman, Deborah L; Epstein, Barbara A
2002-01-01
Online use statistics can provide libraries with a tool to be used when developing an online collection of resources. Statistics can provide information on overall use of a collection, individual print and electronic journal use, and collection use by specific user populations. They can also be used to determine the number of user licenses to purchase. This paper focuses on the issue of use statistics made available for one collection of online resources.
Satellite temperature monitoring and prediction system
NASA Technical Reports Server (NTRS)
Barnett, U. R.; Martsolf, J. D.; Crosby, F. L.
1980-01-01
The paper describes the Florida Satellite Freeze Forecast System (SFFS) in its current state. All data collection options have been demonstrated, and data collected over a three year period have been stored for future analysis. Presently, specific minimum temperature forecasts are issued routinely from November through March. The procedures for issuing these forecast are discussed. The automated data acquisition and processing system is described, and the physical and statistical models employed are examined.
DOT National Transportation Integrated Search
2015-01-01
The Bureau of Transportation Statistics (BTS) is a leader in the collection, analysis, and dissemination of transportation data. The Transportation Services Index (TSI) measures the seasonally adjusted movement of freight traffic and passenger travel...
Assessment strategies for municipal selective waste collection schemes.
Ferreira, Fátima; Avelino, Catarina; Bentes, Isabel; Matos, Cristina; Teixeira, Carlos Afonso
2017-01-01
An important strategy to promote a strong sustainable growth relies on an efficient municipal waste management, and phasing out waste landfilling through waste prevention and recycling emerges as a major target. For this purpose, effective collection schemes are required, in particular those regarding selective waste collection, pursuing a more efficient and high quality recycling of reusable materials. This paper addresses the assessment and benchmarking of selective collection schemes, relevant to guide future operational improvements. In particular, the assessment is based on the monitoring and statistical analysis of a core-set of performance indicators that highlights collection trends, complemented with a performance index that gathers a weighted linear combination of these indicators. This combined analysis underlines a potential tool to support decision makers involved in the process of selecting the collection scheme with best overall performance. The presented approach was applied to a case study conducted in Oporto Municipality, with data gathered from two distinct selective collection schemes. Copyright © 2016 Elsevier Ltd. All rights reserved.
Zheng, Jie; Harris, Marcelline R; Masci, Anna Maria; Lin, Yu; Hero, Alfred; Smith, Barry; He, Yongqun
2016-09-14
Statistics play a critical role in biological and clinical research. However, most reports of scientific results in the published literature make it difficult for the reader to reproduce the statistical analyses performed in achieving those results because they provide inadequate documentation of the statistical tests and algorithms applied. The Ontology of Biological and Clinical Statistics (OBCS) is put forward here as a step towards solving this problem. The terms in OBCS including 'data collection', 'data transformation in statistics', 'data visualization', 'statistical data analysis', and 'drawing a conclusion based on data', cover the major types of statistical processes used in basic biological research and clinical outcome studies. OBCS is aligned with the Basic Formal Ontology (BFO) and extends the Ontology of Biomedical Investigations (OBI), an OBO (Open Biological and Biomedical Ontologies) Foundry ontology supported by over 20 research communities. Currently, OBCS comprehends 878 terms, representing 20 BFO classes, 403 OBI classes, 229 OBCS specific classes, and 122 classes imported from ten other OBO ontologies. We discuss two examples illustrating how the ontology is being applied. In the first (biological) use case, we describe how OBCS was applied to represent the high throughput microarray data analysis of immunological transcriptional profiles in human subjects vaccinated with an influenza vaccine. In the second (clinical outcomes) use case, we applied OBCS to represent the processing of electronic health care data to determine the associations between hospital staffing levels and patient mortality. Our case studies were designed to show how OBCS can be used for the consistent representation of statistical analysis pipelines under two different research paradigms. Other ongoing projects using OBCS for statistical data processing are also discussed. The OBCS source code and documentation are available at: https://github.com/obcs/obcs . The Ontology of Biological and Clinical Statistics (OBCS) is a community-based open source ontology in the domain of biological and clinical statistics. OBCS is a timely ontology that represents statistics-related terms and their relations in a rigorous fashion, facilitates standard data analysis and integration, and supports reproducible biological and clinical research.
Visual Data Analysis for Satellites
NASA Technical Reports Server (NTRS)
Lau, Yee; Bhate, Sachin; Fitzpatrick, Patrick
2008-01-01
The Visual Data Analysis Package is a collection of programs and scripts that facilitate visual analysis of data available from NASA and NOAA satellites, as well as dropsonde, buoy, and conventional in-situ observations. The package features utilities for data extraction, data quality control, statistical analysis, and data visualization. The Hierarchical Data Format (HDF) satellite data extraction routines from NASA's Jet Propulsion Laboratory were customized for specific spatial coverage and file input/output. Statistical analysis includes the calculation of the relative error, the absolute error, and the root mean square error. Other capabilities include curve fitting through the data points to fill in missing data points between satellite passes or where clouds obscure satellite data. For data visualization, the software provides customizable Generic Mapping Tool (GMT) scripts to generate difference maps, scatter plots, line plots, vector plots, histograms, timeseries, and color fill images.
ERIC Educational Resources Information Center
Glancy, Aran W.; Moore, Tamara J.; Guzey, Selcen; Smith, Karl A.
2017-01-01
An understanding of statistics and skills in data analysis are becoming more and more essential, yet research consistently shows that students struggle with these concepts at all levels. This case study documents some of the struggles four groups of fifth-grade students encounter as they collect, organize, and interpret data and then ultimately…
A Secondary Analysis of the Impact of School Management Practices on School Performance
ERIC Educational Resources Information Center
Talbert, Dale A.
2009-01-01
The purpose of this study was to conduct a secondary analysis of the impact of school management practices on school performance utilizing a survey design of School and Staffing (SASS) data collected by the National Center for Education Statistics (NCES) of the U.S. Department of Education, 1999-2000. The study identifies those school management…
Vermont's use-value appraisal property tax program: a forest inventory and analysis
Paul E. Sendak; Donald F. Dennis; Donald F. Dennis
1989-01-01
A statistical report and analysis of the timberland enrolled in the Vermont Use Value Appraisal (UVA) property tax program. The study was conducted using data collected in the fourth forest survey of Vermont (1983). Estimates are presented on land area, timber volumes, tree quality, numbers of live trees, and biomass for timberland enrolled in the UVA program and for...
Stanzel, Sven; Weimer, Marc; Kopp-Schneider, Annette
2013-06-01
High-throughput screening approaches are carried out for the toxicity assessment of a large number of chemical compounds. In such large-scale in vitro toxicity studies several hundred or thousand concentration-response experiments are conducted. The automated evaluation of concentration-response data using statistical analysis scripts saves time and yields more consistent results in comparison to data analysis performed by the use of menu-driven statistical software. Automated statistical analysis requires that concentration-response data are available in a standardised data format across all compounds. To obtain consistent data formats, a standardised data management workflow must be established, including guidelines for data storage, data handling and data extraction. In this paper two procedures for data management within large-scale toxicological projects are proposed. Both procedures are based on Microsoft Excel files as the researcher's primary data format and use a computer programme to automate the handling of data files. The first procedure assumes that data collection has not yet started whereas the second procedure can be used when data files already exist. Successful implementation of the two approaches into the European project ACuteTox is illustrated. Copyright © 2012 Elsevier Ltd. All rights reserved.
Web-based data collection: detailed methods of a questionnaire and data gathering tool
Cooper, Charles J; Cooper, Sharon P; del Junco, Deborah J; Shipp, Eva M; Whitworth, Ryan; Cooper, Sara R
2006-01-01
There have been dramatic advances in the development of web-based data collection instruments. This paper outlines a systematic web-based approach to facilitate this process through locally developed code and to describe the results of using this process after two years of data collection. We provide a detailed example of a web-based method that we developed for a study in Starr County, Texas, assessing high school students' work and health status. This web-based application includes data instrument design, data entry and management, and data tables needed to store the results that attempt to maximize the advantages of this data collection method. The software also efficiently produces a coding manual, web-based statistical summary and crosstab reports, as well as input templates for use by statistical packages. Overall, web-based data entry using a dynamic approach proved to be a very efficient and effective data collection system. This data collection method expedited data processing and analysis and eliminated the need for cumbersome and expensive transfer and tracking of forms, data entry, and verification. The code has been made available for non-profit use only to the public health research community as a free download [1]. PMID:16390556
ERIC Educational Resources Information Center
Bernstein, Michael I.
1982-01-01
Steps a school board can take to minimize the risk of age discrimination suits include reviewing all written policies, forms, files, and collective bargaining agreements for age discriminatory items; preparing a detailed statistical analysis of the age of personnel; and reviewing reduction-in-force procedures. (Author/MLF)
Merolla, Giovanni; Tartarone, Antonio; Porcellini, Giuseppe
2016-01-01
Objectives: To obtain outcomes data on anatomical and reverse total shoulder arthroplasty by analysis of clinical scores and standard radiographs. Subject selection and enrollment: 400 consecutive series of patients replaced with anatomical and reverse total shoulder arthroplasty (minimum 3 years follow-up). Study Design: retrospective monocenter. Preoperative assessment: Demographics, clinical scores (Constant-Murley) as available, shoulder X-ray (AP, outlet and axillary views) . Last follow-up: Postoperative radiographhs and clinical scores. Adverse events and complications to be reported as occurred since implantation. Statistical analysis: Data collected will be summarized and analyzed for statistical significance. PMID:27326389
Texture as a basis for acoustic classification of substrate in the nearshore region
NASA Astrophysics Data System (ADS)
Dennison, A.; Wattrus, N. J.
2016-12-01
Segmentation and classification of substrate type from two locations in Lake Superior, are predicted using multivariate statistical processing of textural measures derived from shallow-water, high-resolution multibeam bathymetric data. During a multibeam sonar survey, both bathymetric and backscatter data are collected. It is well documented that the statistical characteristic of a sonar backscatter mosaic is dependent on substrate type. While classifying the bottom-type on the basis on backscatter alone can accurately predict and map bottom-type, it lacks the ability to resolve and capture fine textural details, an important factor in many habitat mapping studies. Statistical processing can capture the pertinent details about the bottom-type that are rich in textural information. Further multivariate statistical processing can then isolate characteristic features, and provide the basis for an accurate classification scheme. Preliminary results from an analysis of bathymetric data and ground-truth samples collected from the Amnicon River, Superior, Wisconsin, and the Lester River, Duluth, Minnesota, demonstrate the ability to process and develop a novel classification scheme of the bottom type in two geomorphologically distinct areas.
Statistical Sources for Health Science Librarians.
ERIC Educational Resources Information Center
Weise, Frieda
This continuing education course syllabus presents information on the collection of vital and health statistics, lists of agencies or organizations involved in statistical collection and/or dissemination, annotated bibliographies of statistical sources, and guidelines for accessing statistical information. Topics covered include: (1) the reporting…
How Big of a Problem is Analytic Error in Secondary Analyses of Survey Data?
West, Brady T; Sakshaug, Joseph W; Aurelien, Guy Alain S
2016-01-01
Secondary analyses of survey data collected from large probability samples of persons or establishments further scientific progress in many fields. The complex design features of these samples improve data collection efficiency, but also require analysts to account for these features when conducting analysis. Unfortunately, many secondary analysts from fields outside of statistics, biostatistics, and survey methodology do not have adequate training in this area, and as a result may apply incorrect statistical methods when analyzing these survey data sets. This in turn could lead to the publication of incorrect inferences based on the survey data that effectively negate the resources dedicated to these surveys. In this article, we build on the results of a preliminary meta-analysis of 100 peer-reviewed journal articles presenting analyses of data from a variety of national health surveys, which suggested that analytic errors may be extremely prevalent in these types of investigations. We first perform a meta-analysis of a stratified random sample of 145 additional research products analyzing survey data from the Scientists and Engineers Statistical Data System (SESTAT), which describes features of the U.S. Science and Engineering workforce, and examine trends in the prevalence of analytic error across the decades used to stratify the sample. We once again find that analytic errors appear to be quite prevalent in these studies. Next, we present several example analyses of real SESTAT data, and demonstrate that a failure to perform these analyses correctly can result in substantially biased estimates with standard errors that do not adequately reflect complex sample design features. Collectively, the results of this investigation suggest that reviewers of this type of research need to pay much closer attention to the analytic methods employed by researchers attempting to publish or present secondary analyses of survey data.
How Big of a Problem is Analytic Error in Secondary Analyses of Survey Data?
West, Brady T.; Sakshaug, Joseph W.; Aurelien, Guy Alain S.
2016-01-01
Secondary analyses of survey data collected from large probability samples of persons or establishments further scientific progress in many fields. The complex design features of these samples improve data collection efficiency, but also require analysts to account for these features when conducting analysis. Unfortunately, many secondary analysts from fields outside of statistics, biostatistics, and survey methodology do not have adequate training in this area, and as a result may apply incorrect statistical methods when analyzing these survey data sets. This in turn could lead to the publication of incorrect inferences based on the survey data that effectively negate the resources dedicated to these surveys. In this article, we build on the results of a preliminary meta-analysis of 100 peer-reviewed journal articles presenting analyses of data from a variety of national health surveys, which suggested that analytic errors may be extremely prevalent in these types of investigations. We first perform a meta-analysis of a stratified random sample of 145 additional research products analyzing survey data from the Scientists and Engineers Statistical Data System (SESTAT), which describes features of the U.S. Science and Engineering workforce, and examine trends in the prevalence of analytic error across the decades used to stratify the sample. We once again find that analytic errors appear to be quite prevalent in these studies. Next, we present several example analyses of real SESTAT data, and demonstrate that a failure to perform these analyses correctly can result in substantially biased estimates with standard errors that do not adequately reflect complex sample design features. Collectively, the results of this investigation suggest that reviewers of this type of research need to pay much closer attention to the analytic methods employed by researchers attempting to publish or present secondary analyses of survey data. PMID:27355817
A New Femtosecond Laser-Based Three-Dimensional Tomography Technique
NASA Astrophysics Data System (ADS)
Echlin, McLean P.
2011-12-01
Tomographic imaging has dramatically changed science, most notably in the fields of medicine and biology, by producing 3D views of structures which are too complex to understand in any other way. Current tomographic techniques require extensive time both for post-processing and data collection. Femtosecond laser based tomographic techniques have been developed in both standard atmosphere (femtosecond laser-based serial sectioning technique - FSLSS) and in vacuum (Tri-Beam System) for the fast collection (10 5mum3/s) of mm3 sized 3D datasets. Both techniques use femtosecond laser pulses to selectively remove layer-by-layer areas of material with low collateral damage and a negligible heat affected zone. To the authors knowledge, femtosecond lasers have never been used to serial section and these techniques have been entirely and uniquely developed by the author and his collaborators at the University of Michigan and University of California Santa Barbara. The FSLSS was applied to measure the 3D distribution of TiN particles in a 4330 steel. Single pulse ablation morphologies and rates were measured and collected from literature. Simultaneous two-phase ablation of TiN and steel matrix was shown to occur at fluences of 0.9-2 J/cm2. Laser scanning protocols were developed minimizing surface roughness to 0.1-0.4 mum for laser-based sectioning. The FSLSS technique was used to section and 3D reconstruct titanium nitride (TiN) containing 4330 steel. Statistical analysis of 3D TiN particle sizes, distribution parameters, and particle density were measured. A methodology was developed to use the 3D datasets to produce statistical volume elements (SVEs) for toughness modeling. Six FSLSS TiN datasets were sub-sampled into 48 SVEs for statistical analysis and toughness modeling using the Rice-Tracey and Garrison-Moody models. A two-parameter Weibull analysis was performed and variability in the toughness data agreed well with Ruggieri et al. bulk toughness measurements. The Tri-Beam system combines the benefits of laser based material removal (speed, low-damage, automated) with detectors that collect chemical, structural, and topological information. Multi-modal sectioning information was collected after many laser scanning passes demonstrating the capability of the Tri-Beam system.
Statistical Characterization of School Bus Drive Cycles Collected via Onboard Logging Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Duran, A.; Walkowicz, K.
In an effort to characterize the dynamics typical of school bus operation, National Renewable Energy Laboratory (NREL) researchers set out to gather in-use duty cycle data from school bus fleets operating across the country. Employing a combination of Isaac Instruments GPS/CAN data loggers in conjunction with existing onboard telemetric systems resulted in the capture of operating information for more than 200 individual vehicles in three geographically unique domestic locations. In total, over 1,500 individual operational route shifts from Washington, New York, and Colorado were collected. Upon completing the collection of in-use field data using either NREL-installed data acquisition devices ormore » existing onboard telemetry systems, large-scale duty-cycle statistical analyses were performed to examine underlying vehicle dynamics trends within the data and to explore vehicle operation variations between fleet locations. Based on the results of these analyses, high, low, and average vehicle dynamics requirements were determined, resulting in the selection of representative standard chassis dynamometer test cycles for each condition. In this paper, the methodology and accompanying results of the large-scale duty-cycle statistical analysis are presented, including graphical and tabular representations of a number of relationships between key duty-cycle metrics observed within the larger data set. In addition to presenting the results of this analysis, conclusions are drawn and presented regarding potential applications of advanced vehicle technology as it relates specifically to school buses.« less
Hage, Olle; Söderholm, Patrik
2008-01-01
The Swedish producer responsibility ordinance mandates producers to collect and recycle packaging materials. This paper investigates the main determinants of collection rates of household plastic packaging waste in Swedish municipalities. This is done by the use of a regression analysis based on cross-sectional data for 252 Swedish municipalities. The results suggest that local policies, geographic/demographic variables, socio-economic factors and environmental preferences all help explain inter-municipality collection rates. For instance, the collection rate appears to be positively affected by increases in the unemployment rate, the share of private houses, and the presence of immigrants (unless newly arrived) in the municipality. The impacts of distance to recycling industry, urbanization rate and population density on collection outcomes turn out, though, to be both statistically and economically insignificant. A reasonable explanation for this is that the monetary compensation from the material companies to the collection entrepreneurs vary depending on region and is typically higher in high-cost regions. This implies that the plastic packaging collection in Sweden may be cost ineffective. Finally, the analysis also shows that municipalities that employ weight-based waste management fees generally experience higher collection rates than those municipalities in which flat and/or volume-based fees are used.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hage, Olle; Soederholm, Patrik
2008-07-01
The Swedish producer responsibility ordinance mandates producers to collect and recycle packaging materials. This paper investigates the main determinants of collection rates of household plastic packaging waste in Swedish municipalities. This is done by the use of a regression analysis based on cross-sectional data for 252 Swedish municipalities. The results suggest that local policies, geographic/demographic variables, socio-economic factors and environmental preferences all help explain inter-municipality collection rates. For instance, the collection rate appears to be positively affected by increases in the unemployment rate, the share of private houses, and the presence of immigrants (unless newly arrived) in the municipality. Themore » impacts of distance to recycling industry, urbanization rate and population density on collection outcomes turn out, though, to be both statistically and economically insignificant. A reasonable explanation for this is that the monetary compensation from the material companies to the collection entrepreneurs vary depending on region and is typically higher in high-cost regions. This implies that the plastic packaging collection in Sweden may be cost ineffective. Finally, the analysis also shows that municipalities that employ weight-based waste management fees generally experience higher collection rates than those municipalities in which flat and/or volume-based fees are used.« less
Potential of IMU Sensors in Performance Analysis of Professional Alpine Skiers
Yu, Gwangjae; Jang, Young Jae; Kim, Jinhyeok; Kim, Jin Hae; Kim, Hye Young; Kim, Kitae; Panday, Siddhartha Bikram
2016-01-01
In this paper, we present an analysis to identify a sensor location for an inertial measurement unit (IMU) on the body of a skier and propose the best location to capture turn motions for training. We also validate the manner in which the data from the IMU sensor on the proposed location can characterize ski turns and performance with a series of statistical analyses, including a comparison with data collected from foot pressure sensors. The goal of the study is to logically identify the ideal location on the skier’s body to attach the IMU sensor and the best use of the data collected for the skier. The statistical analyses and the hierarchical clustering method indicate that the pelvis is the best location for attachment of an IMU, and numerical validation shows that the data collected from this location can effectively estimate the performance and characteristics of the skier. Moreover, placement of the sensor at this location does not distract the skier’s motion, and the sensor can be easily attached and detached. The findings of this study can be used for the development of a wearable device for the routine training of professional skiers. PMID:27043579
NASA Astrophysics Data System (ADS)
Lee, An-Sheng; Lu, Wei-Li; Huang, Jyh-Jaan; Chang, Queenie; Wei, Kuo-Yen; Lin, Chin-Jung; Liou, Sofia Ya Hsuan
2016-04-01
Through the geology and climate characteristic in Taiwan, generally rivers carry a lot of suspended particles. After these particles settled, they become sediments which are good sorbent for heavy metals in river system. Consequently, sediments can be found recording contamination footprint at low flow energy region, such as estuary. Seven sediment cores were collected along Nankan River, northern Taiwan, which is seriously contaminated by factory, household and agriculture input. Physico-chemical properties of these cores were derived from Itrax-XRF Core Scanner and grain size analysis. In order to interpret these complex data matrices, the multivariate statistical techniques (cluster analysis, factor analysis and discriminant analysis) were introduced to this study. Through the statistical determination, the result indicates four types of sediment. One of them represents contamination event which shows high concentration of Cu, Zn, Pb, Ni and Fe, and low concentration of Si and Zr. Furthermore, three possible contamination sources of this type of sediment were revealed by Factor Analysis. The combination of sediment analysis and multivariate statistical techniques used provides new insights into the contamination depositional history of Nankan River and could be similarly applied to other river systems to determine the scale of anthropogenic contamination.
Logical analysis of diffuse large B-cell lymphomas.
Alexe, G; Alexe, S; Axelrod, D E; Hammer, P L; Weissmann, D
2005-07-01
The goal of this study is to re-examine the oligonucleotide microarray dataset of Shipp et al., which contains the intensity levels of 6817 genes of 58 patients with diffuse large B-cell lymphoma (DLBCL) and 19 with follicular lymphoma (FL), by means of the combinatorics, optimisation, and logic-based methodology of logical analysis of data (LAD). The motivations for this new analysis included the previously demonstrated capabilities of LAD and its expected potential (1) to identify different informative genes than those discovered by conventional statistical methods, (2) to identify combinations of gene expression levels capable of characterizing different types of lymphoma, and (3) to assemble collections of such combinations that if considered jointly are capable of accurately distinguishing different types of lymphoma. The central concept of LAD is a pattern or combinatorial biomarker, a concept that resembles a rule as used in decision tree methods. LAD is able to exhaustively generate the collection of all those patterns which satisfy certain quality constraints, through a systematic combinatorial process guided by clear optimization criteria. Then, based on a set covering approach, LAD aggregates the collection of patterns into classification models. In addition, LAD is able to use the information provided by large collections of patterns in order to extract subsets of variables, which collectively are able to distinguish between different types of disease. For the differential diagnosis of DLBCL versus FL, a model based on eight significant genes is constructed and shown to have a sensitivity of 94.7% and a specificity of 100% on the test set. For the prognosis of good versus poor outcome among the DLBCL patients, a model is constructed on another set consisting also of eight significant genes, and shown to have a sensitivity of 87.5% and a specificity of 90% on the test set. The genes selected by LAD also work well as a basis for other kinds of statistical analysis, indicating their robustness. These two models exhibit accuracies that compare favorably to those in the original study. In addition, the current study also provides a ranking by importance of the genes in the selected significant subsets as well as a library of dozens of combinatorial biomarkers (i.e. pairs or triplets of genes) that can serve as a source of mathematically generated, statistically significant research hypotheses in need of biological explanation.
Dale D. Gormanson; Scott A. Pugh; Charles J. Barnett; Patrick D. Miles; Randall S. Morin; Paul A. Sowers; James A. Westfall
2018-01-01
The U.S. Forest Service Forest Inventory and Analysis (FIA) program collects sample plot data on all forest ownerships across the United States. FIAâs primary objective is to determine the extent, condition, volume, growth, and use of trees on the Nationâs forest land through a comprehensive inventory and analysis of the Nationâs forest resources. The FIA program...
Protein Multiplexed Immunoassay Analysis with R.
Breen, Edmond J
2017-01-01
Plasma samples from 177 control and type 2 diabetes patients collected at three Australian hospitals are screened for 14 analytes using six custom-made multiplex kits across 60 96-well plates. In total 354 samples were collected from the patients, representing one baseline and one end point sample from each patient. R methods and source code for analyzing the analyte fluorescence response obtained from these samples by Luminex Bio-Plex ® xMap multiplexed immunoassay technology are disclosed. Techniques and R procedures for reading Bio-Plex ® result files for statistical analysis and data visualization are also presented. The need for technical replicates and the number of technical replicates are addressed as well as plate layout design strategies. Multinomial regression is used to determine plate to sample covariate balance. Methods for matching clinical covariate information to Bio-Plex ® results and vice versa are given. As well as methods for measuring and inspecting the quality of the fluorescence responses are presented. Both fixed and mixed-effect approaches for immunoassay statistical differential analysis are presented and discussed. A random effect approach to outlier analysis and detection is also shown. The bioinformatics R methodology present here provides a foundation for rigorous and reproducible analysis of the fluorescence response obtained from multiplexed immunoassays.
Symposium Issue on the Energy Information Administration.
ERIC Educational Resources Information Center
Kent, Calvin A.; And Others
1993-01-01
Describes the Energy Information Administration (EIA), a statistical agency which provides credible, timely, and useful energy information for decision makers in all sectors of society. The 10 articles included in the volume cover survey design, data collection, data integration, data analysis, modeling and forecasting, confidentiality, and…
ToxMiner Software Interface for Visualizing and Analyzing ToxCast Data
The ToxCast dataset represents a collection of assays and endpoints that will require both standard statistical approaches as well as customized data analysis workflows. To analyze this unique dataset, we have developed an integrated database with Javabased interface called ToxMi...
Development of a Relay Performance Web Tool for the Mars Network
NASA Technical Reports Server (NTRS)
Allard, Daniel A.; Edwards, Charles D.
2009-01-01
Modern Mars surface missions rely upon orbiting spacecraft to relay communications to and from Earth systems. An important component of this multi-mission relay process is the collection of relay performance statistics supporting strategic trend analysis and tactical anomaly identification and tracking.
Analyzing Mixed-Dyadic Data Using Structural Equation Models
ERIC Educational Resources Information Center
Peugh, James L.; DiLillo, David; Panuzio, Jillian
2013-01-01
Mixed-dyadic data, collected from distinguishable (nonexchangeable) or indistinguishable (exchangeable) dyads, require statistical analysis techniques that model the variation within dyads and between dyads appropriately. The purpose of this article is to provide a tutorial for performing structural equation modeling analyses of cross-sectional…
50 CFR 600.315 - National Standard 2-Scientific Information.
Code of Federal Regulations, 2014 CFR
2014-10-01
...., abundance, environmental, catch statistics, market and trade trends) provide time-series information on... comment should be solicited at appropriate times during the review of scientific information... information or the promise of future data collection or analysis. In some cases, due to time constraints...
Time Series Expression Analyses Using RNA-seq: A Statistical Approach
Oh, Sunghee; Song, Seongho; Grabowski, Gregory; Zhao, Hongyu; Noonan, James P.
2013-01-01
RNA-seq is becoming the de facto standard approach for transcriptome analysis with ever-reducing cost. It has considerable advantages over conventional technologies (microarrays) because it allows for direct identification and quantification of transcripts. Many time series RNA-seq datasets have been collected to study the dynamic regulations of transcripts. However, statistically rigorous and computationally efficient methods are needed to explore the time-dependent changes of gene expression in biological systems. These methods should explicitly account for the dependencies of expression patterns across time points. Here, we discuss several methods that can be applied to model timecourse RNA-seq data, including statistical evolutionary trajectory index (SETI), autoregressive time-lagged regression (AR(1)), and hidden Markov model (HMM) approaches. We use three real datasets and simulation studies to demonstrate the utility of these dynamic methods in temporal analysis. PMID:23586021
Time series expression analyses using RNA-seq: a statistical approach.
Oh, Sunghee; Song, Seongho; Grabowski, Gregory; Zhao, Hongyu; Noonan, James P
2013-01-01
RNA-seq is becoming the de facto standard approach for transcriptome analysis with ever-reducing cost. It has considerable advantages over conventional technologies (microarrays) because it allows for direct identification and quantification of transcripts. Many time series RNA-seq datasets have been collected to study the dynamic regulations of transcripts. However, statistically rigorous and computationally efficient methods are needed to explore the time-dependent changes of gene expression in biological systems. These methods should explicitly account for the dependencies of expression patterns across time points. Here, we discuss several methods that can be applied to model timecourse RNA-seq data, including statistical evolutionary trajectory index (SETI), autoregressive time-lagged regression (AR(1)), and hidden Markov model (HMM) approaches. We use three real datasets and simulation studies to demonstrate the utility of these dynamic methods in temporal analysis.
Tsatsarelis, Thomas; Antonopoulos, Ioannis; Karagiannidis, Avraam; Perkoulidis, George
2007-10-01
This study presents an assessment of the current status of open dumps in Laconia prefecture of Peloponnese in southern Greece, where all open dumps are targeted for closure by 2008. An extensive field survey was conducted in 2005 to register existing sites in the prefecture. The data collected included the site area and age, waste depth, type of disposed waste, distance from nearest populated area, local geographical features and observed practices of open burning and soil coverage. On the basis of the collected data, a GIS database was developed, and the above parameters were statistically analysed. Subsequently, a decision tool for the restoration of open dumps was implemented, which led to the prioritization of site restorations and specific decisions about appropriate restoration steps for each site. The sites requiring restoration were then further classified using Principal Component Analysis, in order to categorize them into groups suitable for similar restoration work, thus facilitating fund allocation and subsequent restoration project management.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-05-26
... Sentenced Population Movement--National Prisoner Statistics, Extension and Revision of Existing Collection...) Title of the Form/Collection: Summary of Sentenced Population Movement--National Prisoner Statistics (3...
An ecological genetic delineation of local seed-source provenance for ecological restoration
Krauss, Siegfried L; Sinclair, Elizabeth A; Bussell, John D; Hobbs, Richard J
2013-01-01
An increasingly important practical application of the analysis of spatial genetic structure within plant species is to help define the extent of local provenance seed collection zones that minimize negative impacts in ecological restoration programs. Here, we derive seed sourcing guidelines from a novel range-wide assessment of spatial genetic structure of 24 populations of Banksia menziesii (Proteaceae), a widely distributed Western Australian tree of significance in local ecological restoration programs. An analysis of molecular variance (AMOVA) of 100 amplified fragment length polymorphism (AFLP) markers revealed significant genetic differentiation among populations (ΦPT = 0.18). Pairwise population genetic dissimilarity was correlated with geographic distance, but not environmental distance derived from 15 climate variables, suggesting overall neutrality of these markers with regard to these climate variables. Nevertheless, Bayesian outlier analysis identified four markers potentially under selection, although these were not correlated with the climate variables. We calculated a global R-statistic using analysis of similarities (ANOSIM) to test the statistical significance of population differentiation and to infer a threshold seed collection zone distance of ∼60 km (all markers) and 100 km (outlier markers) when genetic distance was regressed against geographic distance. Population pairs separated by >60 km were, on average, twice as likely to be significantly genetically differentiated than population pairs separated by <60 km, suggesting that habitat-matched sites within a 30-km radius around a restoration site genetically defines a local provenance seed collection zone for B. menziesii. Our approach is a novel probability-based practical solution for the delineation of a local seed collection zone to minimize negative genetic impacts in ecological restoration. PMID:23919158
Drinking water quality assessment.
Aryal, J; Gautam, B; Sapkota, N
2012-09-01
Drinking water quality is the great public health concern because it is a major risk factor for high incidence of diarrheal diseases in Nepal. In the recent years, the prevalence rate of diarrhoea has been found the highest in Myagdi district. This study was carried out to assess the quality of drinking water from different natural sources, reservoirs and collection taps at Arthunge VDC of Myagdi district. A cross-sectional study was carried out using random sampling method in Arthunge VDC of Myagdi district from January to June,2010. 84 water samples representing natural sources, reservoirs and collection taps from the study area were collected. The physico-chemical and microbiological analysis was performed following standards technique set by APHA 1998 and statistical analysis was carried out using SPSS 11.5. The result was also compared with national and WHO guidelines. Out of 84 water samples (from natural source, reservoirs and tap water) analyzed, drinking water quality parameters (except arsenic and total coliform) of all water samples was found to be within the WHO standards and national standards.15.48% of water samples showed pH (13) higher than the WHO permissible guideline values. Similarly, 85.71% of water samples showed higher Arsenic value (72) than WHO value. Further, the statistical analysis showed no significant difference (P<0.05) of physico-chemical parameters and total coliform count of drinking water for collection taps water samples of winter (January, 2010) and summer (June, 2010). The microbiological examination of water samples revealed the presence of total coliform in 86.90% of water samples. The results obtained from physico-chemical analysis of water samples were within national standard and WHO standards except arsenic. The study also found the coliform contamination to be the key problem with drinking water.
Public and patient involvement in quantitative health research: A statistical perspective.
Hannigan, Ailish
2018-06-19
The majority of studies included in recent reviews of impact for public and patient involvement (PPI) in health research had a qualitative design. PPI in solely quantitative designs is underexplored, particularly its impact on statistical analysis. Statisticians in practice have a long history of working in both consultative (indirect) and collaborative (direct) roles in health research, yet their perspective on PPI in quantitative health research has never been explicitly examined. To explore the potential and challenges of PPI from a statistical perspective at distinct stages of quantitative research, that is sampling, measurement and statistical analysis, distinguishing between indirect and direct PPI. Statistical analysis is underpinned by having a representative sample, and a collaborative or direct approach to PPI may help achieve that by supporting access to and increasing participation of under-represented groups in the population. Acknowledging and valuing the role of lay knowledge of the context in statistical analysis and in deciding what variables to measure may support collective learning and advance scientific understanding, as evidenced by the use of participatory modelling in other disciplines. A recurring issue for quantitative researchers, which reflects quantitative sampling methods, is the selection and required number of PPI contributors, and this requires further methodological development. Direct approaches to PPI in quantitative health research may potentially increase its impact, but the facilitation and partnership skills required may require further training for all stakeholders, including statisticians. © 2018 The Authors Health Expectations published by John Wiley & Sons Ltd.
A nonparametric analysis of plot basal area growth using tree based models
G. L. Gadbury; H. K. lyer; H. T. Schreuder; C. Y. Ueng
1997-01-01
Tree based statistical models can be used to investigate data structure and predict future observations. We used nonparametric and nonlinear models to reexamine the data sets on tree growth used by Bechtold et al. (1991) and Ruark et al. (1991). The growth data were collected by Forest Inventory and Analysis (FIA) teams from 1962 to 1972 (4th cycle) and 1972 to 1982 (...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-12-29
... for OMB Review; Comment Request; Local Area Unemployment Statistics Program ACTION: Notice. SUMMARY... collection request (ICR) titled, ``Local Area Unemployment Statistics Program,'' to the Office of Management... of Collection: Local Area Unemployment Statistics Program. OMB Control Number: 1220-0017. Affected...
Olavarría, Verónica V; Arima, Hisatomi; Anderson, Craig S; Brunser, Alejandro; Muñoz-Venturelli, Paula; Billot, Laurent; Lavados, Pablo M
2017-02-01
Background The HEADPOST Pilot is a proof-of-concept, open, prospective, multicenter, international, cluster randomized, phase IIb controlled trial, with masked outcome assessment. The trial will test if lying flat head position initiated in patients within 12 h of onset of acute ischemic stroke involving the anterior circulation increases cerebral blood flow in the middle cerebral arteries, as measured by transcranial Doppler. The study will also assess the safety and feasibility of patients lying flat for ≥24 h. The trial was conducted in centers in three countries, with ability to perform early transcranial Doppler. A feature of this trial was that patients were randomized to a certain position according to the month of admission to hospital. Objective To outline in detail the predetermined statistical analysis plan for HEADPOST Pilot study. Methods All data collected by participating researchers will be reviewed and formally assessed. Information pertaining to the baseline characteristics of patients, their process of care, and the delivery of treatments will be classified, and for each item, appropriate descriptive statistical analyses are planned with comparisons made between randomized groups. For the outcomes, statistical comparisons to be made between groups are planned and described. Results This statistical analysis plan was developed for the analysis of the results of the HEADPOST Pilot study to be transparent, available, verifiable, and predetermined before data lock. Conclusions We have developed a statistical analysis plan for the HEADPOST Pilot study which is to be followed to avoid analysis bias arising from prior knowledge of the study findings. Trial registration The study is registered under HEADPOST-Pilot, ClinicalTrials.gov Identifier NCT01706094.
NASA Astrophysics Data System (ADS)
Chakravarty, T.; Chowdhury, A.; Ghose, A.; Bhaumik, C.; Balamuralidhar, P.
2014-03-01
Telematics form an important technology enabler for intelligent transportation systems. By deploying on-board diagnostic devices, the signatures of vehicle vibration along with its location and time are recorded. Detailed analyses of the collected signatures offer deep insights into the state of the objects under study. Towards that objective, we carried out experiments by deploying telematics device in one of the office bus that ferries employees to office and back. Data is being collected from 3-axis accelerometer, GPS, speed and the time for all the journeys. In this paper, we present initial results of the above exercise by applying statistical methods to derive information through systematic analysis of the data collected over four months. It is demonstrated that the higher order derivative of the measured Z axis acceleration samples display the properties Weibull distribution when the time axis is replaced by the amplitude of such processed acceleration data. Such an observation offers us a method to predict future behaviour where deviations from prediction are classified as context-based aberrations or progressive degradation of the system. In addition we capture the relationship between speed of the vehicle and median of the jerk energy samples using regression analysis. Such results offer an opportunity to develop a robust method to model road-vehicle interaction thereby enabling us to predict such like driving behaviour and condition based maintenance etc.
The contribution of collective attack tactics in differentiating handball score efficiency.
Rogulj, Nenad; Srhoj, Vatromir; Srhoj, Ljerka
2004-12-01
The prevalence of 19 elements of collective tactics in score efficient and score inefficient teams was analyzed in 90 First Croatian Handball League--Men games during the 1998-1999 season. Prediction variables were used to describe duration, continuity, system, organization and spatial direction of attacks. Analysis of the basic descriptive and distribution statistical parameters revealed normal distribution of all variables and possibility to use multivariate methods. Canonic discrimination analysis and analysis of variance showed the use of collective tactics elements on attacks to differ statistically significantly between the winning and losing teams. Counter-attacks and uninterrupted attacks predominate in winning teams. Other types of attacks such as long position attack, multiply interrupted attack, attack with one circle runner attack player/pivot, attack based on basic principles, attack based on group cooperation, attack based on independent action, attack based on group maneuvering, rightward directed attack and leftward directed attack predominate in losing teams. Winning teams were found to be clearly characterized by quick attacks against unorganized defense, whereas prolonged, interrupted position attacks against organized defense along with frequent and diverse tactical actions were characteristic of losing teams. The choice and frequency of using a particular tactical activity in position attack do not warrant score efficiency but usually are consequential to the limited anthropologic potential and low level of individual technical-tactical skills of the players in low-quality teams.
Identification of the isomers using principal component analysis (PCA) method
NASA Astrophysics Data System (ADS)
Kepceoǧlu, Abdullah; Gündoǧdu, Yasemin; Ledingham, Kenneth William David; Kilic, Hamdi Sukur
2016-03-01
In this work, we have carried out a detailed statistical analysis for experimental data of mass spectra from xylene isomers. Principle Component Analysis (PCA) was used to identify the isomers which cannot be distinguished using conventional statistical methods for interpretation of their mass spectra. Experiments have been carried out using a linear TOF-MS coupled to a femtosecond laser system as an energy source for the ionisation processes. We have performed experiments and collected data which has been analysed and interpreted using PCA as a multivariate analysis of these spectra. This demonstrates the strength of the method to get an insight for distinguishing the isomers which cannot be identified using conventional mass analysis obtained through dissociative ionisation processes on these molecules. The PCA results dependending on the laser pulse energy and the background pressure in the spectrometers have been presented in this work.
DOT National Transportation Integrated Search
1976-09-01
Standardized injury rates and seat belt effectiveness measures are derived from a probability sample of towaway accidents involving 1973-1975 model cars. The data were collected in five different geographic regions. Weighted sample size available for...
NASA Technical Reports Server (NTRS)
Marchenko, V. I.
1974-01-01
During periods of high solar activity fibrinolysis and fibrinogenolysis are increased. A direct correlative relationship is established between the indices of fibrinolysis, fibrinogenolysis and solar flares which were recorded two days before the blood was collected for analysis.
Access and Ownership in the Academic Environment: One Library's Progress Report.
ERIC Educational Resources Information Center
Brin, Beth; Cochran, Elissa
1994-01-01
Describes the methodology used at the University of Arizona Library to address the issue of access versus ownership of library materials. Topics discussed include participatory management; data collection, including focus groups, interlibrary loan statistics, and graduate research citation analysis; and resulting recommendations, including…
The Stratification Analysis of Sediment Data for Lake Michigan
This research paper describes the development of spatial statistical tools that are applied to investigate the spatial trends of sediment data sets for nutrients and carbon in Lake Michigan. All of the sediment data utilized in the present study was collected over a two year per...
75 FR 2888 - Agency Information Collection Activities: Proposed Collection; Comments Requested
Federal Register 2010, 2011, 2012, 2013, 2014
2010-01-19
... has expired, State Court Processing Statistics 2009. The Department of Justice, Office of Justice... which OMB approval has expired, State Court Processing Statistics, 2009. (2) The title of the form/collection: State Court Processing Statistics, 2009. (3) The agency form number, if any, and the applicable...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-12-29
... Statistics (BLS). Title of Collection: Cognitive and Psychological Research. OMB Control Number: 1220-0141... Department of Labor (DOL) is submitting the Bureau of Labor Statistics (BLS) sponsored information collection... Statistics (BLS), Office of Management and Budget, Room 10235, Washington, DC 20503, Telephone: (202) 395...
Concentrations of MUC16 and MUC5AC using three tear collection methods
Nichols, Jason J.
2017-01-01
Purpose To determine the optimal tear collection method for analysis of ocular surface mucins MUC5AC and MUC16. Methods Fifteen subjects without ocular surface disease were recruited. Subjects presented for tear collection on three separate days for three different tear collection methods with the order of method randomized. Methods used to collect tears from right and left eyes included Schirmer’s strip, basal tear collection, and flush tear collection. All samples from the right eyes were individually analyzed for MUC5AC whereas the left eye samples were individually analyzed for MUC16. For each individual sample, 10 μg of protein were loaded per lane into a 1% (w/v) agarose gel and run in electrophoresis buffer for 2 h. After overnight capillary transfer, membranes were incubated with either MUC5AC antibody CLH2 or MUC16 antibody OC125 for western blot analysis. Blots were developed with enhanced chemiluminescence (ECL) and signals captured with the Odyssey Fc (LI-COR). The relative amounts of MUC5AC and MUC16 were quantified with densitometry using software and compared for statistically significant differences between tear collection methods using the Kruskal–Wallis test in SPSS 22 and GraphPad Prism 7.02. Dunn’s multiple comparisons test was used for pairwise post-hoc comparisons. Results Samples containing less than 10 μg of total protein were not used for analysis which left eight samples (out of 45) unusable. The calculated MUC5AC median signal intensities from Schirmer’s strip, basal tears, and flush tears were 2.86 (n = 15, the interquartile range [IQR] = 2.54–3.21), 1.65 (n = 14, IQR = 1.34–3.1), and 1.67 (n = 9, IQR = 1.42–1.72), respectively (H = 9.5, p = 0.009). Post-hoc pairwise comparisons showed a statistically significant difference between Schirmer’s strip and flush tears (p = 0.01). The calculated MUC16 median signal intensities from Schirmer’s strip, basal tears, and flush tears were 1.88 (n = 14, IQR = 1.43–2.61), 5.24 (n = 15, IQR = 4.16–6.21), and 2.45 (n = 7, IQR = 1.85–2.48), respectively (H = 18.1, p = 0.001). Post-hoc pairwise comparison showed statistically significant differences between basal tears and Schirmer’s strip (p = 0.0003) and between basal tears and flush tears (p = 0.006). Conclusions MUC5AC and MUC16 are present in human tear fluid and can be captured using various tear collection methods. Although basal tear collection yielded the highest relative concentration of MUC16, Schirmer’s strip tear collection yielded the highest MUC5AC concentration. Therefore, the tear collection method chosen depends on the mucin of interest. PMID:28761326
Silvestri, Erin E; Yund, Cynthia; Taft, Sarah; Bowling, Charlena Yoder; Chappie, Daniel; Garrahan, Kevin; Brady-Roberts, Eletha; Stone, Harry; Nichols, Tonya L
2017-01-01
In the event of an indoor release of an environmentally persistent microbial pathogen such as Bacillus anthracis, the potential for human exposure will be considered when remedial decisions are made. Microbial site characterization and clearance sampling data collected in the field might be used to estimate exposure. However, there are many challenges associated with estimating environmental concentrations of B. anthracis or other spore-forming organisms after such an event before being able to estimate exposure. These challenges include: (1) collecting environmental field samples that are adequate for the intended purpose, (2) conducting laboratory analyses and selecting the reporting format needed for the laboratory data, and (3) analyzing and interpreting the data using appropriate statistical techniques. This paper summarizes some key challenges faced in collecting, analyzing, and interpreting microbial field data from a contaminated site. Although the paper was written with considerations for B. anthracis contamination, it may also be applicable to other bacterial agents. It explores the implications and limitations of using field data for determining environmental concentrations both before and after decontamination. Several findings were of interest. First, to date, the only validated surface/sampling device combinations are swabs and sponge-sticks on stainless steel surfaces, thus limiting availability of quantitative analytical results which could be used for statistical analysis. Second, agreement needs to be reached with the analytical laboratory on the definition of the countable range and on reporting of data below the limit of quantitation. Finally, the distribution of the microbial field data and statistical methods needed for a particular data set could vary depending on these data that were collected, and guidance is needed on appropriate statistical software for handling microbial data. Further, research is needed to develop better methods to estimate human exposure from pathogens using environmental data collected from a field setting. PMID:26883476
NASA Astrophysics Data System (ADS)
Wright, Robyn; Thornberg, Steven M.
SEDIDAT is a series of compiled IBM-BASIC (version 2.0) programs that direct the collection, statistical calculation, and graphic presentation of particle settling velocity and equivalent spherical diameter for samples analyzed using the settling tube technique. The programs follow a menu-driven format that is understood easily by students and scientists with little previous computer experience. Settling velocity is measured directly (cm,sec) and also converted into Chi units. Equivalent spherical diameter (reported in Phi units) is calculated using a modified Gibbs equation for different particle densities. Input parameters, such as water temperature, settling distance, particle density, run time, and Phi;Chi interval are changed easily at operator discretion. Optional output to a dot-matrix printer includes a summary of moment and graphic statistical parameters, a tabulation of individual and cumulative weight percents, a listing of major distribution modes, and cumulative and histogram plots of a raw time, settling velocity. Chi and Phi data.
Varlamis, Iraklis; Apostolakis, Ioannis; Sifaki-Pistolla, Dimitra; Dey, Nilanjan; Georgoulias, Vassilios; Lionis, Christos
2017-07-01
Micro or macro-level mapping of cancer statistics is a challenging task that requires long-term planning, prospective studies and continuous monitoring of all cancer cases. The objective of the current study is to present how cancer registry data could be processed using data mining techniques in order to improve the statistical analysis outcomes. Data were collected from the Cancer Registry of Crete in Greece (counties of Rethymno and Lasithi) for the period 1998-2004. Data collection was performed on paper forms and manually transcribed to a single data file, thus introducing errors and noise (e.g. missing and erroneous values, duplicate entries etc.). Data were pre-processed and prepared for analysis using data mining tools and algorithms. Feature selection was applied to evaluate the contribution of each collected feature in predicting patients' survival. Several classifiers were trained and evaluated for their ability to predict survival of patients. Finally, statistical analysis of cancer morbidity and mortality rates in the two regions was performed in order to validate the initial findings. Several critical points in the process of data collection, preprocessing and analysis of cancer data were derived from the results, while a road-map for future population data studies was developed. In addition, increased morbidity rates were observed in the counties of Crete (Age Standardized Morbidity/Incidence Rates ASIR= 396.45 ± 2.89 and 274.77 ±2.48 for men and women, respectively) compared to European and world averages (ASIR= 281.6 and 207.3 for men and women in Europe and 203.8 and 165.1 in world level). Significant variation in cancer types between sexes and age groups (the ratio between deaths and reported cases for young patients, less than 34 years old, is at 0.055 when the respective ratio for patients over 75 years old is 0.366) was also observed. This study introduced a methodology for preprocessing and analyzing cancer data, using a combination of data mining techniques that could be a useful tool for other researchers and further enhancement of the cancer registries. Copyright © 2017 Elsevier B.V. All rights reserved.
Gillespie, Paddy; O'Shea, Eamon; Smith, Susan M; Cupples, Margaret E; Murphy, Andrew W
2016-12-01
Data on health care utilization may be collected using a variety of mechanisms within research studies, each of which may have implications for cost and cost effectiveness. The aim of this observational study is to compare data collected from medical records searches and self-report questionnaires for the cost analysis of a cardiac secondary prevention intervention. Secondary data analysis of the Secondary Prevention of Heart Disease in General Practice (SPHERE) randomized controlled trial (RCT). Resource use data for a range of health care services were collected by research nurse searches of medical records and self-report questionnaires and costs of care estimated for each data collection mechanism. A series of statistical analyses were conducted to compare the mean costs for medical records data versus questionnaire data and to conduct incremental analyses for the intervention and control arms in the trial. Data were available to estimate costs for 95% of patients in the intervention and 96% of patients in the control using the medical records data compared to 65% and 66%, respectively, using the questionnaire data. The incremental analysis revealed a statistically significant difference in mean cost of -€796 (95% CI: -1447, -144; P-value: 0.017) for the intervention relative to the control. This compared to no significant difference in mean cost (95% CI: -1446, 860; P-value: 0.619) for the questionnaire analysis. Our findings illustrate the importance of the choice of health care utilization data collection mechanism for the conduct of economic evaluation alongside randomized trials in primary care. This choice will have implications for the costing methodology employed and potentially, for the cost and cost effectiveness outcomes generated. © The Author 2016. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.
2010-01-01
Background Discrimination between clinical and environmental strains within many bacterial species is currently underexplored. Genomic analyses have clearly shown the enormous variability in genome composition between different strains of a bacterial species. In this study we have used Legionella pneumophila, the causative agent of Legionnaire's disease, to search for genomic markers related to pathogenicity. During a large surveillance study in The Netherlands well-characterized patient-derived strains and environmental strains were collected. We have used a mixed-genome microarray to perform comparative-genome analysis of 257 strains from this collection. Results Microarray analysis indicated that 480 DNA markers (out of in total 3360 markers) showed clear variation in presence between individual strains and these were therefore selected for further analysis. Unsupervised statistical analysis of these markers showed the enormous genomic variation within the species but did not show any correlation with a pathogenic phenotype. We therefore used supervised statistical analysis to identify discriminating markers. Genetic programming was used both to identify predictive markers and to define their interrelationships. A model consisting of five markers was developed that together correctly predicted 100% of the clinical strains and 69% of the environmental strains. Conclusions A novel approach for identifying predictive markers enabling discrimination between clinical and environmental isolates of L. pneumophila is presented. Out of over 3000 possible markers, five were selected that together enabled correct prediction of all the clinical strains included in this study. This novel approach for identifying predictive markers can be applied to all bacterial species, allowing for better discrimination between strains well equipped to cause human disease and relatively harmless strains. PMID:20630115
When decision heuristics and science collide.
Yu, Erica C; Sprenger, Amber M; Thomas, Rick P; Dougherty, Michael R
2014-04-01
The ongoing discussion among scientists about null-hypothesis significance testing and Bayesian data analysis has led to speculation about the practices and consequences of "researcher degrees of freedom." This article advances this debate by asking the broader questions that we, as scientists, should be asking: How do scientists make decisions in the course of doing research, and what is the impact of these decisions on scientific conclusions? We asked practicing scientists to collect data in a simulated research environment, and our findings show that some scientists use data collection heuristics that deviate from prescribed methodology. Monte Carlo simulations show that data collection heuristics based on p values lead to biases in estimated effect sizes and Bayes factors and to increases in both false-positive and false-negative rates, depending on the specific heuristic. We also show that using Bayesian data collection methods does not eliminate these biases. Thus, our study highlights the little appreciated fact that the process of doing science is a behavioral endeavor that can bias statistical description and inference in a manner that transcends adherence to any particular statistical framework.
Kuretzki, Carlos Henrique; Campos, Antônio Carlos Ligocki; Malafaia, Osvaldo; Soares, Sandramara Scandelari Kusano de Paula; Tenório, Sérgio Bernardo; Timi, Jorge Rufino Ribas
2016-03-01
The use of information technology is often applied in healthcare. With regard to scientific research, the SINPE(c) - Integrated Electronic Protocols was created as a tool to support researchers, offering clinical data standardization. By the time, SINPE(c) lacked statistical tests obtained by automatic analysis. Add to SINPE(c) features for automatic realization of the main statistical methods used in medicine . The study was divided into four topics: check the interest of users towards the implementation of the tests; search the frequency of their use in health care; carry out the implementation; and validate the results with researchers and their protocols. It was applied in a group of users of this software in their thesis in the strict sensu master and doctorate degrees in one postgraduate program in surgery. To assess the reliability of the statistics was compared the data obtained both automatically by SINPE(c) as manually held by a professional in statistics with experience with this type of study. There was concern for the use of automatic statistical tests, with good acceptance. The chi-square, Mann-Whitney, Fisher and t-Student were considered as tests frequently used by participants in medical studies. These methods have been implemented and thereafter approved as expected. The incorporation of the automatic SINPE (c) Statistical Analysis was shown to be reliable and equal to the manually done, validating its use as a research tool for medical research.
McLean, Rachael M; Farmer, Victoria L; Nettleton, Alice; Cameron, Claire M; Cook, Nancy R; Campbell, Norman R C
2017-12-01
Food frequency questionnaires (FFQs) are often used to assess dietary sodium intake, although 24-hour urinary excretion is the most accurate measure of intake. The authors conducted a systematic review to investigate whether FFQs are a reliable and valid way of measuring usual dietary sodium intake. Results from 18 studies are described in this review, including 16 validation studies. The methods of study design and analysis varied widely with respect to FFQ instrument, number of 24-hour urine collections collected per participant, methods used to assess completeness of urine collections, and statistical analysis. Overall, there was poor agreement between estimates from FFQ and 24-hour urine. The authors suggest a framework for validation and reporting based on a consensus statement (2004), and recommend that all FFQs used to estimate dietary sodium intake undergo validation against multiple 24-hour urine collections. ©2017 Wiley Periodicals, Inc.
NASA Technical Reports Server (NTRS)
Sung, Q. C.; Miller, L. D.
1977-01-01
Three methods were tested for collection of the training sets needed to establish the spectral signatures of the land uses/land covers sought due to the difficulties of retrospective collection of representative ground control data. Computer preprocessing techniques applied to the digital images to improve the final classification results were geometric corrections, spectral band or image ratioing and statistical cleaning of the representative training sets. A minimal level of statistical verification was made based upon the comparisons between the airphoto estimates and the classification results. The verifications provided a further support to the selection of MSS band 5 and 7. It also indicated that the maximum likelihood ratioing technique can achieve more agreeable classification results with the airphoto estimates than the stepwise discriminant analysis.
Stewart, Sarah; Pearson, Janet; Rome, Keith; Dalbeth, Nicola; Vandal, Alain C
2018-01-01
Statistical techniques currently used in musculoskeletal research often inefficiently account for paired-limb measurements or the relationship between measurements taken from multiple regions within limbs. This study compared three commonly used analysis methods with a mixed-models approach that appropriately accounted for the association between limbs, regions, and trials and that utilised all information available from repeated trials. Four analysis were applied to an existing data set containing plantar pressure data, which was collected for seven masked regions on right and left feet, over three trials, across three participant groups. Methods 1-3 averaged data over trials and analysed right foot data (Method 1), data from a randomly selected foot (Method 2), and averaged right and left foot data (Method 3). Method 4 used all available data in a mixed-effects regression that accounted for repeated measures taken for each foot, foot region and trial. Confidence interval widths for the mean differences between groups for each foot region were used as a criterion for comparison of statistical efficiency. Mean differences in pressure between groups were similar across methods for each foot region, while the confidence interval widths were consistently smaller for Method 4. Method 4 also revealed significant between-group differences that were not detected by Methods 1-3. A mixed effects linear model approach generates improved efficiency and power by producing more precise estimates compared to alternative approaches that discard information in the process of accounting for paired-limb measurements. This approach is recommended in generating more clinically sound and statistically efficient research outputs. Copyright © 2017 Elsevier B.V. All rights reserved.
Harrysson, Iliana J; Cook, Jonathan; Sirimanna, Pramudith; Feldman, Liane S; Darzi, Ara; Aggarwal, Rajesh
2014-07-01
To determine how minimally invasive surgical learning curves are assessed and define an ideal framework for this assessment. Learning curves have implications for training and adoption of new procedures and devices. In 2000, a review of the learning curve literature was done by Ramsay et al and it called for improved reporting and statistical evaluation of learning curves. Since then, a body of literature is emerging on learning curves but the presentation and analysis vary. A systematic search was performed of MEDLINE, EMBASE, ISI Web of Science, ERIC, and the Cochrane Library from 1985 to August 2012. The inclusion criteria are minimally invasive abdominal surgery formally analyzing the learning curve and English language. 592 (11.1%) of the identified studies met the selection criteria. Time is the most commonly used proxy for the learning curve (508, 86%). Intraoperative outcomes were used in 316 (53%) of the articles, postoperative outcomes in 306 (52%), technical skills in 102 (17%), and patient-oriented outcomes in 38 (6%) articles. Over time, there was evidence of an increase in the relative amount of laparoscopic and robotic studies (P < 0.001) without statistical evidence of a change in the complexity of analysis (P = 0.121). Assessment of learning curves is needed to inform surgical training and evaluate new clinical procedures. An ideal analysis would account for the degree of complexity of individual cases and the inherent differences between surgeons. There is no single proxy that best represents the success of surgery, and hence multiple outcomes should be collected.
Use of Management Statistics in ARL Libraries. SPEC Kit #153.
ERIC Educational Resources Information Center
Vasi, John
A Systems and Procedures Exchange Center (SPEC) survey conducted in 1986 investigated the collection and use of management statistics in Association of Research Libraries (ARL) member libraries, and SPEC Kit #134 (May 1987) summarized the kinds of statistics collected and the reasons given by the 91 respondents for collecting them. This more…
The Practicality of Statistical Physics Handout Based on KKNI and the Constructivist Approach
NASA Astrophysics Data System (ADS)
Sari, S. Y.; Afrizon, R.
2018-04-01
Statistical physics lecture shows that: 1) the performance of lecturers, social climate, students’ competence and soft skills needed at work are in enough category, 2) students feel difficulties in following the lectures of statistical physics because it is abstract, 3) 40.72% of students needs more understanding in the form of repetition, practice questions and structured tasks, and 4) the depth of statistical physics material needs to be improved gradually and structured. This indicates that learning materials in accordance of The Indonesian National Qualification Framework or Kerangka Kualifikasi Nasional Indonesia (KKNI) with the appropriate learning approach are needed to help lecturers and students in lectures. The author has designed statistical physics handouts which have very valid criteria (90.89%) according to expert judgment. In addition, the practical level of handouts designed also needs to be considered in order to be easy to use, interesting and efficient in lectures. The purpose of this research is to know the practical level of statistical physics handout based on KKNI and a constructivist approach. This research is a part of research and development with 4-D model developed by Thiagarajan. This research activity has reached part of development test at Development stage. Data collection took place by using a questionnaire distributed to lecturers and students. Data analysis using descriptive data analysis techniques in the form of percentage. The analysis of the questionnaire shows that the handout of statistical physics has very practical criteria. The conclusion of this study is statistical physics handouts based on the KKNI and constructivist approach have been practically used in lectures.
Circulation Clusters--An Empirical Approach to Decentralization of Academic Libraries.
ERIC Educational Resources Information Center
McGrath, William E.
1986-01-01
Discusses the issue of centralization or decentralization of academic library collections, and describes a statistical analysis of book circulation at the University of Southwestern Louisiana that yielded subject area clusters as a compromise solution to the problem. Applications of the cluster model for all types of library catalogs are…
Turkish Student Teachers' Concerns about Teaching
ERIC Educational Resources Information Center
Boz, Yezdan
2008-01-01
The purpose of this study was to examine the teaching concerns of Turkish student teachers and how these concerns differ among year groups within the teacher education programme. Data were collected from 339 student teachers using the Teacher Concerns Checklist. Analysis of the data, including both descriptive statistics and multivariate analysis…
Using VITA Service Learning Experiences to Teach Hypothesis Testing and P-Value Analysis
ERIC Educational Resources Information Center
Drougas, Anne; Harrington, Steve
2011-01-01
This paper describes a hypothesis testing project designed to capture student interest and stimulate classroom interaction and communication. Using an online survey instrument, the authors collected student demographic information and data regarding university service learning experiences. Introductory statistics students performed a series of…
Illinois Teacher Supply and Demand, 1984-1985.
ERIC Educational Resources Information Center
Bartolini, Leandro
Statistics are presented on the current status of teacher supply and demand trends in Illinois. This report reviews and discusses the factors affecting teacher supply and demand, changes in student enrollment, teacher retirements, changes in state mandates, and opportunity for employment. An analysis of the data collected on teacher employment…
A Large-Scale Analysis of Variance in Written Language
ERIC Educational Resources Information Center
Johns, Brendan T.; Jamieson, Randall K.
2018-01-01
The collection of very large text sources has revolutionized the study of natural language, leading to the development of several models of language learning and distributional semantics that extract sophisticated semantic representations of words based on the statistical redundancies contained within natural language (e.g., Griffiths, Steyvers,…
GIS Adoption among Senior High School Geography Teachers in Taiwan
ERIC Educational Resources Information Center
Lay, Jinn-Guey; Chen, Yu-Wen; Chi, Yu-Lin
2013-01-01
This article explores the adoption of geographic information system (GIS) knowledge and skills through in-service training for high school geography teachers in Taiwan. Through statistical analysis of primary data collected from a census of Taiwan's high school geography teachers, it explores what motivates these teachers to undertake GIS…
American Samoa's forest resources, 2001.
Joseph A. Donnegan; Sheri S. Mann; Sarah L. Butler; Bruce A. Hiserote
2004-01-01
The Forest Inventory and Analysis Program of the Pacific Northwest Research Station collected, analyzed, and summarized data from field plots, and mapped land cover on four islands in American Samoa. This statistical sample provides estimates of forest area, stem volume, biomass, numbers of trees, damages to trees, and tree size distribution. The summary provides...
Systems Analysis of Alternative Architectures for Riverine Warfare in 2010
2006-12-01
propose system of systems improvements for the RF in 2010. With the RF currently working to establish a command structure, train and equip its forces...opposing force. Measures of performance such as time to first enemy detection and loss exchange ratio were collected from MANA. A detailed statistical
78 FR 50373 - Proposed Information Collection; Comment Request; Annual Capital Expenditures Survey
Federal Register 2010, 2011, 2012, 2013, 2014
2013-08-19
... source of detailed comprehensive statistics on actual business spending for non-farm companies, non- governmental companies, organizations, and associations operating in the United States. Both employer and nonemployer companies are included in the survey. The Bureau of Economic Analysis, the primary Federal user of...
Digital Learning Compass: Distance Education State Almanac 2017. Delaware
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Delaware. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Kansas
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Kansas. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Minnesota
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Minnesota. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Utah
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Utah. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Connecticut
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Connecticut. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Wyoming
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Wyoming. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Montana
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Montana. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Iowa
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Iowa. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Alabama
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Alabama. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Nevada
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Nevada. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Mississippi
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Mississippi. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Kentucky
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Kentucky. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Ohio
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Ohio. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Oklahoma
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Oklahoma. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Texas
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Texas. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Vermont
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Vermont. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Colorado
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Colorado. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Arizona
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Arizona . The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Missouri
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Missouri. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Idaho
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Idaho. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Massachusetts
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Massachusetts. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Tennessee
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Tennessee. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Virginia
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Virginia. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Indiana
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Indiana. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Alaska
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Alaska. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Louisiana
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Louisiana. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Nebraska
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Nebraska. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Maine
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Maine. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Wisconsin
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Wisconsin. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Michigan
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Michigan. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Arkansas
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Arkansas . The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Illinois
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Illinois. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Florida
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Florida. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Maryland
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Maryland. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Oregon
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Oregon. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Washington
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Washington. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Hawaii
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Hawaii. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. California
ERIC Educational Resources Information Center
Seaman, Julia A.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of California. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Georgia
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Georgia. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Pennsylvania
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Pennsylvania. The sample for this analysis is comprised of all active, degree-granting…
Statistics of Poverty: A Bibliography.
ERIC Educational Resources Information Center
Cameron, Colin; And Others
This bibliography is a collection of citations and illustrative tables which is meant to complement and supplement a 1976 work by the United States Department of Health, Education, and Welfare. That work included an analysis of several social service programs such as food stamps, public welfare, housing, and the education of the poor. The…
Forest land area estimates from vegetation continuous fields
Mark D. Nelson; Ronald E. McRoberts; Matthew C. Hansen
2004-01-01
The USDA Forest Service's Forest Inventory and Analysis (FIA) program provides data, information, and knowledge about our Nation's forest resources. FIA regional units collect data from field plots and remotely sensed imagery to produce statistical estimates of forest extent (area); volume, growth, and removals; and health and condition. There is increasing...
The study uses statistical analysis techniques to determine the effects of four heavy metals (cadmium, lead, manganese, and zinc) on the macroinvertebrate community using the data collected in the fall 1987.
USDA-ARS?s Scientific Manuscript database
Recent advances in technology have led to the collection of high-dimensional data not previously encountered in many scientific environments. As a result, scientists are often faced with the challenging task of including these high-dimensional data into statistical models. For example, data from sen...
Patankar, Ravindra
2003-10-01
Statistical fatigue life of a ductile alloy specimen is traditionally divided into three stages, namely, crack nucleation, small crack growth, and large crack growth. Crack nucleation and small crack growth show a wide variation and hence a big spread on cycles versus crack length graph. Relatively, large crack growth shows a lesser variation. Therefore, different models are fitted to the different stages of the fatigue evolution process, thus treating different stages as different phenomena. With these independent models, it is impossible to predict one phenomenon based on the information available about the other phenomenon. Experimentally, it is easier to carry out crack length measurements of large cracks compared to nucleating cracks and small cracks. Thus, it is easier to collect statistical data for large crack growth compared to the painstaking effort it would take to collect statistical data for crack nucleation and small crack growth. This article presents a fracture mechanics-based stochastic model of fatigue crack growth in ductile alloys that are commonly encountered in mechanical structures and machine components. The model has been validated by Ray (1998) for crack propagation by various statistical fatigue data. Based on the model, this article proposes a technique to predict statistical information of fatigue crack nucleation and small crack growth properties that uses the statistical properties of large crack growth under constant amplitude stress excitation. The statistical properties of large crack growth under constant amplitude stress excitation can be obtained via experiments.
NASA Astrophysics Data System (ADS)
Smith, Kimberly A.
The research study investigates the effectiveness of an integrated high school science curriculum on student achievement, knowledge retention and science attitudes using quantitative and qualitative research. Data was collected from tenth grade students, in a small urban high school in Kansas City, Missouri, who were enrolled in a traditional Biology course or an integrated Environmental Science course. Quantitative data was collected in Phase 1 of the study. Data collected for academic achievement included pretest and posttest scores on the CTBS MATN exam. Data collected for knowledge retention included post-posttest scores on the CTBS MATN exam. Data collected for science attitudes were scores on a pretest and posttest using the TOSRA. SPSS was used to analyze the data using independent samples t-tests, one-way ANCOVA's and paired samples statistics. Qualitative data was collected in Phase 2 of the study. Data included responses to open-ended interview questions using three focus groups. Data was analyzed for common themes. Data analysis revealed the integrated Environmental Science course had a statistically significant impact on academic achievement, knowledge retention and positive science attitudes. Gender and socioeconomic status did not influence results. The study also determined that the CTBS MATN exam was not an accurate predictor of scores on state testing as was previously thought.
The statistical analysis of circadian phase and amplitude in constant-routine core-temperature data
NASA Technical Reports Server (NTRS)
Brown, E. N.; Czeisler, C. A.
1992-01-01
Accurate estimation of the phases and amplitude of the endogenous circadian pacemaker from constant-routine core-temperature series is crucial for making inferences about the properties of the human biological clock from data collected under this protocol. This paper presents a set of statistical methods based on a harmonic-regression-plus-correlated-noise model for estimating the phases and the amplitude of the endogenous circadian pacemaker from constant-routine core-temperature data. The methods include a Bayesian Monte Carlo procedure for computing the uncertainty in these circadian functions. We illustrate the techniques with a detailed study of a single subject's core-temperature series and describe their relationship to other statistical methods for circadian data analysis. In our laboratory, these methods have been successfully used to analyze more than 300 constant routines and provide a highly reliable means of extracting phase and amplitude information from core-temperature data.
Data exploration, quality control and statistical analysis of ChIP-exo/nexus experiments
Welch, Rene; Chung, Dongjun; Grass, Jeffrey; Landick, Robert
2017-01-01
Abstract ChIP-exo/nexus experiments rely on innovative modifications of the commonly used ChIP-seq protocol for high resolution mapping of transcription factor binding sites. Although many aspects of the ChIP-exo data analysis are similar to those of ChIP-seq, these high throughput experiments pose a number of unique quality control and analysis challenges. We develop a novel statistical quality control pipeline and accompanying R/Bioconductor package, ChIPexoQual, to enable exploration and analysis of ChIP-exo and related experiments. ChIPexoQual evaluates a number of key issues including strand imbalance, library complexity, and signal enrichment of data. Assessment of these features are facilitated through diagnostic plots and summary statistics computed over regions of the genome with varying levels of coverage. We evaluated our QC pipeline with both large collections of public ChIP-exo/nexus data and multiple, new ChIP-exo datasets from Escherichia coli. ChIPexoQual analysis of these datasets resulted in guidelines for using these QC metrics across a wide range of sequencing depths and provided further insights for modelling ChIP-exo data. PMID:28911122
Data exploration, quality control and statistical analysis of ChIP-exo/nexus experiments.
Welch, Rene; Chung, Dongjun; Grass, Jeffrey; Landick, Robert; Keles, Sündüz
2017-09-06
ChIP-exo/nexus experiments rely on innovative modifications of the commonly used ChIP-seq protocol for high resolution mapping of transcription factor binding sites. Although many aspects of the ChIP-exo data analysis are similar to those of ChIP-seq, these high throughput experiments pose a number of unique quality control and analysis challenges. We develop a novel statistical quality control pipeline and accompanying R/Bioconductor package, ChIPexoQual, to enable exploration and analysis of ChIP-exo and related experiments. ChIPexoQual evaluates a number of key issues including strand imbalance, library complexity, and signal enrichment of data. Assessment of these features are facilitated through diagnostic plots and summary statistics computed over regions of the genome with varying levels of coverage. We evaluated our QC pipeline with both large collections of public ChIP-exo/nexus data and multiple, new ChIP-exo datasets from Escherichia coli. ChIPexoQual analysis of these datasets resulted in guidelines for using these QC metrics across a wide range of sequencing depths and provided further insights for modelling ChIP-exo data. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.
Comparing Methods for Assessing Reliability Uncertainty Based on Pass/Fail Data Collected Over Time
Abes, Jeff I.; Hamada, Michael S.; Hills, Charles R.
2017-12-20
In this paper, we compare statistical methods for analyzing pass/fail data collected over time; some methods are traditional and one (the RADAR or Rationale for Assessing Degradation Arriving at Random) was recently developed. These methods are used to provide uncertainty bounds on reliability. We make observations about the methods' assumptions and properties. Finally, we illustrate the differences between two traditional methods, logistic regression and Weibull failure time analysis, and the RADAR method using a numerical example.
Comparing Methods for Assessing Reliability Uncertainty Based on Pass/Fail Data Collected Over Time
DOE Office of Scientific and Technical Information (OSTI.GOV)
Abes, Jeff I.; Hamada, Michael S.; Hills, Charles R.
In this paper, we compare statistical methods for analyzing pass/fail data collected over time; some methods are traditional and one (the RADAR or Rationale for Assessing Degradation Arriving at Random) was recently developed. These methods are used to provide uncertainty bounds on reliability. We make observations about the methods' assumptions and properties. Finally, we illustrate the differences between two traditional methods, logistic regression and Weibull failure time analysis, and the RADAR method using a numerical example.
Bringing Clouds into Our Lab! - The Influence of Turbulence on the Early Stage Rain Droplets
NASA Astrophysics Data System (ADS)
Yavuz, Mehmet Altug; Kunnen, Rudie; Heijst, Gertjan; Clercx, Herman
2015-11-01
We are investigating a droplet-laden flow in an air-filled turbulence chamber, forced by speaker-driven air jets. The speakers are running in a random manner; yet they allow us to control and define the statistics of the turbulence. We study the motion of droplets with tunable size (Stokes numbers ~ 0.13 - 9) in a turbulent flow, mimicking the early stages of raindrop formation. 3D Particle Tracking Velocimetry (PTV) together with Laser Induced Fluorescence (LIF) methods are chosen as the experimental method to track the droplets and collect data for statistical analysis. Thereby it is possible to study the spatial distribution of the droplets in turbulence using the so-called Radial Distribution Function (RDF), a statistical measure to quantify the clustering of particles. Additionally, 3D-PTV technique allows us to measure velocity statistics of the droplets and the influence of the turbulence on droplet trajectories, both individually and collectively. In this contribution, we will present the clustering probability quantified by the RDF for different Stokes numbers. We will explain the physics underlying the influence of turbulence on droplet cluster behavior. This study supported by FOM/NWO Netherlands.
3 CFR - Enhanced Collection of Relevant Data and Statistics Relating to Women
Code of Federal Regulations, 2012 CFR
2012-01-01
... 3 The President 1 2012-01-01 2012-01-01 false Enhanced Collection of Relevant Data and Statistics Relating to Women Presidential Documents Other Presidential Documents Memorandum of March 4, 2011 Enhanced Collection of Relevant Data and Statistics Relating to Women Memorandum for the Heads of Executive Departments and Agencies I am proud to work...
Statistical analysis of Turbine Engine Diagnostic (TED) field test data
NASA Astrophysics Data System (ADS)
Taylor, Malcolm S.; Monyak, John T.
1994-11-01
During the summer of 1993, a field test of turbine engine diagnostic (TED) software, developed jointly by U.S. Army Research Laboratory and the U.S. Army Ordnance Center and School, was conducted at Fort Stuart, GA. The data were collected in conformance with a cross-over design, some of whose considerations are detailed. The initial analysis of the field test data was exploratory, followed by a more formal investigation. Technical aspects of the data analysis insights that were elicited are reported.
2017-08-30
as being three-fold: 1) a measurement of the integrity of both the central and peripheral visual processing centers; 2) an indicator of detail...visual assessment task 12 integral to the Army’s Class 1 Flight Physical (Ginsburg, 1981 and 1984; Bachman & Behar, 1986). During a Class 1 flight...systems. Meta-analysis has been defined as the statistical analysis of a collection of analytical results for the purpose of integrating the findings
An Overview of R in Health Decision Sciences.
Jalal, Hawre; Pechlivanoglou, Petros; Krijkamp, Eline; Alarid-Escudero, Fernando; Enns, Eva; Hunink, M G Myriam
2017-10-01
As the complexity of health decision science applications increases, high-level programming languages are increasingly adopted for statistical analyses and numerical computations. These programming languages facilitate sophisticated modeling, model documentation, and analysis reproducibility. Among the high-level programming languages, the statistical programming framework R is gaining increased recognition. R is freely available, cross-platform compatible, and open source. A large community of users who have generated an extensive collection of well-documented packages and functions supports it. These functions facilitate applications of health decision science methodology as well as the visualization and communication of results. Although R's popularity is increasing among health decision scientists, methodological extensions of R in the field of decision analysis remain isolated. The purpose of this article is to provide an overview of existing R functionality that is applicable to the various stages of decision analysis, including model design, input parameter estimation, and analysis of model outputs.
Analysis tools for discovering strong parity violation at hadron colliders
NASA Astrophysics Data System (ADS)
Backović, Mihailo; Ralston, John P.
2011-07-01
Several arguments suggest parity violation may be observable in high energy strong interactions. We introduce new analysis tools to describe the azimuthal dependence of multiparticle distributions, or “azimuthal flow.” Analysis uses the representations of the orthogonal group O(2) and dihedral groups DN necessary to define parity completely in two dimensions. Classification finds that collective angles used in event-by-event statistics represent inequivalent tensor observables that cannot generally be represented by a single “reaction plane.” Many new parity-violating observables exist that have never been measured, while many parity-conserving observables formerly lumped together are now distinguished. We use the concept of “event-shape sorting” to suggest separating right- and left-handed events, and we discuss the effects of transverse and longitudinal spin. The analysis tools are statistically robust, and can be applied equally to low or high multiplicity events at the Tevatron, RHIC or RHIC Spin, and the LHC.
76 FR 21780 - Agency Information Collection Activities: Bureau of Justice Statistics
Federal Register 2010, 2011, 2012, 2013, 2014
2011-04-18
... Justice Statistics ACTION: 30-Day notice of information collection under review: Reinstatement, with... Custody, 2011- 2012. The Department of Justice (DOJ), Bureau of Justice Statistics, will be submitting the... information, please contact Allen J. Beck, Ph.D., Bureau of Justice Statistics, 810 Seventh Street, NW...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-12-30
... for OMB Review; Comment Request; Mass Layoff Statistics Program ACTION: Notice. SUMMARY: The... request (ICR) titled, ``Mass Layoff Statistics Program,'' to the Office of Management and Budget (OMB) for... Statistics (BLS). Title of Collection: Mass Layoff Statistics Program. OMB Control Number: 1220-0090...
Statistics Report on TEQSA Registered Higher Education Providers
ERIC Educational Resources Information Center
Australian Government Tertiary Education Quality and Standards Agency, 2015
2015-01-01
This statistics report provides a comprehensive snapshot of national statistics on all parts of the sector for the year 2013, by bringing together data collected directly by TEQSA with data sourced from the main higher education statistics collections managed by the Australian Government Department of Education and Training. The report provides…
Rosen, G D
2006-06-01
Meta-analysis is a vague descriptor used to encompass very diverse methods of data collection analysis, ranging from simple averages to more complex statistical methods. Holo-analysis is a fully comprehensive statistical analysis of all available data and all available variables in a specified topic, with results expressed in a holistic factual empirical model. The objectives and applications of holo-analysis include software production for prediction of responses with confidence limits, translation of research conditions to praxis (field) circumstances, exposure of key missing variables, discovery of theoretically unpredictable variables and interactions, and planning future research. Holo-analyses are cited as examples of the effects on broiler feed intake and live weight gain of exogenous phytases, which account for 70% of variation in responses in terms of 20 highly significant chronological, dietary, environmental, genetic, managemental, and nutrient variables. Even better future accountancy of variation will be facilitated if and when authors of papers routinely provide key data for currently neglected variables, such as temperatures, complete feed formulations, and mortalities.
DARHT Multi-intelligence Seismic and Acoustic Data Analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Stevens, Garrison Nicole; Van Buren, Kendra Lu; Hemez, Francois M.
The purpose of this report is to document the analysis of seismic and acoustic data collected at the Dual-Axis Radiographic Hydrodynamic Test (DARHT) facility at Los Alamos National Laboratory for robust, multi-intelligence decision making. The data utilized herein is obtained from two tri-axial seismic sensors and three acoustic sensors, resulting in a total of nine data channels. The goal of this analysis is to develop a generalized, automated framework to determine internal operations at DARHT using informative features extracted from measurements collected external of the facility. Our framework involves four components: (1) feature extraction, (2) data fusion, (3) classification, andmore » finally (4) robustness analysis. Two approaches are taken for extracting features from the data. The first of these, generic feature extraction, involves extraction of statistical features from the nine data channels. The second approach, event detection, identifies specific events relevant to traffic entering and leaving the facility as well as explosive activities at DARHT and nearby explosive testing sites. Event detection is completed using a two stage method, first utilizing signatures in the frequency domain to identify outliers and second extracting short duration events of interest among these outliers by evaluating residuals of an autoregressive exogenous time series model. Features extracted from each data set are then fused to perform analysis with a multi-intelligence paradigm, where information from multiple data sets are combined to generate more information than available through analysis of each independently. The fused feature set is used to train a statistical classifier and predict the state of operations to inform a decision maker. We demonstrate this classification using both generic statistical features and event detection and provide a comparison of the two methods. Finally, the concept of decision robustness is presented through a preliminary analysis where uncertainty is added to the system through noise in the measurements.« less
Women's health and women's work in health services: what statistics tell us.
Hedman, B; Herner, E
1988-01-01
This article draws together statistical information in several broad areas that relate to women's health, women's reproductive activities and women's occupations in Sweden. The statistical analysis reflects the major changes that have occurred in Swedish society and that have had a major impact on the health and well-being, as well as on the social participation rate, of women. Much of the data is drawn from a recent special effort at Statistic Sweden aimed at influencing the classification, collection and presentation of statistical data in all fields in such a way that family, working, education, health and other conditions of women can be more readily and equitably compared with those of men. In addition, social changes have seen the shifting of the responsibility of health care from the unpaid duties of women in the home to health care institutions, where female employees predominate. These trends are also discussed.
The discrimination of sea ice types using SAR backscatter statistics
NASA Technical Reports Server (NTRS)
Shuchman, Robert A.; Wackerman, Christopher C.; Maffett, Andrew L.; Onstott, Robert G.; Sutherland, Laura L.
1989-01-01
X-band (HH) synthetic aperture radar (SAR) data of sea ice collected during the Marginal Ice Zone Experiment in March and April of 1987 was statistically analyzed with respect to discriminating open water, first-year ice, multiyear ice, and Odden. Odden are large expanses of nilas ice that rapidly form in the Greenland Sea and transform into pancake ice. A first-order statistical analysis indicated that mean versus variance can segment out open water and first-year ice, and skewness versus modified skewness can segment the Odden and multilayer categories. In additions to first-order statistics, a model has been generated for the distribution function of the SAR ice data. Segmentation of ice types was also attempted using textural measurements. In this case, the general co-occurency matrix was evaluated. The textural method did not generate better results than the first-order statistical approach.
Mercer, Theresa G; Frostick, Lynne E; Walmsley, Anthony D
2011-10-15
This paper presents a statistical technique that can be applied to environmental chemistry data where missing values and limit of detection levels prevent the application of statistics. A working example is taken from an environmental leaching study that was set up to determine if there were significant differences in levels of leached arsenic (As), chromium (Cr) and copper (Cu) between lysimeters containing preservative treated wood waste and those containing untreated wood. Fourteen lysimeters were setup and left in natural conditions for 21 weeks. The resultant leachate was analysed by ICP-OES to determine the As, Cr and Cu concentrations. However, due to the variation inherent in each lysimeter combined with the limits of detection offered by ICP-OES, the collected quantitative data was somewhat incomplete. Initial data analysis was hampered by the number of 'missing values' in the data. To recover the dataset, the statistical tool of Statistical Multiple Imputation (SMI) was applied, and the data was re-analysed successfully. It was demonstrated that using SMI did not affect the variance in the data, but facilitated analysis of the complete dataset. Copyright © 2011 Elsevier B.V. All rights reserved.
Using Statistical Analysis Software to Advance Nitro Plasticizer Wettability
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shear, Trevor Allan
Statistical analysis in science is an extremely powerful tool that is often underutilized. Additionally, it is frequently the case that data is misinterpreted or not used to its fullest extent. Utilizing the advanced software JMP®, many aspects of experimental design and data analysis can be evaluated and improved. This overview will detail the features of JMP® and how they were used to advance a project, resulting in time and cost savings, as well as the collection of scientifically sound data. The project analyzed in this report addresses the inability of a nitro plasticizer to coat a gold coated quartz crystalmore » sensor used in a quartz crystal microbalance. Through the use of the JMP® software, the wettability of the nitro plasticizer was increased by over 200% using an atmospheric plasma pen, ensuring good sample preparation and reliable results.« less
Effect of local and global geomagnetic activity on human cardiovascular homeostasis.
Dimitrova, Svetla; Stoilova, Irina; Yanev, Toni; Cholakov, Ilia
2004-02-01
The authors investigated the effects of local and planetary geomagnetic activity on human physiology. They collected data in Sofia, Bulgaria, from a group of 86 volunteers during the periods of the autumnal and vernal equinoxes. They used the factors local/planetary geomagnetic activity, day of measurement, gender, and medication use to apply a four-factor multiple analysis of variance. They also used a post hoc analysis to establish the statistical significance of the differences between the average values of the measured physiological parameters in the separate factor levels. In addition, the authors performed correlation analysis between the physiological parameters examined and geophysical factors. The results revealed that geomagnetic changes had a statistically significant influence on arterial blood pressure. Participants expressed this reaction with weak local geomagnetic changes and when major and severe global geomagnetic storms took place.
Data Model Performance in Data Warehousing
NASA Astrophysics Data System (ADS)
Rorimpandey, G. C.; Sangkop, F. I.; Rantung, V. P.; Zwart, J. P.; Liando, O. E. S.; Mewengkang, A.
2018-02-01
Data Warehouses have increasingly become important in organizations that have large amount of data. It is not a product but a part of a solution for the decision support system in those organizations. Data model is the starting point for designing and developing of data warehouses architectures. Thus, the data model needs stable interfaces and consistent for a longer period of time. The aim of this research is to know which data model in data warehousing has the best performance. The research method is descriptive analysis, which has 3 main tasks, such as data collection and organization, analysis of data and interpretation of data. The result of this research is discussed in a statistic analysis method, represents that there is no statistical difference among data models used in data warehousing. The organization can utilize four data model proposed when designing and developing data warehouse.
Kraeutler, Matthew J; Carver, Trevor J; Belk, John W; McCarty, Eric C
2018-06-01
Kraeutler, MJ, Carver, TJ, Belk, JW, and McCarty, EC. What is the value of a National Football League draft pick? An analysis based on changes made in the collective bargaining agreement. J Strength Cond Res 32(6): 1656-1661, 2018-The purpose of this study was to analyze and compare the value of players drafted in early rounds of the National Football League (NFL) Draft since the new collective bargaining agreement began in 2011. The NFL's player statistics database and database of player contract details were searched for players drafted in the first 3 rounds of the 2011 to 2013 NFL Drafts. Performance outcomes specific to each position were divided by each player's salary to calculate a value statistic. Various demographics, NFL Combine results, and total number of games missed because of injury were also recorded for each player. These statistics were compared within each position between players selected in the first round of the NFL Draft (group A) vs. those drafted in the second or third round (group B). A total of 147 players were included (group A 35, group B 112). Overall, players in group A were significantly taller (p ≤ 0.01) and heavier (p = 0.037) than players in group B. Group B demonstrated significantly greater value statistics than group A for quarterbacks (p = 0.028), wide receivers (p ≤ 0.001), defensive tackles (p = 0.019), and cornerbacks (p ≤ 0.001). No significant differences were found between groups with regard to number of games missed because of injury. Players drafted in the second or third rounds of the NFL Draft often carry more value than those drafted in the first round. NFL teams may wish to more frequently trade down in the Draft rather than trading up.
Supercritical Fluid Extraction and Analysis of Tropospheric Aerosol Particles
NASA Astrophysics Data System (ADS)
Hansen, Kristen J.
An integrated sampling and supercritical fluid extraction (SFE) cell has been designed for whole-sample analysis of organic compounds on tropospheric aerosol particles. The low-volume extraction cell has been interfaced with a sampling manifold for aerosol particle collection in the field. After sample collection, the entire SFE cell was coupled to a gas chromatograph; after on-line extraction, the cryogenically -focused sample was separated and the volatile compounds detected with either a mass spectrometer or a flame ionization detector. A 20-minute extraction at 450 atm and 90 ^circC with pure supercritical CO _2 is sufficient for quantitative extraction of most volatile compounds in aerosol particle samples. A comparison between SFE and thermal desorption, the traditional whole-sample technique for analyses of this type, was performed using ambient aerosol particle samples, as well as samples containing known amounts of standard analytes. The results of these studies indicate that SFE of atmospheric aerosol particles provides quantitative measurement of several classes of organic compounds. SFE provides information that is complementary to that gained by the thermal desorption analysis. The results also indicate that SFE with CO _2 can be validated as an alternative to thermal desorption for quantitative recovery of several organic compounds. In 1989, the organic constituents of atmospheric aerosol particles collected at Niwot Ridge, Colorado, along with various physical and meteorological data, were measured during a collaborative field study. Temporal changes in the composition of samples collected during summertime at the rural site were studied. Thermal desorption-GC/FID was used to quantify selected compounds in samples collected during the field study. The statistical analysis of the 1989 Niwot Ridge data set is presented in this work. Principal component analysis was performed on thirty-one variables selected from the data set in order to ascertain different source and process components, and to examine concentration changes in groups of variables with respect to time of day and meteorological conditions. Seven orthogonal groups of variables resulted from the statistical analysis; the groups serve as molecular markers for different biologic and anthropogenic emission sources. In addition, the results of the statistical analysis were used to investigate how several emission source contributions vary with respect to local atmospheric dynamics. Field studies were conducted in the urban environment in and around Boulder, CO. to characterize the dynamics, chemistry, and emission sources which affect the composition and concentration of different size-fractions of aerosol particles in the Boulder air mass. Relationships between different size fractions of particles and some gas-phase pollutants were elucidated. These field studies included an investigation of seasonal variations in the organic content and concentration of aerosol particles, and how these characteristics are related to local meteorology and to the concentration of some gas-phase pollutants. The elemental and organic composition of aerosol particles was investigated according to particle size in preliminary studies of size-differentiated samples of aerosol particles. In order to aid in future studies of urban aerosol particles, samples were collected at a forest fire near Boulder. Molecular markers specific to wood burning processes will be useful indicators of residential wood burning activities in future field studies.
Kostopoulos, Spiros; Ravazoula, Panagiota; Asvestas, Pantelis; Kalatzis, Ioannis; Xenogiannopoulos, George; Cavouras, Dionisis; Glotsos, Dimitris
2017-06-01
Histopathology image processing, analysis and computer-aided diagnosis have been shown as effective assisting tools towards reliable and intra-/inter-observer invariant decisions in traditional pathology. Especially for cancer patients, decisions need to be as accurate as possible in order to increase the probability of optimal treatment planning. In this study, we propose a new image collection library (HICL-Histology Image Collection Library) comprising 3831 histological images of three different diseases, for fostering research in histopathology image processing, analysis and computer-aided diagnosis. Raw data comprised 93, 116 and 55 cases of brain, breast and laryngeal cancer respectively collected from the archives of the University Hospital of Patras, Greece. The 3831 images were generated from the most representative regions of the pathology, specified by an experienced histopathologist. The HICL Image Collection is free for access under an academic license at http://medisp.bme.teiath.gr/hicl/ . Potential exploitations of the proposed library may span over a board spectrum, such as in image processing to improve visualization, in segmentation for nuclei detection, in decision support systems for second opinion consultations, in statistical analysis for investigation of potential correlations between clinical annotations and imaging findings and, generally, in fostering research on histopathology image processing and analysis. To the best of our knowledge, the HICL constitutes the first attempt towards creation of a reference image collection library in the field of traditional histopathology, publicly and freely available to the scientific community.
Study/experimental/research design: much more than statistics.
Knight, Kenneth L
2010-01-01
The purpose of study, experimental, or research design in scientific manuscripts has changed significantly over the years. It has evolved from an explanation of the design of the experiment (ie, data gathering or acquisition) to an explanation of the statistical analysis. This practice makes "Methods" sections hard to read and understand. To clarify the difference between study design and statistical analysis, to show the advantages of a properly written study design on article comprehension, and to encourage authors to correctly describe study designs. The role of study design is explored from the introduction of the concept by Fisher through modern-day scientists and the AMA Manual of Style. At one time, when experiments were simpler, the study design and statistical design were identical or very similar. With the complex research that is common today, which often includes manipulating variables to create new variables and the multiple (and different) analyses of a single data set, data collection is very different than statistical design. Thus, both a study design and a statistical design are necessary. Scientific manuscripts will be much easier to read and comprehend. A proper experimental design serves as a road map to the study methods, helping readers to understand more clearly how the data were obtained and, therefore, assisting them in properly analyzing the results.
Guam's forest resources, 2002.
Joseph A. Donnegan; Sarah L. Butler; Walter Grabowiecki; Bruce A. Hiserote; David. Limtiaco
2004-01-01
The Forest Inventory and Analysis Program collected, analyzed, and summarized field data on 46 forested plots on the island of Guam. Estimates of forest area, tree stem volume and biomass, the numbers of trees, tree damages, and the distribution of tree sizes were summarized for this statistical sample. Detailed tables and graphical highlights provide a summary of Guam...
The Impact of Social Capital on the Employment of College Graduates
ERIC Educational Resources Information Center
Fengqiao, Yan; Dan, Mao
2015-01-01
This article addresses the impact of social capital on college graduate employment. After reviewing the literature, the authors analyze data collected by Peking University from 34 universities in 2005 and use statistical analysis to clarify the impact of social capital on students' choice of employment or further study, job placement rate,…
Impact of E-Learning and Digitalization in Primary and Secondary Schools
ERIC Educational Resources Information Center
Tunmibi, Sunday; Aregbesola, Ayooluwa; Adejobi, Pascal; Ibrahim, Olaniyi
2015-01-01
This study examines into the impact of e-learning and digitalization in primary and secondary schools, using Greensprings School in Lagos State, Nigeria as a case study. Questionnaire was used as a data collection instrument, and descriptive statistical method was adopted for analysis. Responses from students and teachers reveal that application…
Spectral Analysis; Applications in Water Pollution Control.
ERIC Educational Resources Information Center
Wastler, T. A.
The statistical technique of analyzing data collected at regular intervals to reveal periodic components of the data is described by reference to actual records. The data chosen for illustration include tide height in a river; biochemical oxygen demand and dissolved oxygen in the same river; discharged salt into a river system and its relation to…
A Unified Approach to Measurement Error and Missing Data: Overview and Applications
ERIC Educational Resources Information Center
Blackwell, Matthew; Honaker, James; King, Gary
2017-01-01
Although social scientists devote considerable effort to mitigating measurement error during data collection, they often ignore the issue during data analysis. And although many statistical methods have been proposed for reducing measurement error-induced biases, few have been widely used because of implausible assumptions, high levels of model…
Digital Learning Compass: Distance Education State Almanac 2017. North Dakota
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of North Dakota. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. West Virginia
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of West Virginia. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. South Dakota
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of South Dakota. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. North Carolina
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of North Carolina. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. Rhode Island
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of Rhode Island. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. New Hampshire
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of New Hampshire. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. New Jersey
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of New Jersey. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. New Mexico
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of New Mexico. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. New York
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of New York. The sample for this analysis is comprised of all active, degree-granting…
Digital Learning Compass: Distance Education State Almanac 2017. South Carolina
ERIC Educational Resources Information Center
Seaman, Julia E.; Seaman, Jeff
2017-01-01
This brief report uses data collected under the U.S. Department of Education's National Center for Educational Statistics (NCES) Integrated Postsecondary Education Data System (IPEDS) Fall Enrollment survey to highlight distance education data in the state of South Carolina. The sample for this analysis is comprised of all active, degree-granting…
Guidelines for collecting and maintaining archives for genetic monitoring
Jennifer A. Jackson; Linda Laikre; C. Scott Baker; Katherine C. Kendall; F. W. Allendorf; M. K. Schwartz
2011-01-01
Rapid advances in molecular genetic techniques and the statistical analysis of genetic data have revolutionized the way that populations of animals, plants and microorganisms can be monitored. Genetic monitoring is the practice of using molecular genetic markers to track changes in the abundance, diversity or distribution of populations, species or ecosystems over time...
Palau's forest resources, 2003.
Joseph A. Donnegan; Sarah L. Butler; Olaf Kuegler; Brent J. Stroud; Bruce A. Hiserote; Kashgar. Rengulbai
2007-01-01
The Forest Inventory and Analysis Program collected, analyzed, and summarized field data on 54 forested plots on the islands in the Republic of Palau. Estimates of forest area, tree stem volume and biomass, the numbers of trees, tree damages, and the distribution of tree sizes were summarized for this statistical sample. Detailed tables and graphical highlights provide...
Preliminary Multi-Variable Parametric Cost Model for Space Telescopes
NASA Technical Reports Server (NTRS)
Stahl, H. Philip; Hendrichs, Todd
2010-01-01
This slide presentation reviews creating a preliminary multi-variable cost model for the contract costs of making a space telescope. There is discussion of the methodology for collecting the data, definition of the statistical analysis methodology, single variable model results, testing of historical models and an introduction of the multi variable models.
ERIC Educational Resources Information Center
Wang, Lijuan; Ha, Amy Sau-ching; Wen, Xu
2014-01-01
This research primarily aimed to examine the compatibility of teaching perspectives of teachers with the Physical Education (PE) curriculum in China. The Teaching Perspective Inventory (Pratt, 1998) was used to collect data from 272 PE teachers. Descriptive statistics, MANOVAs, and correlational procedures were used for quantitative data analysis.…
Federal Register 2010, 2011, 2012, 2013, 2014
2010-04-27
... the Bureau of Economic Analysis (BEA) for compiling the U.S. international transactions accounts (ITAs), which BEA publishes quarterly in news releases, on its Web site, and in its monthly journal, the Survey of Current Business. These accounts provide a statistical summary of all U.S. international...
Farmers as Consumers of Agricultural Education Services: Willingness to Pay and Spend Time
ERIC Educational Resources Information Center
Charatsari, Chrysanthi; Papadaki-Klavdianou, Afroditi; Michailidis, Anastasios
2011-01-01
This study assessed farmers' willingness to pay for and spend time attending an Agricultural Educational Program (AEP). Primary data on the demographic and socio-economic variables of farmers were collected from 355 farmers selected randomly from Northern Greece. Descriptive statistics and multivariate analysis methods were used in order to meet…
Structure of Student Time Management Scale (STMS)
ERIC Educational Resources Information Center
Balamurugan, M.
2013-01-01
With the aim of constructing a Student Time Management Scale (STMS), the initial version was administered and data were collected from 523 standard eleventh students. (Mean age = 15.64). The data obtained were subjected to Reliability and Factor analysis using PASW Statistical software version 18. From 42 items 14 were dropped, resulting in the…
ERIC Educational Resources Information Center
Hicks, Catherine
2018-01-01
Purpose: This paper aims to explore predicting employee learning activity via employee characteristics and usage for two online learning tools. Design/methodology/approach: Statistical analysis focused on observational data collected from user logs. Data are analyzed via regression models. Findings: Findings are presented for over 40,000…
V for Voice: Strategies for Bolstering Communication Skills in Statistics
ERIC Educational Resources Information Center
Khachatryan, Davit; Karst, Nathaniel
2017-01-01
With the ease and automation of data collection and plummeting storage costs, organizations are faced with massive amounts of data that present two pressing challenges: technical analysis of the data themselves and communication of the analytics process and its products. Although a plethora of academic and practitioner literature have focused on…
Negotiated Wages and Working Conditions in Ontario Hospitals: 1973.
ERIC Educational Resources Information Center
Ontario Dept. of Labour, Toronto. Research Branch.
This report is a statistical analysis of provisions in collective agreements covering approximately 38,000 full-time employees in 156 hospitals in the Province of Ontario. Part 1 consists of 56 tables giving information on the geographical distribution of hospital contracts, the unions that are party to them, their duration, and the sizes and…
Levitan, Denise M.; Zipper, Carl E.; Donovan, Patricia; Schreiber, Madeline E.; Seal, Robert; Engle, Mark A.; Chermak, John A.; Bodnar, Robert J.; Johnson, Daniel K.; Aylor, Joseph G.
2015-01-01
Soil geochemical anomalies can be used to identify pathfinders in exploration for ore deposits. In this study, compositional data analysis is used with multivariate statistical methods to analyse soil geochemical data collected from the Coles Hill uranium deposit, Virginia, USA, to identify pathfinders associated with this deposit. Elemental compositions and relationships were compared between the collected Coles Hill soil and reference soil samples extracted from a regional subset of a national-scale geochemical survey. Results show that pathfinders for the Coles Hill deposit include light rare earth elements (La and Ce), which, when normalised by their Al content, are correlated with U/Al, and elevated Th/Al values, which are not correlated with U/Al, supporting decoupling of U from Th during soil generation. These results can be used in genetic and weathering models of the Coles Hill deposit, and can also be applied to future prospecting for similar U deposits in the eastern United States, and in regions with similar geological/climatic conditions.
NASA Astrophysics Data System (ADS)
Vagge, Greta; Cutroneo, Laura; Gandolfi, Daniela; Ferretti, Gabriele; Scafidi, Davide; Capello, Marco
2018-05-01
A historical set of daily meteorological data collected at the Hanbury Botanical Gardens (Liguria, northwestern Italy) from 1900 to 1940 was recovered from five manually entered registers. They were digitised and statistically analysed to check their reliability and study their trends and variations. In particular, air temperature, precipitation and their extreme values were considered, together with wind direction, sea state, sky conditions and relative humidity. The results show a decreasing trend in mean annual temperature of approximately 0.2 °C/decade due to a decrease in maximum air temperature. Annual cumulative precipitation increased by 65.2 mm/decade over the study period. The data analysis showed a summer temperature decrease in 1912 and a severe drought in 1921. Moreover, the years with most days with extreme temperatures were associated to the negative phases of the North Atlantic oscillation (NAO). During the study period, SW winds were prevailing. Sky conditions followed seasonal trends, while slight sea was the most frequent sea state.
Kirgiz, Irina A; Calloway, Cassandra
2017-04-01
Tape lifting and FTA paper scraping methods were directly compared to traditional double swabbing for collecting touch DNA from car steering wheels (n = 70 cars). Touch DNA was collected from the left or right side of each steering wheel (randomized) using two sterile cotton swabs, while the other side was sampled using water-soluble tape or FTA paper cards. DNA was extracted and quantified in duplicate using qPCR. Quantifiable amounts of DNA were detected for 100% of the samples (n = 140) collected independent of the method. However, the DNA collection yield was dependent on the collection method. A statistically significant difference in DNA yield was observed between FTA scraping and double swabbing methods (p = 0.0051), with FTA paper collecting a two-fold higher amount. Statistical analysis showed no significant difference in DNA yields between the double swabbing and tape lifting techniques (p = 0.21). Based on the DNA concentration required for 1 ng input, 47% of the samples collected using FTA paper would be expected to yield a short tandem repeat (STR) profile compared to 30% and 23% using double swabbing or tape, respectively. Further, 55% and 77% of the samples collected using double swabbing or tape, respectively, did not yield a high enough DNA concentration for the 0.5 ng of DNA input recommended for conventional STR kits and would be expected to result in a partial or no profile compared to 35% of the samples collected using FTA paper. STR analysis was conducted for a subset of the higher concentrated samples to confirm that the DNA collected from the steering wheel was from the driver. 32 samples were selected with DNA amounts of at least 1 ng total DNA (100 pg/μl when concentrated if required). A mixed STR profile was observed for 26 samples (88%) and the last driver was the major DNA contributor for 29 samples (94%). For one sample, the last driver was the minor DNA contributor. A full STR profile of the last driver was observed for 21 samples (69%) and a partial profile was observed for nine samples (25%); STR analysis failed for two samples collected using tape (6%). In conclusion, we show that the FTA paper scraping method has the potential to collect higher DNA yields from touch DNA evidence deposited on non-porous surfaces often encountered in criminal cases compared to conventional methods. Copyright © 2017 Elsevier Ltd and Faculty of Forensic and Legal Medicine. All rights reserved.
Statistical Analysis of Time-Series from Monitoring of Active Volcanic Vents
NASA Astrophysics Data System (ADS)
Lachowycz, S.; Cosma, I.; Pyle, D. M.; Mather, T. A.; Rodgers, M.; Varley, N. R.
2016-12-01
Despite recent advances in the collection and analysis of time-series from volcano monitoring, and the resulting insights into volcanic processes, challenges remain in forecasting and interpreting activity from near real-time analysis of monitoring data. Statistical methods have potential to characterise the underlying structure and facilitate intercomparison of these time-series, and so inform interpretation of volcanic activity. We explore the utility of multiple statistical techniques that could be widely applicable to monitoring data, including Shannon entropy and detrended fluctuation analysis, by their application to various data streams from volcanic vents during periods of temporally variable activity. Each technique reveals changes through time in the structure of some of the data that were not apparent from conventional analysis. For example, we calculate the Shannon entropy (a measure of the randomness of a signal) of time-series from the recent dome-forming eruptions of Volcán de Colima (Mexico) and Soufrière Hills (Montserrat). The entropy of real-time seismic measurements and the count rate of certain volcano-seismic event types from both volcanoes is found to be temporally variable, with these data generally having higher entropy during periods of lava effusion and/or larger explosions. In some instances, the entropy shifts prior to or coincident with changes in seismic or eruptive activity, some of which were not clearly recognised by real-time monitoring. Comparison with other statistics demonstrates the sensitivity of the entropy to the data distribution, but that it is distinct from conventional statistical measures such as coefficient of variation. We conclude that each analysis technique examined could provide valuable insights for interpretation of diverse monitoring time-series.
Technology, Data Bases and System Analysis for Space-to-Ground Optical Communications
NASA Technical Reports Server (NTRS)
Lesh, James
1995-01-01
Optical communications is becoming an ever-increasingly important option for designers of space-to- ground communications links, whether it be for government or commercial applications. In this paper the technology being developed by NASA for use in space-to-ground optical communications is presented. Next, a program which is collecting a long term data base of atmospheric visibility statistics for optical propagation through the atmosphere will be described. Finally, a methodology for utilizing the statistics of the atmospheric data base in the analysis of space-to-ground links will be presented. This methodology takes into account the effects of station availability, is useful when comparing optical communications with microwave systems, and provides a rationale establishing the recommended link margin.
Calibrated Noise Measurements with Induced Receiver Gain Fluctuations
NASA Technical Reports Server (NTRS)
Racette, Paul; Walker, David; Gu, Dazhen; Rajola, Marco; Spevacek, Ashly
2011-01-01
The lack of well-developed techniques for modeling changing statistical moments in our observations has stymied the application of stochastic process theory in science and engineering. These limitations were encountered when modeling the performance of radiometer calibration architectures and algorithms in the presence of non stationary receiver fluctuations. Analyses of measured signals have traditionally been limited to a single measurement series. Whereas in a radiometer that samples a set of noise references, the data collection can be treated as an ensemble set of measurements of the receiver state. Noise Assisted Data Analysis is a growing field of study with significant potential for aiding the understanding and modeling of non stationary processes. Typically, NADA entails adding noise to a signal to produce an ensemble set on which statistical analysis is performed. Alternatively as in radiometric measurements, mixing a signal with calibrated noise provides, through the calibration process, the means to detect deviations from the stationary assumption and thereby a measurement tool to characterize the signal's non stationary properties. Data sets comprised of calibrated noise measurements have been limited to those collected with naturally occurring fluctuations in the radiometer receiver. To examine the application of NADA using calibrated noise, a Receiver Gain Modulation Circuit (RGMC) was designed and built to modulate the gain of a radiometer receiver using an external signal. In 2010, an RGMC was installed and operated at the National Institute of Standards and Techniques (NIST) using their Noise Figure Radiometer (NFRad) and national standard noise references. The data collected is the first known set of calibrated noise measurements from a receiver with an externally modulated gain. As an initial step, sinusoidal and step-function signals were used to modulate the receiver gain, to evaluate the circuit characteristics and to study the performance of a variety of calibration algorithms. The receiver noise temperature and time-bandwidth product of the NFRad are calculated from the data. Statistical analysis using temporal-dependent calibration algorithms reveals that the natural occurring fluctuations in the receiver are stationary over long intervals (100s of seconds); however the receiver exhibits local non stationarity over the interval over which one set of reference measurements are collected. A variety of calibration algorithms have been applied to the data to assess algorithms' performance with the gain fluctuation signals. This presentation will describe the RGMC, experiment design and a comparative analysis of calibration algorithms.
Melching, C.S.; Coupe, R.H.
1995-01-01
During water years 1985-91, the U.S. Geological Survey (USGS) and the Illinois Environmental Protection Agency (IEPA) cooperated in the collection and analysis of concurrent and split stream-water samples from selected sites in Illinois. Concurrent samples were collected independently by field personnel from each agency at the same time and sent to the IEPA laboratory, whereas the split samples were collected by USGS field personnel and divided into aliquots that were sent to each agency's laboratory for analysis. The water-quality data from these programs were examined by means of the Wilcoxon signed ranks test to identify statistically significant differences between results of the USGS and IEPA analyses. The data sets for constituents and properties identified by the Wilcoxon test as having significant differences were further examined by use of the paired t-test, mean relative percentage difference, and scattergrams to determine if the differences were important. Of the 63 constituents and properties in the concurrent-sample analysis, differences in only 2 (pH and ammonia) were statistically significant and large enough to concern water-quality engineers and planners. Of the 27 constituents and properties in the split-sample analysis, differences in 9 (turbidity, dissolved potassium, ammonia, total phosphorus, dissolved aluminum, dissolved barium, dissolved iron, dissolved manganese, and dissolved nickel) were statistically significant and large enough to con- cern water-quality engineers and planners. The differences in concentration between pairs of the concurrent samples were compared to the precision of the laboratory or field method used. The differences in concentration between pairs of the concurrent samples were compared to the precision of the laboratory or field method used. The differences in concentration between paris of split samples were compared to the precision of the laboratory method used and the interlaboratory precision of measuring a given concentration or property. Consideration of method precision indicated that differences between concurrent samples were insignificant for all concentrations and properties except pH, and that differences between split samples were significant for all concentrations and properties. Consideration of interlaboratory precision indicated that the differences between the split samples were not unusually large. The results for the split samples illustrate the difficulty in obtaining comparable and accurate water-quality data.
Analysis of repeated measurement data in the clinical trials
Singh, Vineeta; Rana, Rakesh Kumar; Singhal, Richa
2013-01-01
Statistics is an integral part of Clinical Trials. Elements of statistics span Clinical Trial design, data monitoring, analyses and reporting. A solid understanding of statistical concepts by clinicians improves the comprehension and the resulting quality of Clinical Trials. In biomedical research it has been seen that researcher frequently use t-test and ANOVA to compare means between the groups of interest irrespective of the nature of the data. In Clinical Trials we record the data on the patients more than two times. In such a situation using the standard ANOVA procedures is not appropriate as it does not consider dependencies between observations within subjects in the analysis. To deal with such types of study data Repeated Measure ANOVA should be used. In this article the application of One-way Repeated Measure ANOVA has been demonstrated by using the software SPSS (Statistical Package for Social Sciences) Version 15.0 on the data collected at four time points 0 day, 15th day, 30th day, and 45th day of multicentre clinical trial conducted on Pandu Roga (~Iron Deficiency Anemia) with an Ayurvedic formulation Dhatrilauha. PMID:23930038
NASA Astrophysics Data System (ADS)
Deidda, Roberto; Mascaro, Giuseppe; Hellies, Matteo; Baldini, Luca; Roberto, Nicoletta
2013-04-01
COSMO Sky-Med (CSK) is an important programme of the Italian Space Agency aiming at supporting environmental monitoring and management of exogenous, endogenous and anthropogenic risks through X-band Synthetic Aperture Radar (X-SAR) on board of 4 satellites forming a constellation. Most of typical SAR applications are focused on land or ocean observation. However, X-band SAR can be detect precipitation that results in a specific signature caused by the combination of attenuation of surface returns induced by precipitation and enhancement of backscattering determined by the hydrometeors in the SAR resolution volume. Within CSK programme, we conducted an intercomparison between the statistical properties of precipitation fields derived by CSK SARs and those derived by the CNR Polar 55C (C-band) ground based weather radar located in Rome (Italy). This contribution presents main results of this research which was aimed at the robust characterisation of rainfall statistical properties across different scales by means of scale-invariance analysis and multifractal theory. The analysis was performed on a dataset of more two years of precipitation observations collected by the CNR Polar 55C radar and rainfall fields derived from available images collected by the CSK satellites during intense rainfall events. Scale-invariance laws and multifractal properties were detected on the most intense rainfall events derived from the CNR Polar 55C radar for spatial scales from 4 km to 64 km. The analysis on X-SAR retrieved rainfall fields, although based on few images, leaded to similar results and confirmed the existence of scale-invariance and multifractal properties for scales larger than 4 km. These outcomes encourage investigating SAR methodologies for future development of meteo-hydrological forecasting models based on multifractal theory.
Ahmad, Sheikh Saeed; Aziz, Neelam; Butt, Amna; Shabbir, Rabia; Erum, Summra
2015-09-01
One of the features of medical geography that has made it so useful in health research is statistical spatial analysis, which enables the quantification and qualification of health events. The main objective of this research was to study the spatial distribution patterns of malaria in Rawalpindi district using spatial statistical techniques to identify the hot spots and the possible risk factor. Spatial statistical analyses were done in ArcGIS, and satellite images for land use classification were processed in ERDAS Imagine. Four hundred and fifty water samples were also collected from the study area to identify the presence or absence of any microbial contamination. The results of this study indicated that malaria incidence varied according to geographical location, with eco-climatic condition and showing significant positive spatial autocorrelation. Hotspots or location of clusters were identified using Getis-Ord Gi* statistic. Significant clustering of malaria incidence occurred in rural central part of the study area including Gujar Khan, Kaller Syedan, and some part of Kahuta and Rawalpindi Tehsil. Ordinary least square (OLS) regression analysis was conducted to analyze the relationship of risk factors with the disease cases. Relationship of different land cover with the disease cases indicated that malaria was more related with agriculture, low vegetation, and water class. Temporal variation of malaria cases showed significant positive association with the meteorological variables including average monthly rainfall and temperature. The results of the study further suggested that water supply and sewage system and solid waste collection system needs a serious attention to prevent any outbreak in the study area.
75 FR 37839 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2010-06-30
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting... Systems, Bureau of Labor Statistics, Room 4080, 2 Massachusetts Avenue, NE., [[Page 37840
75 FR 5346 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2010-02-02
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting... Clearance Officer, Division of Management Systems, Bureau of Labor Statistics, Room 4080, 2 Massachusetts...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-05-17
... DEPARTMENT OF COMMERCE National Oceanic and Atmospheric Administration Proposed Information Collection; Comment Request; Marine Recreational Fisheries Statistics Survey AGENCY: National Oceanic and... Andrews, NOAA, National Marine Fisheries Service, Fisheries Statistics Division, Phone: (301) 713-2328 or...
77 FR 36296 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2012-06-18
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting... Nora Kincaid, BLS Clearance Officer, Division of Management Systems, Bureau of Labor Statistics, Room...
78 FR 41958 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2013-07-12
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting... Nora Kincaid, BLS Clearance Officer, Division of Management Systems, Bureau of Labor Statistics, Room...
76 FR 71076 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2011-11-16
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting comments on the proposed extension of the ``BLS Occupational Safety and Health Statistics (OSHS...
Okumu, Clarice; Oyugi, Boniface
2018-01-01
This study intended to compare the clients' satisfaction with the quality of childbirth services in a private and public facility amongst mothers who have delivered within the last twenty four to seventy hours. This was a cross-sectional comparative research design with both quantitative and qualitative data collection and analysis methods. Data were collected through a focused group discussion guide and structured questionnaire collecting information on clients' satisfaction with quality of childbirth services. The study was conducted amongst women of reproductive age (WRA) between 15-49 years in Tigoni District hospital (public hospital) and Limuru Nursing home (private hospital). For quantitative data we conducted descriptive analysis and Mann-Whitney test using SPSS version 20.0 while qualitative data was manually analyzed manually using thematic analysis. A higher proportion of clients from private facility 98.1% were attended within 0-30 minutes of arrival to the facility as compared to 87% from public facility. The overall mean score showed that the respondents in public facility gave to satisfaction with the services was 4.46 out of a maximum of 5.00 score while private facility gave 4.60. The level of satisfaction amongst respondents in the public facility on pain relief after delivery was statistically significantly higher than the respondents in private facilities (U = 8132.50, p<0.001) while the level of satisfaction amongst respondents in the public facility on functional equipment was statistically significantly higher than the respondents in private facilities (U = 9206.50, p = 0.001). Moreover, level of satisfaction with the way staff responded to questions and concerns during labour and delivery was statistically significantly higher than the respondents in private facilities (U = 9964.50, p = 0.022). In overall, majority of clients from both public and private facilities expressed satisfaction with quality of services from admission till discharge in both public and private facilities and were willing to recommend other to come and deliver in the respective facilities.
Iacono, William G; Malone, Stephen M; Vaidyanathan, Uma; Vrieze, Scott I
2014-12-01
This article provides an introductory overview of the investigative strategy employed to evaluate the genetic basis of 17 endophenotypes examined as part of a 20-year data collection effort from the Minnesota Center for Twin and Family Research. Included are characterization of the study samples, descriptive statistics for key properties of the psychophysiological measures, and rationale behind the steps taken in the molecular genetic study design. The statistical approach included (a) biometric analysis of twin and family data, (b) heritability analysis using 527,829 single nucleotide polymorphisms (SNPs), (c) genome-wide association analysis of these SNPs and 17,601 autosomal genes, (d) follow-up analyses of candidate SNPs and genes hypothesized to have an association with each endophenotype, (e) rare variant analysis of nonsynonymous SNPs in the exome, and (f) whole genome sequencing association analysis using 27 million genetic variants. These methods were used in the accompanying empirical articles comprising this special issue, Genome-Wide Scans of Genetic Variants for Psychophysiological Endophenotypes. Copyright © 2014 Society for Psychophysiological Research.
Effectiveness of propolis on oral health: a meta-analysis.
Hwu, Yueh-Juen; Lin, Feng-Yu
2014-12-01
The use of propolis mouth rinse or gel as a supplementary intervention has increased during the last decade in Taiwan. However, the effect of propolis on oral health is not well understood. The purpose of this meta-analysis was to present the best available evidence regarding the effects of propolis use on oral health, including oral infection, dental plaque, and stomatitis. Researchers searched seven electronic databases for relevant articles published between 1969 and 2012. Data were collected using inclusion and exclusion criteria. The Joanna Briggs Institute Meta Analysis of Statistics Assessment and Review Instrument was used to evaluate the quality of the identified articles. Eight trials published from 1997 to 2011 with 194 participants had extractable data. The result of the meta-analysis indicated that, although propolis had an effect on reducing dental plaque, this effect was not statistically significant. The results were not statistically significant for oral infection or stomatitis. Although there are a number of promising indications, in view of the limited number and quality of studies and the variation in results among studies, this review highlights the need for additional well-designed trials to draw conclusions that are more robust.
Sources of Safety Data and Statistical Strategies for Design and Analysis: Clinical Trials.
Zink, Richard C; Marchenko, Olga; Sanchez-Kam, Matilde; Ma, Haijun; Jiang, Qi
2018-03-01
There has been an increased emphasis on the proactive and comprehensive evaluation of safety endpoints to ensure patient well-being throughout the medical product life cycle. In fact, depending on the severity of the underlying disease, it is important to plan for a comprehensive safety evaluation at the start of any development program. Statisticians should be intimately involved in this process and contribute their expertise to study design, safety data collection, analysis, reporting (including data visualization), and interpretation. In this manuscript, we review the challenges associated with the analysis of safety endpoints and describe the safety data that are available to influence the design and analysis of premarket clinical trials. We share our recommendations for the statistical and graphical methodologies necessary to appropriately analyze, report, and interpret safety outcomes, and we discuss the advantages and disadvantages of safety data obtained from clinical trials compared to other sources. Clinical trials are an important source of safety data that contribute to the totality of safety information available to generate evidence for regulators, sponsors, payers, physicians, and patients. This work is a result of the efforts of the American Statistical Association Biopharmaceutical Section Safety Working Group.
Water quality analysis of the Rapur area, Andhra Pradesh, South India using multivariate techniques
NASA Astrophysics Data System (ADS)
Nagaraju, A.; Sreedhar, Y.; Thejaswi, A.; Sayadi, Mohammad Hossein
2017-10-01
The groundwater samples from Rapur area were collected from different sites to evaluate the major ion chemistry. The large number of data can lead to difficulties in the integration, interpretation, and representation of the results. Two multivariate statistical methods, hierarchical cluster analysis (HCA) and factor analysis (FA), were applied to evaluate their usefulness to classify and identify geochemical processes controlling groundwater geochemistry. Four statistically significant clusters were obtained from 30 sampling stations. This has resulted two important clusters viz., cluster 1 (pH, Si, CO3, Mg, SO4, Ca, K, HCO3, alkalinity, Na, Na + K, Cl, and hardness) and cluster 2 (EC and TDS) which are released to the study area from different sources. The application of different multivariate statistical techniques, such as principal component analysis (PCA), assists in the interpretation of complex data matrices for a better understanding of water quality of a study area. From PCA, it is clear that the first factor (factor 1), accounted for 36.2% of the total variance, was high positive loading in EC, Mg, Cl, TDS, and hardness. Based on the PCA scores, four significant cluster groups of sampling locations were detected on the basis of similarity of their water quality.
NASA Astrophysics Data System (ADS)
Skorobogatiy, Maksim; Sadasivan, Jayesh; Guerboukha, Hichem
2018-05-01
In this paper, we first discuss the main types of noise in a typical pump-probe system, and then focus specifically on terahertz time domain spectroscopy (THz-TDS) setups. We then introduce four statistical models for the noisy pulses obtained in such systems, and detail rigorous mathematical algorithms to de-noise such traces, find the proper averages and characterise various types of experimental noise. Finally, we perform a comparative analysis of the performance, advantages and limitations of the algorithms by testing them on the experimental data collected using a particular THz-TDS system available in our laboratories. We conclude that using advanced statistical models for trace averaging results in the fitting errors that are significantly smaller than those obtained when only a simple statistical average is used.
A statistical metadata model for clinical trials' data management.
Vardaki, Maria; Papageorgiou, Haralambos; Pentaris, Fragkiskos
2009-08-01
We introduce a statistical, process-oriented metadata model to describe the process of medical research data collection, management, results analysis and dissemination. Our approach explicitly provides a structure for pieces of information used in Clinical Study Data Management Systems, enabling a more active role for any associated metadata. Using the object-oriented paradigm, we describe the classes of our model that participate during the design of a clinical trial and the subsequent collection and management of the relevant data. The advantage of our approach is that we focus on presenting the structural inter-relation of these classes when used during datasets manipulation by proposing certain transformations that model the simultaneous processing of both data and metadata. Our solution reduces the possibility of human errors and allows for the tracking of all changes made during datasets lifecycle. The explicit modeling of processing steps improves data quality and assists in the problem of handling data collected in different clinical trials. The case study illustrates the applicability of the proposed framework demonstrating conceptually the simultaneous handling of datasets collected during two randomized clinical studies. Finally, we provide the main considerations for implementing the proposed framework into a modern Metadata-enabled Information System.
Johnson, Gregory R.; Kangas, Joshua D.; Dovzhenko, Alexander; Trojok, Rüdiger; Voigt, Karsten; Majarian, Timothy D.; Palme, Klaus; Murphy, Robert F.
2017-01-01
Quantitative image analysis procedures are necessary for the automated discovery of effects of drug treatment in large collections of fluorescent micrographs. When compared to their mammalian counterparts, the effects of drug conditions on protein localization in plant species are poorly understood and underexplored. To investigate this relationship, we generated a large collection of images of single plant cells after various drug treatments. For this, protoplasts were isolated from six transgenic lines of A. thaliana expressing fluorescently tagged proteins. Nine drugs at three concentrations were applied to protoplast cultures followed by automated image acquisition. For image analysis, we developed a cell segmentation protocol for detecting drug effects using a Hough-transform based region of interest detector and a novel cross-channel texture feature descriptor. In order to determine treatment effects, we summarized differences between treated and untreated experiments with an L1 Cramér-von Mises statistic. The distribution of these statistics across all pairs of treated and untreated replicates was compared to the variation within control replicates to determine the statistical significance of observed effects. Using this pipeline, we report the dose dependent drug effects in the first high-content Arabidopsis thaliana drug screen of its kind. These results can function as a baseline for comparison to other protein organization modeling approaches in plant cells. PMID:28245335
Bellenguez, Céline; Strange, Amy; Freeman, Colin; Donnelly, Peter; Spencer, Chris C A
2012-01-01
High-throughput genotyping arrays provide an efficient way to survey single nucleotide polymorphisms (SNPs) across the genome in large numbers of individuals. Downstream analysis of the data, for example in genome-wide association studies (GWAS), often involves statistical models of genotype frequencies across individuals. The complexities of the sample collection process and the potential for errors in the experimental assay can lead to biases and artefacts in an individual's inferred genotypes. Rather than attempting to model these complications, it has become a standard practice to remove individuals whose genome-wide data differ from the sample at large. Here we describe a simple, but robust, statistical algorithm to identify samples with atypical summaries of genome-wide variation. Its use as a semi-automated quality control tool is demonstrated using several summary statistics, selected to identify different potential problems, and it is applied to two different genotyping platforms and sample collections. The algorithm is written in R and is freely available at www.well.ox.ac.uk/chris-spencer chris.spencer@well.ox.ac.uk Supplementary data are available at Bioinformatics online.
Ichthyoplankton abundance and variance in a large river system concerns for long-term monitoring
Holland-Bartels, Leslie E.; Dewey, Michael R.; Zigler, Steven J.
1995-01-01
System-wide spatial patterns of ichthyoplankton abundance and variability were assessed in the upper Mississippi and lower Illinois rivers to address the experimental design and statistical confidence in density estimates. Ichthyoplankton was sampled from June to August 1989 in primary milieus (vegetated and non-vegated backwaters and impounded areas, main channels and main channel borders) in three navigation pools (8, 13 and 26) of the upper Mississippi River and in a downstream reach of the Illinois River. Ichthyoplankton densities varied among stations of similar aquatic landscapes (milieus) more than among subsamples within a station. An analysis of sampling effort indicated that the collection of single samples at many stations in a given milieu type is statistically and economically preferable to the collection of multiple subsamples at fewer stations. Cluster analyses also revealed that stations only generally grouped by their preassigned milieu types. Pilot studies such as this can define station groupings and sources of variation beyond an a priori habitat classification. Thus the minimum intensity of sampling required to achieve a desired statistical confidence can be identified before implementing monitoring efforts.
Adams, James; Kruger, Uwe; Geis, Elizabeth; Gehn, Eva; Fimbres, Valeria; Pollard, Elena; Mitchell, Jessica; Ingram, Julie; Hellmers, Robert; Quig, David; Hahn, Juergen
2017-01-01
Introduction A number of previous studies examined a possible association of toxic metals and autism, and over half of those studies suggest that toxic metal levels are different in individuals with Autism Spectrum Disorders (ASD). Additionally, several studies found that those levels correlate with the severity of ASD. Methods In order to further investigate these points, this paper performs the most detailed statistical analysis to date of a data set in this field. First morning urine samples were collected from 67 children and adults with ASD and 50 neurotypical controls of similar age and gender. The samples were analyzed to determine the levels of 10 urinary toxic metals (UTM). Autism-related symptoms were assessed with eleven behavioral measures. Statistical analysis was used to distinguish participants on the ASD spectrum and neurotypical participants based upon the UTM data alone. The analysis also included examining the association of autism severity with toxic metal excretion data using linear and nonlinear analysis. “Leave-one-out” cross-validation was used to ensure statistical independence of results. Results and Discussion Average excretion levels of several toxic metals (lead, tin, thallium, antimony) were significantly higher in the ASD group. However, ASD classification using univariate statistics proved difficult due to large variability, but nonlinear multivariate statistical analysis significantly improved ASD classification with Type I/II errors of 15% and 18%, respectively. These results clearly indicate that the urinary toxic metal excretion profiles of participants in the ASD group were significantly different from those of the neurotypical participants. Similarly, nonlinear methods determined a significantly stronger association between the behavioral measures and toxic metal excretion. The association was strongest for the Aberrant Behavior Checklist (including subscales on Irritability, Stereotypy, Hyperactivity, and Inappropriate Speech), but significant associations were found for UTM with all eleven autism-related assessments with cross-validation R2 values ranging from 0.12–0.48. PMID:28068407
Accounting for measurement error: a critical but often overlooked process.
Harris, Edward F; Smith, Richard N
2009-12-01
Due to instrument imprecision and human inconsistencies, measurements are not free of error. Technical error of measurement (TEM) is the variability encountered between dimensions when the same specimens are measured at multiple sessions. A goal of a data collection regimen is to minimise TEM. The few studies that actually quantify TEM, regardless of discipline, report that it is substantial and can affect results and inferences. This paper reviews some statistical approaches for identifying and controlling TEM. Statistically, TEM is part of the residual ('unexplained') variance in a statistical test, so accounting for TEM, which requires repeated measurements, enhances the chances of finding a statistically significant difference if one exists. The aim of this paper was to review and discuss common statistical designs relating to types of error and statistical approaches to error accountability. This paper addresses issues of landmark location, validity, technical and systematic error, analysis of variance, scaled measures and correlation coefficients in order to guide the reader towards correct identification of true experimental differences. Researchers commonly infer characteristics about populations from comparatively restricted study samples. Most inferences are statistical and, aside from concerns about adequate accounting for known sources of variation with the research design, an important source of variability is measurement error. Variability in locating landmarks that define variables is obvious in odontometrics, cephalometrics and anthropometry, but the same concerns about measurement accuracy and precision extend to all disciplines. With increasing accessibility to computer-assisted methods of data collection, the ease of incorporating repeated measures into statistical designs has improved. Accounting for this technical source of variation increases the chance of finding biologically true differences when they exist.
78 FR 17920 - Notice of Intent to Seek Reinstatement of an Information Collection
Federal Register 2010, 2011, 2012, 2013, 2014
2013-03-25
... DEPARTMENT OF AGRICULTURE National Agricultural Statistics Service Notice of Intent to Seek Reinstatement of an Information Collection AGENCY: National Agricultural Statistics Service, USDA. ACTION... notice announces the intention of the National Agricultural Statistics Service (NASS) to seek...
77 FR 17405 - Notice of Intent To Revise a Previously Approved Information Collection
Federal Register 2010, 2011, 2012, 2013, 2014
2012-03-26
... DEPARTMENT OF AGRICULTURE National Agricultural Statistics Service Notice of Intent To Revise a Previously Approved Information Collection AGENCY: National Agricultural Statistics Service, USDA. ACTION... notice announces the intent of the National Agricultural Statistics Service (NASS) to seek reinstatement...
76 FR 45505 - Notice of Intent To Revise a Currently Approved Information Collection
Federal Register 2010, 2011, 2012, 2013, 2014
2011-07-29
... cooperative agreement between the Center for Disease Control (CDC) and the National Agricultural Statistics... DEPARTMENT OF AGRICULTURE National Agricultural Statistics Service Notice of Intent To Revise a Currently Approved Information Collection AGENCY: National Agricultural Statistics Service, USDA. ACTION...
33 CFR 207.800 - Collection of navigation statistics.
Code of Federal Regulations, 2013 CFR
2013-07-01
... statistics. 207.800 Section 207.800 Navigation and Navigable Waters CORPS OF ENGINEERS, DEPARTMENT OF THE ARMY, DEPARTMENT OF DEFENSE NAVIGATION REGULATIONS § 207.800 Collection of navigation statistics. (a... Revenue Service, Customs Service, Maritime Administration, Department of Transportation, and Department of...
33 CFR 207.800 - Collection of navigation statistics.
Code of Federal Regulations, 2014 CFR
2014-07-01
... statistics. 207.800 Section 207.800 Navigation and Navigable Waters CORPS OF ENGINEERS, DEPARTMENT OF THE ARMY, DEPARTMENT OF DEFENSE NAVIGATION REGULATIONS § 207.800 Collection of navigation statistics. (a... Revenue Service, Customs Service, Maritime Administration, Department of Transportation, and Department of...
33 CFR 207.800 - Collection of navigation statistics.
Code of Federal Regulations, 2011 CFR
2011-07-01
... statistics. 207.800 Section 207.800 Navigation and Navigable Waters CORPS OF ENGINEERS, DEPARTMENT OF THE ARMY, DEPARTMENT OF DEFENSE NAVIGATION REGULATIONS § 207.800 Collection of navigation statistics. (a... Revenue Service, Customs Service, Maritime Administration, Department of Transportation, and Department of...
33 CFR 207.800 - Collection of navigation statistics.
Code of Federal Regulations, 2012 CFR
2012-07-01
... statistics. 207.800 Section 207.800 Navigation and Navigable Waters CORPS OF ENGINEERS, DEPARTMENT OF THE ARMY, DEPARTMENT OF DEFENSE NAVIGATION REGULATIONS § 207.800 Collection of navigation statistics. (a... Revenue Service, Customs Service, Maritime Administration, Department of Transportation, and Department of...
78 FR 70059 - Agency Information Collection Activities: Proposed Collection; Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2013-11-22
... (as opposed to quantitative statistical methods). In consultation with research experts, we have... qualitative interviews (as opposed to quantitative statistical methods). In consultation with research experts... utilization of qualitative interviews (as opposed to quantitative statistical methods). In consultation with...
76 FR 6161 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2011-02-03
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting... comments to Carol Rowan, BLS Clearance Officer, Division of Management Systems, Bureau of Labor Statistics...
77 FR 27798 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2012-05-11
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting... comments to Carol Rowan, BLS Clearance Officer, Division of Management Systems, Bureau of Labor Statistics...
76 FR 71075 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2011-11-16
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting... Statistics, Room 4080, 2 Massachusetts Avenue NE., Washington, DC 20212. Written comments also may be...
76 FR 60930 - Proposed Collection, Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2011-09-30
... DEPARTMENT OF LABOR Bureau of Labor Statistics Proposed Collection, Comment Request ACTION: Notice... requirements on respondents can be properly assessed. The Bureau of Labor Statistics (BLS) is soliciting comments concerning the proposed extension of the ``Mass Layoff Statistics Program.'' A copy of the...
Ramsthaler, F; Kreutz, K; Verhoff, M A
2007-11-01
It has been generally accepted in skeletal sex determination that the use of metric methods is limited due to the population dependence of the multivariate algorithms. The aim of the study was to verify the applicability of software-based sex estimations outside the reference population group for which discriminant equations have been developed. We examined 98 skulls from recent forensic cases of known age, sex, and Caucasian ancestry from cranium collections in Frankfurt and Mainz (Germany) to determine the accuracy of sex determination using the statistical software solution Fordisc which derives its database and functions from the US American Forensic Database. In a comparison between metric analysis using Fordisc and morphological determination of sex, average accuracy for both sexes was 86 vs 94%, respectively, and males were identified more accurately than females. The ratio of the true test result rate to the false test result rate was not statistically different for the two methodological approaches at a significance level of 0.05 but was statistically different at a level of 0.10 (p=0.06). Possible explanations for this difference comprise different ancestry, age distribution, and socio-economic status compared to the Fordisc reference sample. It is likely that a discriminant function analysis on the basis of more similar European reference samples will lead to more valid and reliable sexing results. The use of Fordisc as a single method for the estimation of sex of recent skeletal remains in Europe cannot be recommended without additional morphological assessment and without a built-in software update based on modern European reference samples.
Windshield splatter analysis with the Galaxy metagenomic pipeline
Kosakovsky Pond, Sergei; Wadhawan, Samir; Chiaromonte, Francesca; Ananda, Guruprasad; Chung, Wen-Yu; Taylor, James; Nekrutenko, Anton
2009-01-01
How many species inhabit our immediate surroundings? A straightforward collection technique suitable for answering this question is known to anyone who has ever driven a car at highway speeds. The windshield of a moving vehicle is subjected to numerous insect strikes and can be used as a collection device for representative sampling. Unfortunately the analysis of biological material collected in that manner, as with most metagenomic studies, proves to be rather demanding due to the large number of required tools and considerable computational infrastructure. In this study, we use organic matter collected by a moving vehicle to design and test a comprehensive pipeline for phylogenetic profiling of metagenomic samples that includes all steps from processing and quality control of data generated by next-generation sequencing technologies to statistical analyses and data visualization. To the best of our knowledge, this is also the first publication that features a live online supplement providing access to exact analyses and workflows used in the article. PMID:19819906
Design and analysis of multiple diseases genome-wide association studies without controls.
Chen, Zhongxue; Huang, Hanwen; Ng, Hon Keung Tony
2012-11-15
In genome-wide association studies (GWAS), multiple diseases with shared controls is one of the case-control study designs. If data obtained from these studies are appropriately analyzed, this design can have several advantages such as improving statistical power in detecting associations and reducing the time and cost in the data collection process. In this paper, we propose a study design for GWAS which involves multiple diseases but without controls. We also propose corresponding statistical data analysis strategy for GWAS with multiple diseases but no controls. Through a simulation study, we show that the statistical association test with the proposed study design is more powerful than the test with single disease sharing common controls, and it has comparable power to the overall test based on the whole dataset including the controls. We also apply the proposed method to a real GWAS dataset to illustrate the methodologies and the advantages of the proposed design. Some possible limitations of this study design and testing method and their solutions are also discussed. Our findings indicate that the proposed study design and statistical analysis strategy could be more efficient than the usual case-control GWAS as well as those with shared controls. Copyright © 2012 Elsevier B.V. All rights reserved.
Becker, Betsy Jane; Aloe, Ariel M; Duvendack, Maren; Stanley, T D; Valentine, Jeffrey C; Fretheim, Atle; Tugwell, Peter
2017-09-01
To outline issues of importance to analytic approaches to the synthesis of quasi-experiments (QEs) and to provide a statistical model for use in analysis. We drew on studies of statistics, epidemiology, and social-science methodology to outline methods for synthesis of QE studies. The design and conduct of QEs, effect sizes from QEs, and moderator variables for the analysis of those effect sizes were discussed. Biases, confounding, design complexities, and comparisons across designs offer serious challenges to syntheses of QEs. Key components of meta-analyses of QEs were identified, including the aspects of QE study design to be coded and analyzed. Of utmost importance are the design and statistical controls implemented in the QEs. Such controls and any potential sources of bias and confounding must be modeled in analyses, along with aspects of the interventions and populations studied. Because of such controls, effect sizes from QEs are more complex than those from randomized experiments. A statistical meta-regression model that incorporates important features of the QEs under review was presented. Meta-analyses of QEs provide particular challenges, but thorough coding of intervention characteristics and study methods, along with careful analysis, should allow for sound inferences. Copyright © 2017 Elsevier Inc. All rights reserved.
Detection of semi-volatile organic compounds in permeable ...
Abstract The Edison Environmental Center (EEC) has a research and demonstration permeable parking lot comprised of three different permeable systems: permeable asphalt, porous concrete and interlocking concrete permeable pavers. Water quality and quantity analysis has been ongoing since January, 2010. This paper describes a subset of the water quality analysis, analysis of semivolatile organic compounds (SVOCs) to determine if hydrocarbons were in water infiltrated through the permeable surfaces. SVOCs were analyzed in samples collected from 11 dates over a 3 year period, from 2/8/2010 to 4/1/2013.Results are broadly divided into three categories: 42 chemicals were never detected; 12 chemicals (11 chemical test) were detected at a rate of less than 10% or less; and 22 chemicals were detected at a frequency of 10% or greater (ranging from 10% to 66.5% detections). Fundamental and exploratory statistical analyses were performed on these latter analyses results by grouping results by surface type. The statistical analyses were limited due to low frequency of detections and dilutions of samples which impacted detection limits. The infiltrate data through three permeable surfaces were analyzed as non-parametric data by the Kaplan-Meier estimation method for fundamental statistics; there were some statistically observable difference in concentration between pavement types when using Tarone-Ware Comparison Hypothesis Test. Additionally Spearman Rank order non-parame
NASA Astrophysics Data System (ADS)
Chakraborthy, Parthasarathi; Chattopadhyay, Surajit
2013-02-01
Endeavor of the present paper is to investigate the statistical properties of the total ozone concentration time series over Arosa, Switzerland (9.68°E, 46.78°N). For this purpose, different statistical data analysis procedures have been employed for analyzing the mean monthly total ozone concentration data, collected over a period of 40 years (1932-1971), at the above location. Based on the computations on the available data set, the study reports different degrees of variations in different months. The month of July is reported as the month of lowest variability. April and May are found to be the most correlated months with respect to total ozone concentration.
Statistical analysis of field data for aircraft warranties
NASA Astrophysics Data System (ADS)
Lakey, Mary J.
Air Force and Navy maintenance data collection systems were researched to determine their scientific applicability to the warranty process. New and unique algorithms were developed to extract failure distributions which were then used to characterize how selected families of equipment typically fails. Families of similar equipment were identified in terms of function, technology and failure patterns. Statistical analyses and applications such as goodness-of-fit test, maximum likelihood estimation and derivation of confidence intervals for the probability density function parameters were applied to characterize the distributions and their failure patterns. Statistical and reliability theory, with relevance to equipment design and operational failures were also determining factors in characterizing the failure patterns of the equipment families. Inferences about the families with relevance to warranty needs were then made.
1992-04-01
contractor’s existing data collection, analysis and corrective action system shall be utilized, with modification only as necessary to meet the...either from test or from analysis of field data . The procedures of MIL-STD-756B assume that the reliability of a 18 DEFINE IDENTIFY SOFTWARE LIFE CYCLE...to generate sufficient data to report a statistically valid reliability figure for a class of software. Casual data gathering accumulates data more
Breast Density Assessment by Dual Energy X-ray Absorptiometry in Women and Girls
2008-07-01
daughter pairs next year. Task 6. Data Management and Analysis. All current participants’’ data collected during the study visits have been...projects although in some woman the areola continues to form a secondary mound. Statistical Analysis. All data management and anal- yses were done using the...from the viewpoint of qaulity control. Washington (DC): Graduate School of the Department of Agriculture; 1939. 19. Irwin ML, Aiello EJ, McTiernan A
Human Systems Engineering and Program Success - A Retrospective Content Analysis
2016-01-01
collected from the 546 documents and entered into SPSS Statistics Version 22.0 for Windows. HSI words within the sampled doc- uments ranged from zero to...engineers. The approach used a retrospective content analysis of documents from weapon systems acquisi- tion programs, namely Major Defense Acquisition...January 2016, Vol. 23 No. 1 : 78–101 January 2016 The interaction between humans and the systems they use affects program success, as well as life-cycle
Acoustic fill factors for a 120 inch diameter fairing
NASA Technical Reports Server (NTRS)
Lee, Y. Albert
1992-01-01
Data from the acoustic test of a 120-inch diameter payload fairing were collected and an analysis of acoustic fill factors were performed. Correction factors for obtaining a weighted spatial average of the interior sound pressure level (SPL) were derived based on this database and a normalized 200-inch diameter fairing database. The weighted fill factors were determined and compared with statistical energy analysis (VAPEPS code) derived fill factors. The comparison is found to be reasonable.
1998-01-01
Ferrography on High Performance Aircraft Engine Lubricating Oils Allison M. Toms, Sharon 0. Hem, Tim Yarborough Joint Oil Analysis Program Technical...turbine engines by spectroscopy (AES and FT-IR) and direct reading and analytical ferrography . A statistical analysis of the data collected is...presented. Key Words: Analytical ferrography ; atomic emission spectroscopy; condition monitoring; direct reading ferrography ; Fourier transform infrared
Integrated Data Collection Analysis (IDCA) Program - Statistical Analysis of RDX Standard Data Sets
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sandstrom, Mary M.; Brown, Geoffrey W.; Preston, Daniel N.
2015-10-30
The Integrated Data Collection Analysis (IDCA) program is conducting a Proficiency Test for Small- Scale Safety and Thermal (SSST) testing of homemade explosives (HMEs). Described here are statistical analyses of the results for impact, friction, electrostatic discharge, and differential scanning calorimetry analysis of the RDX Type II Class 5 standard. The material was tested as a well-characterized standard several times during the proficiency study to assess differences among participants and the range of results that may arise for well-behaved explosive materials. The analyses show that there are detectable differences among the results from IDCA participants. While these differences are statisticallymore » significant, most of them can be disregarded for comparison purposes to assess potential variability when laboratories attempt to measure identical samples using methods assumed to be nominally the same. The results presented in this report include the average sensitivity results for the IDCA participants and the ranges of values obtained. The ranges represent variation about the mean values of the tests of between 26% and 42%. The magnitude of this variation is attributed to differences in operator, method, and environment as well as the use of different instruments that are also of varying age. The results appear to be a good representation of the broader safety testing community based on the range of methods, instruments, and environments included in the IDCA Proficiency Test.« less
75 FR 1410 - Agency Information Collection Activities: Existing Collection; Comments Requested
Federal Register 2010, 2011, 2012, 2013, 2014
2010-01-11
... DEPARTMENT OF JUSTICE Bureau of Justice Statistics [OMB Number 1121-0094] Agency Information... of Justice (DOJ), Bureau of Justice Statistics, will be submitting the following information... contact Todd D. Minton, Bureau of Justice Statistics, 810 Seventh Street, NW., Washington, DC 20531 (phone...
Nateglinide versus repaglinide for type 2 diabetes mellitus in China.
Li, Chanjuan; Xia, Jielai; Zhang, Gaokui; Wang, Suzhen; Wang, Ling
2009-12-01
The purpose of this study is to evaluate efficacy and safety of nateglinide tablet administration in comparison with those of repaglinide tablet as control on treating type 2 diabetes mellitus in China. Pooled-analysis with analysis of covariance (ANCOVA) method was applied to assess the efficacy and safety based on original data collected from four independent randomized clinical trials with similar research protocols. However meta-analysis was applied based on the outcomes of the four studies. The results by meta-analysis were comparable to those obtained by pooled-analysis. The means of HbA(1c), and fasting blood glucose in both the nateglinide and repaglinide groups were reduced significantly after 12 weeks duration but no statistical differences in reduction between the two groups. The adverse reaction rates were 9.89 and 6.51% in the nateglinide and repaglinide groups respectively, with the rate difference showing no statistical significance, and the Odds Ratio of adverse reaction rate (95% confidence interval) was 1.59 (0.99, 2.55). Both nateglinide and repaglinide administration have similarly significant effects on reducing HbA(1c) and FBG. However, the adverse reaction rate in the nateglinide group is higher than that in the latter using repaglinide but no statistical significance difference as revealed in the four clinical trials detailed below.
Making the Grade: A Report on Standards in Work-Based Learning for Young People.
ERIC Educational Resources Information Center
Hughes, Maria
The reasons for the deterioration of the inspection grades awarded for work-based learning (WBL) provision in England were examined. The main data collection activities were as follows: (1) a statistical analysis of the Training Standards Council (TSC) and Adult Learning Inspectorate (ALI) inspection grades in 1998-2001; (2) a qualitative review…
2012-06-01
generalized behavioral model characterized after the fictional Seldon equations (the one elaborated upon by Isaac Asimov in the 1951 novel, The...Foundation). Asimov described the Seldon equations as essentially statistical models with historical data of a sufficient size and variability that they
Home Safety, Safe Behaviors of Elderly People, and Fall Accidents At Home
ERIC Educational Resources Information Center
Erkal, Sibel
2010-01-01
The present study analyzed home safety and safe behaviors against fall accidents of elderly people living at home. The study group comprised 121 people aged 65+ living in the catchment area of Ankara Mamak Halil Ulgen Health Center. Data were collected via a personal information form and Home-Screen Scale. Statistical analysis used an independent…
ERIC Educational Resources Information Center
Aderibigbe, Semiyu Adejare; Ajasa, Folorunso Adekemi
2013-01-01
Purpose: The purpose of this paper is to explore the perceptions of college tutors on peer coaching as a tool for professional development to determine its formal institutionalisation. Design/methodology/approach: A survey questionnaire was used for data collection, while analysis of data was done using descriptive statistics. Findings: The…
The Effect of Automobile Safety on Vehicle Type Choice: An Empirical Study.
ERIC Educational Resources Information Center
McCarthy, Patrick S.
An analysis was made of the extent to which the safety characteristics of new vehicles affect consumer purchase decisions. Using an extensive data set that combines vehicle data collected by the Automobile Club of Southern California Target Car Program with the responses from a national household survey of new car buyers, a statistical model of…
ERIC Educational Resources Information Center
General Accounting Office, Washington, DC. Div. of Human Resources.
In response to Congressional requests, this report describes the extent of Hispanic American health and nutrition data available from federal sources. Oversampling of a minority group by a national survey is necessary for valid analysis of group characteristics. Among the four national health and nutrition surveys conducted by the Department of…
Time-dynamics of the two-color emission from vertical-external-cavity surface-emitting lasers
NASA Astrophysics Data System (ADS)
Chernikov, A.; Wichmann, M.; Shakfa, M. K.; Scheller, M.; Moloney, J. V.; Koch, S. W.; Koch, M.
2012-01-01
The temporal stability of a two-color vertical-external-cavity surface-emitting laser is studied using single-shot streak-camera measurements. The collected data is evaluated via quantitative statistical analysis schemes. Dynamically stable and unstable regions for the two-color operation are identified and the dependence on the pump conditions is analyzed.
ERIC Educational Resources Information Center
Sherman, Joel D.; Gregory, Barbra; Poirier, Jeffrey M.
This report is an annual collection of school district financial data. Specifically, this report presents analyses of school district revenues from the 1997-98 school year. The report is designed to address the following questions about the financing of public elementary and secondary education at the state and district levels: How much money per…
ERIC Educational Resources Information Center
Ram, Rati
Educational expenditures in 18 Organisation for Economic Cooperation and Development (OECD) countries for the years 1975 and 1985 are investigated in this report. Data collection is based on analysis of UNESCO's 1989 "Statistical Yearbook" and OECD data. Although data deficiencies allow only a broad assessment, a conclusion is that…
ERIC Educational Resources Information Center
Seng, Ernest Lim Kok; Ling, Tan Pei
2013-01-01
This study aims to investigate student satisfaction on quality education services provided by institutions of higher learning in Malaysia. Their level of satisfaction based primarily on the data collected through five dimensions of education service quality. A random sample of 250 students studying in an institution of higher learning was selected…
Institute for Training Minority Group Research and Evaluation Specialists. Final Report.
ERIC Educational Resources Information Center
Brown, Roscoe C., Jr.
The Institute for Training Minority Group Research and Evaluation Specialists comprised 4 programs in 1: (1) a 6-week graduate course at New York University (NYU) during the 1970 summer session for 20 minority group persons that provided training in research design, statistics, data collection and analysis, and report writing; (2) a program of…
Fragmentation statistics for FIA: designing an approach
Rachel Riemann; Andrew Lister; Michael Hoppus; Tonya Lister
2002-01-01
The USDA Forest Inventory and Analysis (FIA) program collects data on the amount of forest, as well as on characteristics such as forest type, tree volume, species composition, and size and age classes. However, little data are obtained nationwide on forest fragmentation-how that forest is distributed and in what land use/land cover context-factors that can...
ERIC Educational Resources Information Center
Burns, Jim A.; King, Richard
An international team of researchers studied the following aspects of training in the United Kingdom's food and beverage sector: structure and characteristics, business and social context, training and recruitment, and future training requirements. Data were collected from an analysis of social and labor/employment statistics, literature review,…
ERIC Educational Resources Information Center
Library Research Service, 2010
2010-01-01
Issues 283 through 289 of "Fast Facts" from the Library Research Service present data collected from libraries in Colorado and throughout the nation. Topics addressed in these "Fast Facts" from 2010 include the relationship between computer access in libraries and use of traditional services, analysis of the third year of data…
Dodge, Kent A.; Hornberger, Michelle I.; Turner, Matthew A.
2017-01-19
Water, bed sediment, and biota were sampled in selected streams from Butte to near Missoula, Montana, as part of a monitoring program in the upper Clark Fork Basin of western Montana. The sampling program was led by the U.S. Geological Survey, in cooperation with the U.S. Environmental Protection Agency, to characterize aquatic resources in the Clark Fork Basin, with emphasis on trace elements associated with historic mining and smelting activities. Sampling sites were located on the Clark Fork and selected tributaries. Water samples were collected periodically at 20 sites from October 2014 through September 2015. Bed-sediment and biota samples were collected once at 13 sites during August 2015.This report presents the analytical results and quality-assurance data for water-quality, bed-sediment, and biota samples collected at sites from October 2014 through September 2015. Water-quality data include concentrations of selected major ions, trace elements, and suspended sediment. At 12 sites, samples for analysis of dissolved organic carbon and turbidity were collected. In addition, samples for analysis of nitrogen (nitrate plus nitrite) were collected at two sites. Daily values of mean suspended-sediment concentration and suspended-sediment discharge were determined for three sites. Seasonal daily values of turbidity were determined for four sites. Bed-sediment data include trace-element concentrations in the fine-grained fraction. Biological data include trace-element concentrations in whole-body tissue of aquatic benthic insects. Statistical summaries of water-quality, bed-sediment, and biological data for sites in the upper Clark Fork Basin are provided for the period of record.
Statistical analysis of excitation energies in actinide and rare-earth nuclei
NASA Astrophysics Data System (ADS)
Levon, A. I.; Magner, A. G.; Radionov, S. V.
2018-04-01
Statistical analysis of distributions of the collective states in actinide and rare-earth nuclei is performed in terms of the nearest-neighbor spacing distribution (NNSD). Several approximations, such as the linear approach to the level repulsion density and that suggested by Brody to the NNSDs were applied for the analysis. We found an intermediate character of the experimental spectra between the order and the chaos for a number of rare-earth and actinide nuclei. The spectra are closer to the Wigner distribution for energies limited by 3 MeV, and to the Poisson distribution for data including higher excitation energies and higher spins. The latter result is in agreement with the theoretical calculations. These features are confirmed by the cumulative distributions, where the Wigner contribution dominates at smaller spacings while the Poisson one is more important at larger spacings, and our linear approach improves the comparison with experimental data at all desired spacings.
Sando, Steven K.; Clark, Melanie L.; Cleasby, Thomas E.; Barnhart, Elliott P.
2015-01-01
Trend results for sites in the Tenmile Creek watershed generally are more variable and difficult to interpret than for sites in the Boulder River watershed. Trend results for Tenmile Creek above City Diversion (site 11) and Minnehaha Creek near Rimini (site 12) for water years 2000–13 indicate decreasing trends in FACs of cadmium, copper, and zinc. The magnitudes of the decreasing trends in FACs of copper generally are moderate and statistically significant for sites 11 and 12. The magnitudes of the decreasing trends in FACs of cadmium and zinc for site 11 are minor to small and not statistically significant; however, the magnitudes for site 12 are moderate and statistically significant. In general, patterns in FACs for Tenmile Creek near Rimini (site 13) are not well represented by fitted trends within the short data collection period, which might indicate that the trend-analysis structure of the study is not appropriate for describing trends in FACs for site 13. The large decreasing trend in FACs of suspended sediment is the strongest indication of change in water quality during the short period of record for site 13; however, this trend is not statistically significant.
The Australasian Resuscitation in Sepsis Evaluation (ARISE) trial statistical analysis plan.
Delaney, Anthony P; Peake, Sandra L; Bellomo, Rinaldo; Cameron, Peter; Holdgate, Anna; Howe, Belinda; Higgins, Alisa; Presneill, Jeffrey; Webb, Steve
2013-09-01
The Australasian Resuscitation in Sepsis Evaluation (ARISE) study is an international, multicentre, randomised, controlled trial designed to evaluate the effectiveness of early goal-directed therapy compared with standard care for patients presenting to the emergency department with severe sepsis. In keeping with current practice, and considering aspects of trial design and reporting specific to non-pharmacological interventions, our plan outlines the principles and methods for analysing and reporting the trial results. The document is prepared before completion of recruitment into the ARISE study, without knowledge of the results of the interim analysis conducted by the data safety and monitoring committee and before completion of the two related international studies. Our statistical analysis plan was designed by the ARISE chief investigators, and reviewed and approved by the ARISE steering committee. We reviewed the data collected by the research team as specified in the study protocol and detailed in the study case report form. We describe information related to baseline characteristics, characteristics of delivery of the trial interventions, details of resuscitation, other related therapies and other relevant data with appropriate comparisons between groups. We define the primary, secondary and tertiary outcomes for the study, with description of the planned statistical analyses. We have developed a statistical analysis plan with a trial profile, mock-up tables and figures. We describe a plan for presenting baseline characteristics, microbiological and antibiotic therapy, details of the interventions, processes of care and concomitant therapies and adverse events. We describe the primary, secondary and tertiary outcomes with identification of subgroups to be analysed. We have developed a statistical analysis plan for the ARISE study, available in the public domain, before the completion of recruitment into the study. This will minimise analytical bias and conforms to current best practice in conducting clinical trials.
Scanning probe recognition microscopy investigation of tissue scaffold properties
Fan, Yuan; Chen, Qian; Ayres, Virginia M; Baczewski, Andrew D; Udpa, Lalita; Kumar, Shiva
2007-01-01
Scanning probe recognition microscopy is a new scanning probe microscopy technique which enables selective scanning along individual nanofibers within a tissue scaffold. Statistically significant data for multiple properties can be collected by repetitively fine-scanning an identical region of interest. The results of a scanning probe recognition microscopy investigation of the surface roughness and elasticity of a series of tissue scaffolds are presented. Deconvolution and statistical methods were developed and used for data accuracy along curved nanofiber surfaces. Nanofiber features were also independently analyzed using transmission electron microscopy, with results that supported the scanning probe recognition microscopy-based analysis. PMID:18203431
Scanning probe recognition microscopy investigation of tissue scaffold properties.
Fan, Yuan; Chen, Qian; Ayres, Virginia M; Baczewski, Andrew D; Udpa, Lalita; Kumar, Shiva
2007-01-01
Scanning probe recognition microscopy is a new scanning probe microscopy technique which enables selective scanning along individual nanofibers within a tissue scaffold. Statistically significant data for multiple properties can be collected by repetitively fine-scanning an identical region of interest. The results of a scanning probe recognition microscopy investigation of the surface roughness and elasticity of a series of tissue scaffolds are presented. Deconvolution and statistical methods were developed and used for data accuracy along curved nanofiber surfaces. Nanofiber features were also independently analyzed using transmission electron microscopy, with results that supported the scanning probe recognition microscopy-based analysis.
Standardized data collection to build prediction models in oncology: a prototype for rectal cancer.
Meldolesi, Elisa; van Soest, Johan; Damiani, Andrea; Dekker, Andre; Alitto, Anna Rita; Campitelli, Maura; Dinapoli, Nicola; Gatta, Roberto; Gambacorta, Maria Antonietta; Lanzotti, Vito; Lambin, Philippe; Valentini, Vincenzo
2016-01-01
The advances in diagnostic and treatment technology are responsible for a remarkable transformation in the internal medicine concept with the establishment of a new idea of personalized medicine. Inter- and intra-patient tumor heterogeneity and the clinical outcome and/or treatment's toxicity's complexity, justify the effort to develop predictive models from decision support systems. However, the number of evaluated variables coming from multiple disciplines: oncology, computer science, bioinformatics, statistics, genomics, imaging, among others could be very large thus making traditional statistical analysis difficult to exploit. Automated data-mining processes and machine learning approaches can be a solution to organize the massive amount of data, trying to unravel important interaction. The purpose of this paper is to describe the strategy to collect and analyze data properly for decision support and introduce the concept of an 'umbrella protocol' within the framework of 'rapid learning healthcare'.
75 FR 3926 - Submission for OMB Emergency Review: Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2010-01-25
... DEPARTMENT OF LABOR Bureau of Labor Statistics Submission for OMB Emergency Review: Comment.... Agency: Bureau of Labor Statistics. Type of Review: New collection. Title of Collection: Quarterly Census... appropriation tasks the Bureau of Labor Statistics (BLS) Quarterly Census of Employment and Wages (QCEW) program...
77 FR 36477 - Notice of Intent To Revise and Extend a Currently Approved Information Collection
Federal Register 2010, 2011, 2012, 2013, 2014
2012-06-19
... DEPARTMENT OF AGRICULTURE National Agricultural Statistics Service Notice of Intent To Revise and Extend a Currently Approved Information Collection AGENCY: National Agricultural Statistics Service, USDA... Reduction Act of 1995 this notice announces the intention of the National Agricultural Statistics Service...
Automated Cognitive Health Assessment From Smart Home-Based Behavior Data.
Dawadi, Prafulla Nath; Cook, Diane Joyce; Schmitter-Edgecombe, Maureen
2016-07-01
Smart home technologies offer potential benefits for assisting clinicians by automating health monitoring and well-being assessment. In this paper, we examine the actual benefits of smart home-based analysis by monitoring daily behavior in the home and predicting clinical scores of the residents. To accomplish this goal, we propose a clinical assessment using activity behavior (CAAB) approach to model a smart home resident's daily behavior and predict the corresponding clinical scores. CAAB uses statistical features that describe characteristics of a resident's daily activity performance to train machine learning algorithms that predict the clinical scores. We evaluate the performance of CAAB utilizing smart home sensor data collected from 18 smart homes over two years. We obtain a statistically significant correlation ( r=0.72) between CAAB-predicted and clinician-provided cognitive scores and a statistically significant correlation ( r=0.45) between CAAB-predicted and clinician-provided mobility scores. These prediction results suggest that it is feasible to predict clinical scores using smart home sensor data and learning-based data analysis.
A Flexible Approach for the Statistical Visualization of Ensemble Data
DOE Office of Scientific and Technical Information (OSTI.GOV)
Potter, K.; Wilson, A.; Bremer, P.
2009-09-29
Scientists are increasingly moving towards ensemble data sets to explore relationships present in dynamic systems. Ensemble data sets combine spatio-temporal simulation results generated using multiple numerical models, sampled input conditions and perturbed parameters. While ensemble data sets are a powerful tool for mitigating uncertainty, they pose significant visualization and analysis challenges due to their complexity. We present a collection of overview and statistical displays linked through a high level of interactivity to provide a framework for gaining key scientific insight into the distribution of the simulation results as well as the uncertainty associated with the data. In contrast to methodsmore » that present large amounts of diverse information in a single display, we argue that combining multiple linked statistical displays yields a clearer presentation of the data and facilitates a greater level of visual data analysis. We demonstrate this approach using driving problems from climate modeling and meteorology and discuss generalizations to other fields.« less
Miyagi, Atsushi
2017-09-01
Detailed exploration of sensory perception as well as preference across gender and age for a certain food is very useful for developing a vendible food commodity related to physiological and psychological motivation for food preference. Sensory tests including color, sweetness, bitterness, fried peanut aroma, textural preference and overall liking of deep-fried peanuts with varying frying time (2, 4, 6, 9, 12 and 15 min) at 150 °C were carried out using 417 healthy Japanese consumers. To determine the influence of gender and age on sensory evaluation, systematic statistical analysis including one-way analysis of variance, polynomial regression analysis and multiple regression analysis was conducted using the collected data. The results indicated that females were more sensitive to bitterness than males. This may affect sensory preference; female subjects favored peanuts prepared with a shorter frying time more than male subjects did. With advancing age, textural preference played a more important role in overall preference. Older subjects liked deeper-fried peanuts, which are more brittle, more than younger subjects did. In the present study, systematic statistical analysis based on collected sensory evaluation data using deep-fried peanuts was conducted and the tendency of sensory perception and preference across gender and age was clarified. These results may be useful for engineering optimal strategies to target specific segments to gain greater acceptance in the market. © 2017 Society of Chemical Industry. © 2017 Society of Chemical Industry.
Statistical assessment on a combined analysis of GRYN-ROMN-UCBN upland vegetation vital signs
Irvine, Kathryn M.; Rodhouse, Thomas J.
2014-01-01
As of 2013, Rocky Mountain and Upper Columbia Basin Inventory and Monitoring Networks have multiple years of vegetation data and Greater Yellowstone Network has three years of vegetation data and monitoring is ongoing in all three networks. Our primary objective is to assess whether a combined analysis of these data aimed at exploring correlations with climate and weather data is feasible. We summarize the core survey design elements across protocols and point out the major statistical challenges for a combined analysis at present. The dissimilarity in response designs between ROMN and UCBN-GRYN network protocols presents a statistical challenge that has not been resolved yet. However, the UCBN and GRYN data are compatible as they implement a similar response design; therefore, a combined analysis is feasible and will be pursued in future. When data collected by different networks are combined, the survey design describing the merged dataset is (likely) a complex survey design. A complex survey design is the result of combining datasets from different sampling designs. A complex survey design is characterized by unequal probability sampling, varying stratification, and clustering (see Lohr 2010 Chapter 7 for general overview). Statistical analysis of complex survey data requires modifications to standard methods, one of which is to include survey design weights within a statistical model. We focus on this issue for a combined analysis of upland vegetation from these networks, leaving other topics for future research. We conduct a simulation study on the possible effects of equal versus unequal probability selection of points on parameter estimates of temporal trend using available packages within the R statistical computing package. We find that, as written, using lmer or lm for trend detection in a continuous response and clm and clmm for visually estimated cover classes with “raw” GRTS design weights specified for the weight argument leads to substantially different results and/or computational instability. However, when only fixed effects are of interest, the survey package (svyglm and svyolr) may be suitable for a model-assisted analysis for trend. We provide possible directions for future research into combined analysis for ordinal and continuous vital sign indictors.
a Comparative Analysis of Five Cropland Datasets in Africa
NASA Astrophysics Data System (ADS)
Wei, Y.; Lu, M.; Wu, W.
2018-04-01
The food security, particularly in Africa, is a challenge to be resolved. The cropland area and spatial distribution obtained from remote sensing imagery are vital information. In this paper, according to cropland area and spatial location, we compare five global cropland datasets including CCI Land Cover, GlobCover, MODIS Collection 5, GlobeLand30 and Unified Cropland in circa 2010 of Africa in terms of cropland area and spatial location. The accuracy of cropland area calculated from five datasets was analyzed compared with statistic data. Based on validation samples, the accuracies of spatial location for the five cropland products were assessed by error matrix. The results show that GlobeLand30 has the best fitness with the statistics, followed by MODIS Collection 5 and Unified Cropland, GlobCover and CCI Land Cover have the lower accuracies. For the accuracy of spatial location of cropland, GlobeLand30 reaches the highest accuracy, followed by Unified Cropland, MODIS Collection 5 and GlobCover, CCI Land Cover has the lowest accuracy. The spatial location accuracy of five datasets in the Csa with suitable farming condition is generally higher than in the Bsk.
Uncovering stable and occasional human mobility patterns: A case study of the Beijing subway
NASA Astrophysics Data System (ADS)
Yong, Nuo; Ni, Shunjiang; Shen, Shifei; Chen, Peng; Ji, Xuewei
2018-02-01
There have generally been two kinds of approaches to the empirical study of human mobility. At the group level, some valuable information might be submerged in statistical noise, while due to the diversity of individual purpose and preference, there is still no general statistical regularity of human mobility at the individual level. In this paper, we considered group-level human mobility as the combination of several basic patterns and analyzed the collective mobility by category. Utilizing matrix factorization and correlation analysis, we extracted some of the stable/occasional components from the collective human mobility in the Beijing subway and found that the departure and arrival mobility patterns have different characteristics, both in time and space, under various conditions. We classified individual records into different patterns and analyzed the most likely trip distance by category. The proposed method can decompose stable/occasional mobility patterns from the collective mobility and identify passengers belonging to different patterns, helping us to better understand the origin of different mobility patterns and provide guidance for emergency management of large crowds.
Onsite Gaseous Centrifuge Enrichment Plant UF6 Cylinder Destructive Analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Anheier, Norman C.; Cannon, Bret D.; Qiao, Hong
2012-07-17
The IAEA safeguards approach for gaseous centrifuge enrichment plants (GCEPs) includes measurements of gross, partial, and bias defects in a statistical sampling plan. These safeguard methods consist principally of mass and enrichment nondestructive assay (NDA) verification. Destructive assay (DA) samples are collected from a limited number of cylinders for high precision offsite mass spectrometer analysis. DA is typically used to quantify bias defects in the GCEP material balance. Under current safeguards measures, the operator collects a DA sample from a sample tap following homogenization. The sample is collected in a small UF6 sample bottle, then sealed and shipped under IAEAmore » chain of custody to an offsite analytical laboratory. Current practice is expensive and resource intensive. We propose a new and novel approach for performing onsite gaseous UF6 DA analysis that provides rapid and accurate assessment of enrichment bias defects. DA samples are collected using a custom sampling device attached to a conventional sample tap. A few micrograms of gaseous UF6 is chemically adsorbed onto a sampling coupon in a matter of minutes. The collected DA sample is then analyzed onsite using Laser Ablation Absorption Ratio Spectrometry-Destructive Assay (LAARS-DA). DA results are determined in a matter of minutes at sufficient accuracy to support reliable bias defect conclusions, while greatly reducing DA sample volume, analysis time, and cost.« less
Chhabra, Anmol; Quinn, Andrea; Ries, Amanda
2018-01-01
Accurate history collection is integral to medication reconciliation. Studies support pharmacy involvement in the process, but assessment of global time spent is limited. The authors hypothesized the location of a medication-focused interview would impact time spent. The objective was to compare time spent by pharmacists and nurses based on the location of a medication-focused interview. Time spent by the interviewing pharmacist, admitting nurse, and centralized pharmacist verifying admission orders was collected. Patient groups were based on whether the interview was conducted in the emergency department (ED) or medical floor. The primary end point was a composite of the 3 time points. Secondary end points were individual time components and number and types of transcription discrepancies identified during medical floor interviews. Pharmacists and nurses spent an average of ten fewer minutes per ED patient versus a medical floor patient ( P = .028). Secondary end points were not statistically significant. Transcription discrepancies were identified at a rate of 1 in 4 medications. Post hoc analysis revealed the time spent by pharmacists and nurses was 2.4 minutes shorter per medication when interviewed in the ED ( P < .001). The primary outcome was statistically and clinically significant. Limitations included inability to blind and lack of cost-saving analysis. Pharmacist involvement in ED medication reconciliation leads to time savings during the admission process.
Analysis of pediatric blood lead levels in New York City for 1970-1976.
Billick, I H; Curran, A S; Shier, D R
1979-01-01
A study was completed of more than 170,000 records of pediatric venous blood levels and supporting demographic information collected in New York City during 1970-1976. The geometric mean (GM) blood lead level shows a consistent cyclical variation superimposed on an overall decreasing trend with time for all ages and ethnic groups studied. The GM blood lead levels for blacks are significantly greater than those for either Hispanics or whites. Regression analysis indicates a significant statistical association between GM blood lead level and ambient air lead level, after appropriate adjustments are made for age and ethnic group. These highly significant statistical relationships provide extremely strong incentives and directions for research into casual factors related to blood lead levels in children. PMID:499123
Moral sensitivity in Primary Health Care nurses.
Nora, Carlise Rigon Dalla; Zoboli, Elma Lourdes Campos Pavone; Vieira, Margarida M
2017-04-01
to characterize the profile and describe the moral sensitivity of primary health care nurses. this is a quantitative, transversal, exploratory, descriptive study. The data were collected through the Moral Sensitivity Questionnaire translated and adapted to Brazil. 100 primary health care nurses participated, from Rio Grande do Sul, Brazil. The data collection took place during the months of March and July 2016, in an online form. The analysis of the data occurred through descriptive statistical analysis. the nurses had an average moral sensitivity of 4.5 (out of 7). The dimensions with the greatest moral sensitivity were: interpersonal orientation, professional knowledge, moral conflict and moral meaning. the nurses of Rio Grande do Sul have a moderate moral sensitivity, which may contribute to a lower quality in Primary Health Care.
Automatic Generation of Algorithms for the Statistical Analysis of Planetary Nebulae Images
NASA Technical Reports Server (NTRS)
Fischer, Bernd
2004-01-01
Analyzing data sets collected in experiments or by observations is a Core scientific activity. Typically, experimentd and observational data are &aught with uncertainty, and the analysis is based on a statistical model of the conjectured underlying processes, The large data volumes collected by modern instruments make computer support indispensible for this. Consequently, scientists spend significant amounts of their time with the development and refinement of the data analysis programs. AutoBayes [GF+02, FS03] is a fully automatic synthesis system for generating statistical data analysis programs. Externally, it looks like a compiler: it takes an abstract problem specification and translates it into executable code. Its input is a concise description of a data analysis problem in the form of a statistical model as shown in Figure 1; its output is optimized and fully documented C/C++ code which can be linked dynamically into the Matlab and Octave environments. Internally, however, it is quite different: AutoBayes derives a customized algorithm implementing the given model using a schema-based process, and then further refines and optimizes the algorithm into code. A schema is a parameterized code template with associated semantic constraints which define and restrict the template s applicability. The schema parameters are instantiated in a problem-specific way during synthesis as AutoBayes checks the constraints against the original model or, recursively, against emerging sub-problems. AutoBayes schema library contains problem decomposition operators (which are justified by theorems in a formal logic in the domain of Bayesian networks) as well as machine learning algorithms (e.g., EM, k-Means) and nu- meric optimization methods (e.g., Nelder-Mead simplex, conjugate gradient). AutoBayes augments this schema-based approach by symbolic computation to derive closed-form solutions whenever possible. This is a major advantage over other statistical data analysis systems which use numerical approximations even in cases where closed-form solutions exist. AutoBayes is implemented in Prolog and comprises approximately 75.000 lines of code. In this paper, we take one typical scientific data analysis problem-analyzing planetary nebulae images taken by the Hubble Space Telescope-and show how AutoBayes can be used to automate the implementation of the necessary anal- ysis programs. We initially follow the analysis described by Knuth and Hajian [KHO2] and use AutoBayes to derive code for the published models. We show the details of the code derivation process, including the symbolic computations and automatic integration of library procedures, and compare the results of the automatically generated and manually implemented code. We then go beyond the original analysis and use AutoBayes to derive code for a simple image segmentation procedure based on a mixture model which can be used to automate a manual preproceesing step. Finally, we combine the original approach with the simple segmentation which yields a more detailed analysis. This also demonstrates that AutoBayes makes it easy to combine different aspects of data analysis.
Federal Register 2010, 2011, 2012, 2013, 2014
2012-03-19
... Transportation Statistics [Docket: RITA 2008-0002 BTS Paperwork Reduction Notice] Agency Information Collection... of Transportation Statistics (BTS), DOT. ACTION: Notice. SUMMARY: In compliance with the Paperwork Reduction Act of 1995, Public Law 104-13, the Bureau of Transportation Statistics invites the general public...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-07-27
... for OMB Review; Comment Request; Report on Current Employment Statistics ACTION: Notice. SUMMARY: The Department of Labor (DOL) is submitting the revised Bureau of Labor Statistics (BLS) sponsored information collection request (ICR) titled, ``Report on Current Employment Statistics,'' to the Office of Management and...
Federal Register 2010, 2011, 2012, 2013, 2014
2012-11-16
... Statistics Service Notice of Intent To Request Revision and Extension of a Currently Approved Information Collection AGENCY: National Agricultural Statistics Service, USDA. ACTION: Notice and request for comments... National Agricultural Statistics Service (NASS) to request revision and extension of a currently approved...
78 FR 26611 - Notice of Intent To Seek Approval To Conduct an Information Collection
Federal Register 2010, 2011, 2012, 2013, 2014
2013-05-07
... Statistics Service Notice of Intent To Seek Approval To Conduct an Information Collection AGENCY: National Agricultural Statistics Service, USDA. ACTION: Notice and request for comments. SUMMARY: In accordance with the Paperwork Reduction Act of 1995, this notice announces the intention of the National Agricultural Statistics...
Burrow, J Gordon
2016-05-01
This small-scale study examined the role that bare footprint collection and measurement processes have on the Reel method of measurement in forensic podiatry and its use in the Criminal Justice System. Previous research indicated that the Reel method was a valid and reliable measurement system for bare footprint analysis but various collection systems have been used to collect footprint data and both manual and digital measurement processes were utilized in forensic podiatry and other disciplines. This study contributes to the debate about collecting bare footprints; the techniques employed to quantify various Reel measurements and considered whether there was asymmetry between feet and footprints of the same person. An inductive, quantitative paradigm used the Podotrack gathering procedure for footprint collection and the subsequent dynamic footprints subjected to Adobe Photoshop techniques of calculating the Reel linear variables. Statistical analyses using paired-sample t tests were conducted to test hypotheses and compare data sets. Standard error of mean (SEM) showed variation between feet and the findings provide support for the Reel study and measurement method. Copyright © 2016 The Chartered Society of Forensic Sciences. Published by Elsevier Ireland Ltd. All rights reserved.
Quantifying biodiversity using digital cameras and automated image analysis.
NASA Astrophysics Data System (ADS)
Roadknight, C. M.; Rose, R. J.; Barber, M. L.; Price, M. C.; Marshall, I. W.
2009-04-01
Monitoring the effects on biodiversity of extensive grazing in complex semi-natural habitats is labour intensive. There are also concerns about the standardization of semi-quantitative data collection. We have chosen to focus initially on automating the most time consuming aspect - the image analysis. The advent of cheaper and more sophisticated digital camera technology has lead to a sudden increase in the number of habitat monitoring images and information that is being collected. We report on the use of automated trail cameras (designed for the game hunting market) to continuously capture images of grazer activity in a variety of habitats at Moor House National Nature Reserve, which is situated in the North of England at an average altitude of over 600m. Rainfall is high, and in most areas the soil consists of deep peat (1m to 3m), populated by a mix of heather, mosses and sedges. The cameras have been continuously in operation over a 6 month period, daylight images are in full colour and night images (IR flash) are black and white. We have developed artificial intelligence based methods to assist in the analysis of the large number of images collected, generating alert states for new or unusual image conditions. This paper describes the data collection techniques, outlines the quantitative and qualitative data collected and proposes online and offline systems that can reduce the manpower overheads and increase focus on important subsets in the collected data. By converting digital image data into statistical composite data it can be handled in a similar way to other biodiversity statistics thus improving the scalability of monitoring experiments. Unsupervised feature detection methods and supervised neural methods were tested and offered solutions to simplifying the process. Accurate (85 to 95%) categorization of faunal content can be obtained, requiring human intervention for only those images containing rare animals or unusual (undecidable) conditions, and enabling automatic deletion of images generated by erroneous triggering (e.g. cloud movements). This is the first step to a hierarchical image processing framework, where situation subclasses such as birds or climatic conditions can be fed into more appropriate automated or semi-automated data mining software.
Statistical analysis of cyprinid ethoxyresorufin-O-deethylase data in a large French watershed.
Flammarion, P; Migeon, B; Garric, J
1998-01-01
A comparison of ethoxyresorufin-O-deethylase (EROD) data collected in 1995 in various sites in the Rhône watershed (France) was carried out to quantify the influence of factors such as contamination and biological parameters on EROD levels and within-group variabilities. Three species of cyprinids were collected and fish chemical contamination was measured. A log transformation of EROD data provided both normalization and homogeneity of variances. The influence of female sexual maturation on the variability and EROD dimorphism was quantified. A relationship with contaminant bioaccumulation was observed. A comparison with EROD data collected during previous studies by the same laboratory was made to validate the results.
Moyle, Phillip R.; Causey, J. Douglas
2001-01-01
This report provides chemical analyses for 31 samples collected from various phosphate mine sites in southeastern Idaho (25), northern Utah (2), and western Wyoming (4). The sampling effort was undertaken as a reconnaissance and does not constitute a characterization of mine wastes. Twenty-five samples were collected from waste rock dumps, 2 from stockpiles, and 1 each from slag, tailings, mill shale, and an outcrop. All samples were analyzed for a suite of major, minor, and trace elements. Although the analytical data set for the 31 samples is too small for detailed statistical analysis, a summary of general observations is made.
Study/Experimental/Research Design: Much More Than Statistics
Knight, Kenneth L.
2010-01-01
Abstract Context: The purpose of study, experimental, or research design in scientific manuscripts has changed significantly over the years. It has evolved from an explanation of the design of the experiment (ie, data gathering or acquisition) to an explanation of the statistical analysis. This practice makes “Methods” sections hard to read and understand. Objective: To clarify the difference between study design and statistical analysis, to show the advantages of a properly written study design on article comprehension, and to encourage authors to correctly describe study designs. Description: The role of study design is explored from the introduction of the concept by Fisher through modern-day scientists and the AMA Manual of Style. At one time, when experiments were simpler, the study design and statistical design were identical or very similar. With the complex research that is common today, which often includes manipulating variables to create new variables and the multiple (and different) analyses of a single data set, data collection is very different than statistical design. Thus, both a study design and a statistical design are necessary. Advantages: Scientific manuscripts will be much easier to read and comprehend. A proper experimental design serves as a road map to the study methods, helping readers to understand more clearly how the data were obtained and, therefore, assisting them in properly analyzing the results. PMID:20064054
Evaluating collective significance of climatic trends: A comparison of methods on synthetic data
NASA Astrophysics Data System (ADS)
Huth, Radan; Dubrovský, Martin
2017-04-01
The common approach to determine whether climatic trends are significantly different from zero is to conduct individual (local) tests at each single site (station or gridpoint). Whether the number of sites where the trends are significantly non-zero can or cannot occur by random, is almost never evaluated in trend studies. That is, collective (global) significance of trends is ignored. We compare three approaches to evaluating collective statistical significance of trends at a network of sites, using the following statistics: (i) the number of successful local tests (a successful test means here a test in which the null hypothesis of no trend is rejected); this is a standard way of assessing collective significance in various applications in atmospheric sciences; (ii) the smallest p-value among the local tests (Walker test); and (iii) the counts of positive and negative trends regardless of their magnitudes and local significance. The third approach is a new procedure that we propose; the rationale behind it is that it is reasonable to assume that the prevalence of one sign of trends at individual sites is indicative of a high confidence in the trend not being zero, regardless of the (in)significance of individual local trends. A potentially large amount of information contained in trends that are not locally significant, which are typically deemed irrelevant and neglected, is thus not lost and is retained in the analysis. In this contribution we examine the feasibility of the proposed way of significance testing on synthetic data, produced by a multi-site stochastic generator, and compare it with the two other ways of assessing collective significance, which are well established now. The synthetic dataset, mimicking annual mean temperature on an array of stations (or gridpoints), is constructed assuming a given statistical structure characterized by (i) spatial separation (density of the station network), (ii) local variance, (iii) temporal and spatial autocorrelations, and (iv) the trend magnitude. The probabilistic distributions of the three test statistics (null distributions) and critical values of the tests are determined from multiple realizations of the synthetic dataset, in which no trend is imposed at each site (that is, any trend is a result of random fluctuations only). The procedure is then evaluated by determining the type II error (the probability of a false detection of a trend) in the presence of a trend with a known magnitude, for which the synthetic dataset with an imposed spatially uniform non-zero trend is used. A sensitivity analysis is conducted for various combinations of the trend magnitude and spatial autocorrelation.
Mathematics authentic assessment on statistics learning: the case for student mini projects
NASA Astrophysics Data System (ADS)
Fauziah, D.; Mardiyana; Saputro, D. R. S.
2018-03-01
Mathematics authentic assessment is a form of meaningful measurement of student learning outcomes for the sphere of attitude, skill and knowledge in mathematics. The construction of attitude, skill and knowledge achieved through the fulfilment of tasks which involve active and creative role of the students. One type of authentic assessment is student mini projects, started from planning, data collecting, organizing, processing, analysing and presenting the data. The purpose of this research is to learn the process of using authentic assessments on statistics learning which is conducted by teachers and to discuss specifically the use of mini projects to improving students’ learning in the school of Surakarta. This research is an action research, where the data collected through the results of the assessments rubric of student mini projects. The result of data analysis shows that the average score of rubric of student mini projects result is 82 with 96% classical completeness. This study shows that the application of authentic assessment can improve students’ mathematics learning outcomes. Findings showed that teachers and students participate actively during teaching and learning process, both inside and outside of the school. Student mini projects also provide opportunities to interact with other people in the real context while collecting information and giving presentation to the community. Additionally, students are able to exceed more on the process of statistics learning using authentic assessment.
Kinoshita, Manabu; Sakai, Mio; Arita, Hideyuki; Shofuda, Tomoko; Chiba, Yasuyoshi; Kagawa, Naoki; Watanabe, Yoshiyuki; Hashimoto, Naoya; Fujimoto, Yasunori; Yoshimine, Toshiki; Nakanishi, Katsuyuki; Kanemura, Yonehiro
2016-01-01
Reports have suggested that tumor textures presented on T2-weighted images correlate with the genetic status of glioma. Therefore, development of an image analyzing framework that is capable of objective and high throughput image texture analysis for large scale image data collection is needed. The current study aimed to address the development of such a framework by introducing two novel parameters for image textures on T2-weighted images, i.e., Shannon entropy and Prewitt filtering. Twenty-two WHO grade 2 and 28 grade 3 glioma patients were collected whose pre-surgical MRI and IDH1 mutation status were available. Heterogeneous lesions showed statistically higher Shannon entropy than homogenous lesions (p = 0.006) and ROC curve analysis proved that Shannon entropy on T2WI was a reliable indicator for discrimination of homogenous and heterogeneous lesions (p = 0.015, AUC = 0.73). Lesions with well-defined borders exhibited statistically higher Edge mean and Edge median values using Prewitt filtering than those with vague lesion borders (p = 0.0003 and p = 0.0005 respectively). ROC curve analysis also proved that both Edge mean and median values were promising indicators for discrimination of lesions with vague and well defined borders and both Edge mean and median values performed in a comparable manner (p = 0.0002, AUC = 0.81 and p < 0.0001, AUC = 0.83, respectively). Finally, IDH1 wild type gliomas showed statistically lower Shannon entropy on T2WI than IDH1 mutated gliomas (p = 0.007) but no difference was observed between IDH1 wild type and mutated gliomas in Edge median values using Prewitt filtering. The current study introduced two image metrics that reflect lesion texture described on T2WI. These two metrics were validated by readings of a neuro-radiologist who was blinded to the results. This observation will facilitate further use of this technique in future large scale image analysis of glioma.
Load Model Verification, Validation and Calibration Framework by Statistical Analysis on Field Data
NASA Astrophysics Data System (ADS)
Jiao, Xiangqing; Liao, Yuan; Nguyen, Thai
2017-11-01
Accurate load models are critical for power system analysis and operation. A large amount of research work has been done on load modeling. Most of the existing research focuses on developing load models, while little has been done on developing formal load model verification and validation (V&V) methodologies or procedures. Most of the existing load model validation is based on qualitative rather than quantitative analysis. In addition, not all aspects of model V&V problem have been addressed by the existing approaches. To complement the existing methods, this paper proposes a novel load model verification and validation framework that can systematically and more comprehensively examine load model's effectiveness and accuracy. Statistical analysis, instead of visual check, quantifies the load model's accuracy, and provides a confidence level of the developed load model for model users. The analysis results can also be used to calibrate load models. The proposed framework can be used as a guidance to systematically examine load models for utility engineers and researchers. The proposed method is demonstrated through analysis of field measurements collected from a utility system.
75 FR 1415 - Submission for OMB Review: Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2010-01-11
... Department of Labor--Bureau of Labor Statistics (BLS), Office of Management and Budget, Room 10235... Statistics. Type of Review: Revision of a currently approved collection. Title of Collection: The Consumer... sector. The data are collected from a national probability sample of households designed to represent the...
Feature-Based Statistical Analysis of Combustion Simulation Data
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bennett, J; Krishnamoorthy, V; Liu, S
2011-11-18
We present a new framework for feature-based statistical analysis of large-scale scientific data and demonstrate its effectiveness by analyzing features from Direct Numerical Simulations (DNS) of turbulent combustion. Turbulent flows are ubiquitous and account for transport and mixing processes in combustion, astrophysics, fusion, and climate modeling among other disciplines. They are also characterized by coherent structure or organized motion, i.e. nonlocal entities whose geometrical features can directly impact molecular mixing and reactive processes. While traditional multi-point statistics provide correlative information, they lack nonlocal structural information, and hence, fail to provide mechanistic causality information between organized fluid motion and mixing andmore » reactive processes. Hence, it is of great interest to capture and track flow features and their statistics together with their correlation with relevant scalar quantities, e.g. temperature or species concentrations. In our approach we encode the set of all possible flow features by pre-computing merge trees augmented with attributes, such as statistical moments of various scalar fields, e.g. temperature, as well as length-scales computed via spectral analysis. The computation is performed in an efficient streaming manner in a pre-processing step and results in a collection of meta-data that is orders of magnitude smaller than the original simulation data. This meta-data is sufficient to support a fully flexible and interactive analysis of the features, allowing for arbitrary thresholds, providing per-feature statistics, and creating various global diagnostics such as Cumulative Density Functions (CDFs), histograms, or time-series. We combine the analysis with a rendering of the features in a linked-view browser that enables scientists to interactively explore, visualize, and analyze the equivalent of one terabyte of simulation data. We highlight the utility of this new framework for combustion science; however, it is applicable to many other science domains.« less
DNA analysis in Disaster Victim Identification.
Montelius, Kerstin; Lindblom, Bertil
2012-06-01
DNA profiling and matching is one of the primary methods to identify missing persons in a disaster, as defined by the Interpol Disaster Victim Identification Guide. The process to identify a victim by DNA includes: the collection of the best possible ante-mortem (AM) samples, the choice of post-mortem (PM) samples, DNA-analysis, matching and statistical weighting of the genetic relationship or match. Each disaster has its own scenario, and each scenario defines its own methods for identification of the deceased.
Charm dimuon production in neutrino-nucleon interactions in the NOMAD experiment
NASA Astrophysics Data System (ADS)
Petti, Roberto; Samoylov, Oleg
2012-09-01
We present our new measurement of charm dimuon production in neutrino-iron interactions based upon the full statistics collected by the NOMAD experiment. After background subtraction we observe 15,340 charm dimuon events, providing the largest sample currently available. The analysis exploits the large inclusive charged current sample (about 9 million events after all analysis cuts) to constrain the total systematic uncertainty to about 2%. The extraction of strange sea and charm production parameters is also discussed.
Charm dimuon production in neutrino-nucleon interactions in the NOMAD experiment
NASA Astrophysics Data System (ADS)
Petti, R.; Samoylov, O. B.
2011-12-01
We present our new measurement of charm dimuon production in neutrino-iron interactions based upon the full statistics collected by the NOMAD experiment. After background subtraction we observe 15,340 charm dimuon events, providing the largest sample currently available. The analysis exploits the large inclusive charged current sample (about 9 million events after all analysis cuts) to constrain the total systematic uncertainty to ˜2%. The extraction of strange sea and charm production parameters is also discussed.
Karami, Manoochehr; Khazaei, Salman
2017-12-06
Clinical decision makings according studies result require the valid and correct data collection, andanalysis. However, there are some common methodological and statistical issues which may ignore by authors. In individual matched case- control design bias arising from the unconditional analysis instead of conditional analysis. Using an unconditional logistic for matched data causes the imposition of a large number of nuisance parameters which may result in seriously biased estimates.
1992-02-01
configuration. We have spent the last year observing two firms as they experimented with modular manufacturing. The following report will track the progress of...the transitions as they I moved through the year . Incorporated into the analysis is the statistical interpretation of data collected from each firm, as...during the year . FEBRUARY The most noticeable change this month was the introduction of the new ergonomic chairs for the operators. Previously the
The Problem of Auto-Correlation in Parasitology
Pollitt, Laura C.; Reece, Sarah E.; Mideo, Nicole; Nussey, Daniel H.; Colegrave, Nick
2012-01-01
Explaining the contribution of host and pathogen factors in driving infection dynamics is a major ambition in parasitology. There is increasing recognition that analyses based on single summary measures of an infection (e.g., peak parasitaemia) do not adequately capture infection dynamics and so, the appropriate use of statistical techniques to analyse dynamics is necessary to understand infections and, ultimately, control parasites. However, the complexities of within-host environments mean that tracking and analysing pathogen dynamics within infections and among hosts poses considerable statistical challenges. Simple statistical models make assumptions that will rarely be satisfied in data collected on host and parasite parameters. In particular, model residuals (unexplained variance in the data) should not be correlated in time or space. Here we demonstrate how failure to account for such correlations can result in incorrect biological inference from statistical analysis. We then show how mixed effects models can be used as a powerful tool to analyse such repeated measures data in the hope that this will encourage better statistical practices in parasitology. PMID:22511865
Multivariate frequency domain analysis of protein dynamics
NASA Astrophysics Data System (ADS)
Matsunaga, Yasuhiro; Fuchigami, Sotaro; Kidera, Akinori
2009-03-01
Multivariate frequency domain analysis (MFDA) is proposed to characterize collective vibrational dynamics of protein obtained by a molecular dynamics (MD) simulation. MFDA performs principal component analysis (PCA) for a bandpass filtered multivariate time series using the multitaper method of spectral estimation. By applying MFDA to MD trajectories of bovine pancreatic trypsin inhibitor, we determined the collective vibrational modes in the frequency domain, which were identified by their vibrational frequencies and eigenvectors. At near zero temperature, the vibrational modes determined by MFDA agreed well with those calculated by normal mode analysis. At 300 K, the vibrational modes exhibited characteristic features that were considerably different from the principal modes of the static distribution given by the standard PCA. The influences of aqueous environments were discussed based on two different sets of vibrational modes, one derived from a MD simulation in water and the other from a simulation in vacuum. Using the varimax rotation, an algorithm of the multivariate statistical analysis, the representative orthogonal set of eigenmodes was determined at each vibrational frequency.
Indelicato, Serena; Bongiorno, David; Tuzzolino, Nicola; Mannino, Maria Rosaria; Muscarella, Rosalia; Fradella, Pasquale; Gargano, Maria Elena; Nicosia, Salvatore; Ceraulo, Leopoldo
2018-03-14
Multivariate analysis was performed on a large data set of groundwater and leachate samples collected during 9 years of operation of the Bellolampo municipal solid waste landfill (located above Palermo, Italy). The aim was to obtain the most likely correlations among the data. The analysis results are presented. Groundwater samples were collected in the period 2004-2013, whereas the leachate analysis refers to the period 2006-2013. For groundwater, statistical data evaluation revealed notable differences among the samples taken from the numerous wells located around the landfill. Characteristic parameters revealed by principal component analysis (PCA) were more deeply investigated, and corresponding thematic maps were drawn. The composition of the leachate was also thoroughly investigated. Several chemical macro-descriptors were calculated, and the results are presented. A comparison of PCA results for the leachate and groundwater data clearly reveals that the groundwater's main components substantially differ from those of the leachate. This outcome strongly suggests excluding leachate permeation through the multiple landfill lining.
Federal Register 2010, 2011, 2012, 2013, 2014
2013-01-07
... DEPARTMENT OF JUSTICE [OMB Number 1121-0094] Agency Information Collection Activities: Existing...: 60-day notice. The Department of Justice (DOJ), Bureau of Justice Statistics, will be submitting the... information, please contact Todd D. Minton, Bureau of Justice Statistics, 810 Seventh Street NW., Washington...
Using Statistics for Database Management in an Academic Library.
ERIC Educational Resources Information Center
Hyland, Peter; Wright, Lynne
1996-01-01
Collecting statistical data about database usage by library patrons aids in the management of CD-ROM and database offerings, collection development, and evaluation of training programs. Two approaches to data collection are presented which should be used together: an automated or nonintrusive method which monitors search sessions while the…
Collecting and Using Networked Statistics: Current Status, Future Goals
ERIC Educational Resources Information Center
Hiott, Judith
2004-01-01
For more than five years the Houston Public Library has collected statistics for measuring networked collections and services based on emerging guidelines. While the guidelines have provided authority and stability to the process, the clarification process continues. The development of information discovery software, such as federated search tools…
76 FR 12823 - Enhanced Collection of Relevant Data and Statistics Relating to Women
Federal Register 2010, 2011, 2012, 2013, 2014
2011-03-09
... greater understanding of policies and programs. Preparation of this report revealed the vast data... Collection of Relevant Data and Statistics Relating to Women Memorandum for the Heads of Executive... accompanying website collection of relevant data, will assist Government officials in crafting policies in...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-03-02
... DEPARTMENT OF HOMELAND SECURITY Coast Guard [USCG-2011-0017] Information Collection Request to... Secretary to collect, analyze and publish reports, information, and statistics on marine casualties. Need... the Secretary (delegated to the Coast Guard) reports, information, and statistics on casualties...
Federal Register 2010, 2011, 2012, 2013, 2014
2013-03-14
... collecting, compiling, and publishing export trade statistics for the United States under the provisions of... for collecting export trade data, which is used by the Census Bureau for statistical purposes only... timely statistics as well as address the enforcement and security concerns of CBP, the Bureau of Industry...
Geo-statistical analysis of Culicoides spp. distribution and abundance in Sicily, Italy.
Blanda, Valeria; Blanda, Marcellocalogero; La Russa, Francesco; Scimeca, Rossella; Scimeca, Salvatore; D'Agostino, Rosalia; Auteri, Michelangelo; Torina, Alessandra
2018-02-01
Biting midges belonging to Culicoides imicola, Culicoides obsoletus complex and Culicoides pulicaris complex (Diptera: Ceratopogonidae) are increasingly implicated as vectors of bluetongue virus in Palaearctic regions. Culicoides obsoletus complex includes C. obsoletus (sensu stricto), C. scoticus, C. dewulfi and C. chiopterus. Culicoides pulicaris and C. lupicaris belong to the Culicoides pulicaris complex. The aim of this study was a geo-statistical analysis of the abundance and spatial distribution of Culicoides spp. involved in bluetongue virus transmission. As part of the national bluetongue surveillance plan 7081 catches were collected in 897 Sicilian farms from 2000 to 2013. Onderstepoort-type blacklight traps were used for sample collection and each catch was analysed for the presence of Culicoides spp. and for the presence and abundance of Culicoides vector species (C. imicola, C. pulicaris / C. obsoletus complexes). A geo-statistical analysis was carried out monthly via the interpolation of measured values based on the Inverse Distance Weighted method, using a GIS tool. Raster maps were reclassified into seven classes according to the presence and abundance of Culicoides, in order to obtain suitable maps for Map Algebra operations. Sicilian provinces showing a very high abundance of Culicoides vector species were Messina (80% of the whole area), Palermo (20%) and Catania (12%). A total of 5654 farms fell within the very high risk area for bluetongue (21% of the 26,676 farms active in Sicily); of these, 3483 farms were in Messina, 1567 in Palermo and 604 in Catania. Culicoides imicola was prevalent in Palermo, C. pulicaris in Messina and C. obsoletus complex was very abundant over the whole island with the highest abundance value in Messina. Our study reports the results of a geo-statistical analysis concerning the abundance and spatial distribution of Culicoides spp. in Sicily throughout the fourteen year study. It provides useful decision support in the field of epidemiology, allowing the identification of areas to be monitored as bases for improved surveillance plans. Moreover, this knowledge can become a tool for the evaluation of virus transmission risks, especially if related to vector competence.
Narayanan, Roshni; Nugent, Rebecca; Nugent, Kenneth
2015-10-01
Accreditation Council for Graduate Medical Education guidelines require internal medicine residents to develop skills in the interpretation of medical literature and to understand the principles of research. A necessary component is the ability to understand the statistical methods used and their results, material that is not an in-depth focus of most medical school curricula and residency programs. Given the breadth and depth of the current medical literature and an increasing emphasis on complex, sophisticated statistical analyses, the statistical foundation and education necessary for residents are uncertain. We reviewed the statistical methods and terms used in 49 articles discussed at the journal club in the Department of Internal Medicine residency program at Texas Tech University between January 1, 2013 and June 30, 2013. We collected information on the study type and on the statistical methods used for summarizing and comparing samples, determining the relations between independent variables and dependent variables, and estimating models. We then identified the typical statistics education level at which each term or method is learned. A total of 14 articles came from the Journal of the American Medical Association Internal Medicine, 11 from the New England Journal of Medicine, 6 from the Annals of Internal Medicine, 5 from the Journal of the American Medical Association, and 13 from other journals. Twenty reported randomized controlled trials. Summary statistics included mean values (39 articles), category counts (38), and medians (28). Group comparisons were based on t tests (14 articles), χ2 tests (21), and nonparametric ranking tests (10). The relations between dependent and independent variables were analyzed with simple regression (6 articles), multivariate regression (11), and logistic regression (8). Nine studies reported odds ratios with 95% confidence intervals, and seven analyzed test performance using sensitivity and specificity calculations. These papers used 128 statistical terms and context-defined concepts, including some from data analysis (56), epidemiology-biostatistics (31), modeling (24), data collection (12), and meta-analysis (5). Ten different software programs were used in these articles. Based on usual undergraduate and graduate statistics curricula, 64.3% of the concepts and methods used in these papers required at least a master's degree-level statistics education. The interpretation of the current medical literature can require an extensive background in statistical methods at an education level exceeding the material and resources provided to most medical students and residents. Given the complexity and time pressure of medical education, these deficiencies will be hard to correct, but this project can serve as a basis for developing a curriculum in study design and statistical methods needed by physicians-in-training.
ERIC Educational Resources Information Center
Sengupta, Atanu; Pal, Naibedya Prasun
2012-01-01
Primary education is essential for the economic development in any country. Most studies give more emphasis to the final output (such as literacy, enrolment etc.) rather than the delivery of the entire primary education system. In this paper, we study the school level data from an Indian district, collected under the official DISE statistics. We…
ERIC Educational Resources Information Center
Kapoor, Kanta
2010-01-01
Purpose: The purpose of this paper is to quantify the use of electronic journals in comparison with the print collections in the Guru Gobind Singh Indraprastha University Library. Design/methodology/approach: A detailed analysis was made of the use of lending services, the Xerox facility and usage of electronic journals such as Science Direct,…
National Disability Registers Report on Causes of Intellectual Disability in Taiwan: 2000-2007
ERIC Educational Resources Information Center
Lin, Jin-Ding; Yen, Chia-Feng; Wu, Jia-Ling; Kang, Shih-Wan
2009-01-01
The main purposes of the present analysis were to describe the causes of intellectual disability (ID) and examine its overtime change from 2000 to 2007 in Taiwan. Data of the present study mainly come from the public web-access information which collected by the Department of Statistics, Ministry of the Interiors, Taipei, Taiwan. Data were…
ERIC Educational Resources Information Center
Hollenbeck, Kevin
A study examined the effect of education and training on the economy and on employment outcomes. Data collected during a 1982 nationwide telephone survey of 3,500 employers were used as the basis for statistical models of voluntary and involuntary job separations and job promotions. Four major conclusions resulted from the modeling process…
ERIC Educational Resources Information Center
Whitehead, Michele L.; Gutierrez, Laura; Miller, Melody
2014-01-01
The purpose of this study is to gain an understanding of current academic medical library circulation polices and examine methods libraries utilize to meet patron needs. Key informants were selected from five states. Statistics regarding financial practices, users, services, space access, and circulation practices were collected via survey…
Status of the NASA Robotic Mission Conjunction Assessment Effort
NASA Technical Reports Server (NTRS)
Newman, Lauri Kraft
2007-01-01
This viewgraph presentation discusses NASA's processes and tools used to mitigate threats to NASA's robotic assets. The topics include: 1) Background; 2) Goddard Stakeholders and Mission Support; 3) ESC and TDRS Mission Descriptions; 4) TDRS Conjunction Assessment Process; 5) ESMO Conjunction Assessment Process; 6) Recent Operations Experiences; 7) Statistics Collected for ESC Regime; and 8) Current and Future Analysis Items.
Measuring: From Paces to Feet. Used Numbers: Real Data in the Classroom. Grades 3-4.
ERIC Educational Resources Information Center
Corwin, Rebecca B.; Russell, Susan Jo
A unit of study that introduces measuring as a way of collecting data is presented. Suitable for students in grades 3 and 4, it provides a foundation for further work in statistics and data analysis. The investigations may extend from one to four class sessions and are grouped into three parts: "Introduction to Measurement"; "Using Standard…
ERIC Educational Resources Information Center
Artinian, Vrej-Armen
An extensive investigation of elementary school classrooms was conducted through the collection and statistical analysis of student and teacher responses to questions concerning the educational environment. Several asepcts of the classroom are discussed, including the spatial, thermal, luminous, and aural environments. Questions were organized so…
Schools and Data: The Educator's Guide for Using Data to Improve Decision Making
ERIC Educational Resources Information Center
Creighton, Theodore B.
2006-01-01
Since the first edition of "Schools and Data", the No Child Left Behind Act has swept the country, and data-based decision making is no longer an option for educators. Today's educational climate makes it imperative for all schools to collect data and use statistical analysis to help create clear goals and recognize strategies for…
ERIC Educational Resources Information Center
Lavigne, Heather J.; Shakman, Karen; Zweig, Jacqueline; Greller, Sara L.
2016-01-01
This study describes how principals reported spending their time and what professional development they reported participating in, based on data collected through the Schools and Staffing Survey by the National Center for Education Statistics during the 2011/12 school year. The study analyzes schools by grade level, poverty level, and within…
ERIC Educational Resources Information Center
Poole, Sonja Martin
2011-01-01
Using data from the National Center for Educational Statistics, this article examines the relationship between strength of state accountability policy (i.e., external accountability) and internal accountability, defined as a school-level system in which collective behaviors and conditions exist that direct the attention and effort of the internal…
Issues with PISA's Use of Its Data in the Context of International Education Policy Convergence
ERIC Educational Resources Information Center
Murphy, Dan
2014-01-01
In its analysis of data it collected in 2006, the Programme for International Student Assessment (PISA) survey by the Organisation for Economic Co-operation and Development (OECD) used a variety of statistical methods to arrive at the 'key findings' that school autonomy, school competition and the public posting of student achievement data were…
We will present observations and statistical analysis on indoor air and soil gas data, collected over four years (including parts of five winter seasons) along with data on meteorological and hydrological variations at an unoccupied pre-1920 duplex. The monitoring program has now...
A Comparative Study of Pre-Service Education for Preschool Teachers in China and the United States
ERIC Educational Resources Information Center
Gong, Xin; Wang, Pengcheng
2017-01-01
This study provides a comparative analysis of the pre-service education system for preschool educators in China and the United States. Based on collected data and materials (literature, policy documents, and statistical data), we compare two areas of pre-service training: (1) the formal system; (2) the informal system. In the formal system, most…
Al-Qarni, Mohammed A; Shakeela, Nasim Vahid; Alamri, Mohammed Abdullah; Alshaikh, Yahya A
2016-10-01
Eco-friendly or green dentistry can be a reality by effectively designing dental clinics and using more eco-friendly materials in the clinical practice. To determine the awareness of eco-friendly dentistry among dental faculty and students in preparation for future implementation. Assessment of knowledge regarding eco-friendly dentistry was done using an 18 item self-administered questionnaire among 160 participants. After baseline data collection, the intervention was done by educating participants with a power point presentation. The post-intervention data was then collected for analysis. Statistical analysis was done using Wilcoxon's signed rank test and one-way ANOVA. The educational intervention increased the knowledge about eco-friendly dentistry confirming the importance of continuing education. There was a statistically significant gain in knowledge among the participants after the presentation. The gain was highest for department of Preventive Dental Sciences (PDS) followed by Substitute Dental Sciences (SDS), No specialty, Maxillofacial Dental Sciences (MDS), and Restorative Dental Sciences (RDS) respectively. (F=5.5091, p<0.05). Lack of knowledge of green dentistry amongst the dental fraternity is highly prevailing. The same can be substantiated with effective training in the respective fields if channelized through the curriculum in an educational set-up.
Modeling Longitudinal Data Containing Non-Normal Within Subject Errors
NASA Technical Reports Server (NTRS)
Feiveson, Alan; Glenn, Nancy L.
2013-01-01
The mission of the National Aeronautics and Space Administration’s (NASA) human research program is to advance safe human spaceflight. This involves conducting experiments, collecting data, and analyzing data. The data are longitudinal and result from a relatively few number of subjects; typically 10 – 20. A longitudinal study refers to an investigation where participant outcomes and possibly treatments are collected at multiple follow-up times. Standard statistical designs such as mean regression with random effects and mixed–effects regression are inadequate for such data because the population is typically not approximately normally distributed. Hence, more advanced data analysis methods are necessary. This research focuses on four such methods for longitudinal data analysis: the recently proposed linear quantile mixed models (lqmm) by Geraci and Bottai (2013), quantile regression, multilevel mixed–effects linear regression, and robust regression. This research also provides computational algorithms for longitudinal data that scientists can directly use for human spaceflight and other longitudinal data applications, then presents statistical evidence that verifies which method is best for specific situations. This advances the study of longitudinal data in a broad range of applications including applications in the sciences, technology, engineering and mathematics fields.
CORSSA: Community Online Resource for Statistical Seismicity Analysis
NASA Astrophysics Data System (ADS)
Zechar, J. D.; Hardebeck, J. L.; Michael, A. J.; Naylor, M.; Steacy, S.; Wiemer, S.; Zhuang, J.
2011-12-01
Statistical seismology is critical to the understanding of seismicity, the evaluation of proposed earthquake prediction and forecasting methods, and the assessment of seismic hazard. Unfortunately, despite its importance to seismology-especially to those aspects with great impact on public policy-statistical seismology is mostly ignored in the education of seismologists, and there is no central repository for the existing open-source software tools. To remedy these deficiencies, and with the broader goal to enhance the quality of statistical seismology research, we have begun building the Community Online Resource for Statistical Seismicity Analysis (CORSSA, www.corssa.org). We anticipate that the users of CORSSA will range from beginning graduate students to experienced researchers. More than 20 scientists from around the world met for a week in Zurich in May 2010 to kick-start the creation of CORSSA: the format and initial table of contents were defined; a governing structure was organized; and workshop participants began drafting articles. CORSSA materials are organized with respect to six themes, each will contain between four and eight articles. CORSSA now includes seven articles with an additional six in draft form along with forums for discussion, a glossary, and news about upcoming meetings, special issues, and recent papers. Each article is peer-reviewed and presents a balanced discussion, including illustrative examples and code snippets. Topics in the initial set of articles include: introductions to both CORSSA and statistical seismology, basic statistical tests and their role in seismology; understanding seismicity catalogs and their problems; basic techniques for modeling seismicity; and methods for testing earthquake predictability hypotheses. We have also begun curating a collection of statistical seismology software packages.
Applying Triple-Matrix Masking for Privacy Preserving Data Collection and Sharing in HIV Studies.
Pei, Qinglin; Chen, Shigang; Xiao, Yao; Wu, Samuel S
2016-01-01
Many HIV research projects are plagued by the high missing rate of selfreported information during data collection. Also, due to the sensitive nature of the HIV research data, privacy protection is always a concern for data sharing in HIV studies. This paper applies a data masking approach, called triple-matrix masking [1], to the context of HIV research for ensuring privacy protection during the process of data collection and data sharing. Using a set of generated HIV patient data, we show step by step how the data are randomly transformed (masked) before leaving the patients' individual data collection device (which ensures that nobody sees the actual data) and how the masked data are further transformed by a masking service provider and a data collector. We demonstrate that the masked data retain statistical utility of the original data, yielding the exactly same inference results in the planned logistic regression on the effect of age on the adherence to antiretroviral therapy and in the Cox proportional hazard model for the age effect on time to viral load suppression. Privacy-preserving data collection method may help resolve the privacy protection issue in HIV research. The individual sensitive data can be completely hidden while the same inference results can still be obtained from the masked data, with the use of common statistical analysis methods.
A Numerical Simulation and Statistical Modeling of High Intensity Radiated Fields Experiment Data
NASA Technical Reports Server (NTRS)
Smith, Laura J.
2004-01-01
Tests are conducted on a quad-redundant fault tolerant flight control computer to establish upset characteristics of an avionics system in an electromagnetic field. A numerical simulation and statistical model are described in this work to analyze the open loop experiment data collected in the reverberation chamber at NASA LaRC as a part of an effort to examine the effects of electromagnetic interference on fly-by-wire aircraft control systems. By comparing thousands of simulation and model outputs, the models that best describe the data are first identified and then a systematic statistical analysis is performed on the data. All of these efforts are combined which culminate in an extrapolation of values that are in turn used to support previous efforts used in evaluating the data.
Chua, Felicia H Z; Thien, Ady; Ng, Lee Ping; Seow, Wan Tew; Low, David C Y; Chang, Kenneth T E; Lian, Derrick W Q; Loh, Eva; Low, Sharon Y Y
2017-03-01
Posterior fossa syndrome (PFS) is a serious complication faced by neurosurgeons and their patients, especially in paediatric medulloblastoma patients. The uncertain aetiology of PFS, myriad of cited risk factors and therapeutic challenges make this phenomenon an elusive entity. The primary objective of this study was to identify associative factors related to the development of PFS in medulloblastoma patient post-tumour resection. This is a retrospective study based at a single institution. Patient data and all related information were collected from the hospital records, in accordance to a list of possible risk factors associated with PFS. These included pre-operative tumour volume, hydrocephalus, age, gender, extent of resection, metastasis, ventriculoperitoneal shunt insertion, post-operative meningitis and radiological changes in MRI. Additional variables included molecular and histological subtypes of each patient's medulloblastoma tumour. Statistical analysis was employed to determine evidence of each variable's significance in PFS permanence. A total of 19 patients with appropriately complete data was identified. Initial univariate analysis did not show any statistical significance. However, multivariate analysis for MRI-specific changes reported bilateral DWI restricted diffusion changes involving both right and left sides of the surgical cavity was of statistical significance for PFS permanence. The authors performed a clinical study that evaluated possible risk factors for permanent PFS in paediatric medulloblastoma patients. Analysis of collated results found that post-operative DWI restriction in bilateral regions within the surgical cavity demonstrated statistical significance as a predictor of PFS permanence-a novel finding in the current literature.
Signatures of criticality arise from random subsampling in simple population models.
Nonnenmacher, Marcel; Behrens, Christian; Berens, Philipp; Bethge, Matthias; Macke, Jakob H
2017-10-01
The rise of large-scale recordings of neuronal activity has fueled the hope to gain new insights into the collective activity of neural ensembles. How can one link the statistics of neural population activity to underlying principles and theories? One attempt to interpret such data builds upon analogies to the behaviour of collective systems in statistical physics. Divergence of the specific heat-a measure of population statistics derived from thermodynamics-has been used to suggest that neural populations are optimized to operate at a "critical point". However, these findings have been challenged by theoretical studies which have shown that common inputs can lead to diverging specific heat. Here, we connect "signatures of criticality", and in particular the divergence of specific heat, back to statistics of neural population activity commonly studied in neural coding: firing rates and pairwise correlations. We show that the specific heat diverges whenever the average correlation strength does not depend on population size. This is necessarily true when data with correlations is randomly subsampled during the analysis process, irrespective of the detailed structure or origin of correlations. We also show how the characteristic shape of specific heat capacity curves depends on firing rates and correlations, using both analytically tractable models and numerical simulations of a canonical feed-forward population model. To analyze these simulations, we develop efficient methods for characterizing large-scale neural population activity with maximum entropy models. We find that, consistent with experimental findings, increases in firing rates and correlation directly lead to more pronounced signatures. Thus, previous reports of thermodynamical criticality in neural populations based on the analysis of specific heat can be explained by average firing rates and correlations, and are not indicative of an optimized coding strategy. We conclude that a reliable interpretation of statistical tests for theories of neural coding is possible only in reference to relevant ground-truth models.
Leitao, Mario M; Byrum, Graham V; Abu-Rustum, Nadeem R; Brown, Carol L; Chi, Dennis S; Sonoda, Yukio; Levine, Douglas A; Gardner, Ginger J; Barakat, Richard R
2010-11-01
A prior analysis of patients undergoing laparotomy for ovarian malignancies at our institution revealed an increased rate of intra-abdominal collections using HA-CMC film during debulking surgery. The primary objective of the current study was to determine whether the use of HA-CMC is associated with the development of postoperative intra-abdominal collections in patients undergoing laparotomy for uterine or cervical malignancies. We retrospectively identified all laparotomies performed for these malignancies from 3/1/05 to 12/31/07. We identified cases involving the use of HA-CMC via billing records and operative reports. Intra-abdominal collections were defined as localized intraperitoneal fluid accumulations in the absence of re-accumulating ascites. We noted incidences of intra-abdominal collections, as well as other complications. Appropriate statistical tests were applied using SPSS 15.0. We identified 169 laparotomies in which HA-CMC was used and 347 in which HA-CMC was not used. The following were statistically similar in both cohorts: age, body mass index (BMI), primary site, surgery for recurrent disease, prior intraperitoneal surgery, and extent of current surgery. Intra-abdominal collections were seen in 6 (3.6%) of 169 HA-CMC cases compared to 10 (2.9%) of 347 non-HA-CMC cases (p=0.7). The rate of infected collections was similar in both groups (1.2% vs. 1.4%). In the subgroup that underwent tumor debulking, intra-abdominal collections were seen in 3 (11.5%) of 26 HA-CMC cases compared to 2 (5.4%) of 37 non-HA-CMC cases (p=0.6). HA-CMC use does not appear to be associated with postoperative intra-abdominal collections in patients undergoing laparotomy for uterine or cervical cancer. Copyright © 2010 Elsevier Inc. All rights reserved.
Statistical analysis of 4 types of neck whiplash injuries based on classical meridian theory.
Chen, Yemeng; Zhao, Yan; Xue, Xiaolin; Li, Hui; Wu, Xiuyan; Zhang, Qunce; Zheng, Xin; Wang, Tianfang
2015-01-01
As one component of the Chinese medicine meridian system, the meridian sinew (Jingjin, (see text), tendino-musculo) is specially described as being for acupuncture treatment of the musculoskeletal system because of its dynamic attributes and tender point correlations. In recent decades, the therapeutic importance of the sinew meridian has become revalued in clinical application. Based on this theory, the authors have established therapeutic strategies of acupuncture treatment in Whiplash-Associated Disorders (WAD) by categorizing four types of neck symptom presentations. The advantage of this new system is to make it much easier for the clinician to find effective acupuncture points. This study attempts to prove the significance of the proposed therapeutic strategies by analyzing data collected from a clinical survey of various WAD using non-supervised statistical methods, such as correlation analysis, factor analysis, and cluster analysis. The clinical survey data have successfully verified discrete characteristics of four neck syndromes, based upon the range of motion (ROM) and tender point location findings. A summary of the relationships among the symptoms of the four neck syndromes has shown the correlation coefficient as having a statistical significance (P < 0.01 or P < 0.05), especially with regard to ROM. Furthermore, factor and cluster analyses resulted in a total of 11 categories of general symptoms, which implies syndrome factors are more related to the Liver, as originally described in classical theory. The hypothesis of meridian sinew syndromes in WAD is clearly supported by the statistical analysis of the clinical trials. This new discovery should be beneficial in improving therapeutic outcomes.
Bayesian inference for joint modelling of longitudinal continuous, binary and ordinal events.
Li, Qiuju; Pan, Jianxin; Belcher, John
2016-12-01
In medical studies, repeated measurements of continuous, binary and ordinal outcomes are routinely collected from the same patient. Instead of modelling each outcome separately, in this study we propose to jointly model the trivariate longitudinal responses, so as to take account of the inherent association between the different outcomes and thus improve statistical inferences. This work is motivated by a large cohort study in the North West of England, involving trivariate responses from each patient: Body Mass Index, Depression (Yes/No) ascertained with cut-off score not less than 8 at the Hospital Anxiety and Depression Scale, and Pain Interference generated from the Medical Outcomes Study 36-item short-form health survey with values returned on an ordinal scale 1-5. There are some well-established methods for combined continuous and binary, or even continuous and ordinal responses, but little work was done on the joint analysis of continuous, binary and ordinal responses. We propose conditional joint random-effects models, which take into account the inherent association between the continuous, binary and ordinal outcomes. Bayesian analysis methods are used to make statistical inferences. Simulation studies show that, by jointly modelling the trivariate outcomes, standard deviations of the estimates of parameters in the models are smaller and much more stable, leading to more efficient parameter estimates and reliable statistical inferences. In the real data analysis, the proposed joint analysis yields a much smaller deviance information criterion value than the separate analysis, and shows other good statistical properties too. © The Author(s) 2014.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wild, M.; Rouhani, S.
1995-02-01
A typical site investigation entails extensive sampling and monitoring. In the past, sampling plans have been designed on purely ad hoc bases, leading to significant expenditures and, in some cases, collection of redundant information. In many instances, sampling costs exceed the true worth of the collected data. The US Environmental Protection Agency (EPA) therefore has advocated the use of geostatistics to provide a logical framework for sampling and analysis of environmental data. Geostatistical methodology uses statistical techniques for the spatial analysis of a variety of earth-related data. The use of geostatistics was developed by the mining industry to estimate oremore » concentrations. The same procedure is effective in quantifying environmental contaminants in soils for risk assessments. Unlike classical statistical techniques, geostatistics offers procedures to incorporate the underlying spatial structure of the investigated field. Sample points spaced close together tend to be more similar than samples spaced further apart. This can guide sampling strategies and determine complex contaminant distributions. Geostatistic techniques can be used to evaluate site conditions on the basis of regular, irregular, random and even spatially biased samples. In most environmental investigations, it is desirable to concentrate sampling in areas of known or suspected contamination. The rigorous mathematical procedures of geostatistics allow for accurate estimates at unsampled locations, potentially reducing sampling requirements. The use of geostatistics serves as a decision-aiding and planning tool and can significantly reduce short-term site assessment costs, long-term sampling and monitoring needs, as well as lead to more accurate and realistic remedial design criteria.« less
Spontaneous collective synchronization in the Kuramoto model with additional non-local interactions
NASA Astrophysics Data System (ADS)
Gupta, Shamik
2017-10-01
In the context of the celebrated Kuramoto model of globally-coupled phase oscillators of distributed natural frequencies, which serves as a paradigm to investigate spontaneous collective synchronization in many-body interacting systems, we report on a very rich phase diagram in presence of thermal noise and an additional non-local interaction on a one-dimensional periodic lattice. Remarkably, the phase diagram involves both equilibrium and non-equilibrium phase transitions. In two contrasting limits of the dynamics, we obtain exact analytical results for the phase transitions. These two limits correspond to (i) the absence of thermal noise, when the dynamics reduces to that of a non-linear dynamical system, and (ii) the oscillators having the same natural frequency, when the dynamics becomes that of a statistical system in contact with a heat bath and relaxing to a statistical equilibrium state. In the former case, our exact analysis is based on the use of the so-called Ott-Antonsen ansatz to derive a reduced set of nonlinear partial differential equations for the macroscopic evolution of the system. Our results for the case of statistical equilibrium are on the other hand obtained by extending the well-known transfer matrix approach for nearest-neighbor Ising model to consider non-local interactions. The work offers a case study of exact analysis in many-body interacting systems. The results obtained underline the crucial role of additional non-local interactions in either destroying or enhancing the possibility of observing synchrony in mean-field systems exhibiting spontaneous synchronization.
Statistical models for fever forecasting based on advanced body temperature monitoring.
Jordan, Jorge; Miro-Martinez, Pau; Vargas, Borja; Varela-Entrecanales, Manuel; Cuesta-Frau, David
2017-02-01
Body temperature monitoring provides health carers with key clinical information about the physiological status of patients. Temperature readings are taken periodically to detect febrile episodes and consequently implement the appropriate medical countermeasures. However, fever is often difficult to assess at early stages, or remains undetected until the next reading, probably a few hours later. The objective of this article is to develop a statistical model to forecast fever before a temperature threshold is exceeded to improve the therapeutic approach to the subjects involved. To this end, temperature series of 9 patients admitted to a general internal medicine ward were obtained with a continuous monitoring Holter device, collecting measurements of peripheral and core temperature once per minute. These series were used to develop different statistical models that could quantify the probability of having a fever spike in the following 60 minutes. A validation series was collected to assess the accuracy of the models. Finally, the results were compared with the analysis of some series by experienced clinicians. Two different models were developed: a logistic regression model and a linear discrimination analysis model. Both of them exhibited a fever peak forecasting accuracy greater than 84%. When compared with experts' assessment, both models identified 35 (97.2%) of 36 fever spikes. The models proposed are highly accurate in forecasting the appearance of fever spikes within a short period in patients with suspected or confirmed febrile-related illnesses. Copyright © 2016 Elsevier Inc. All rights reserved.
Chandrasekaran, A; Ravisankar, R; Harikrishnan, N; Satapathy, K K; Prasad, M V R; Kanagasabapathy, K V
2015-02-25
Anthropogenic activities increase the accumulation of heavy metals in the soil environment. Soil pollution significantly reduces environmental quality and affects the human health. In the present study soil samples were collected at different locations of Yelagiri Hills, Tamilnadu, India for heavy metal analysis. The samples were analyzed for twelve selected heavy metals (Mg, Al, K, Ca, Ti, Fe, V, Cr, Mn, Co, Ni and Zn) using energy dispersive X-ray fluorescence (EDXRF) spectroscopy. Heavy metals concentration in soil were investigated using enrichment factor (EF), geo-accumulation index (Igeo), contamination factor (CF) and pollution load index (PLI) to determine metal accumulation, distribution and its pollution status. Heavy metal toxicity risk was assessed using soil quality guidelines (SQGs) given by target and intervention values of Dutch soil standards. The concentration of Ni, Co, Zn, Cr, Mn, Fe, Ti, K, Al, Mg were mainly controlled by natural sources. Multivariate statistical methods such as correlation matrix, principal component analysis and cluster analysis were applied for the identification of heavy metal sources (anthropogenic/natural origin). Geo-statistical methods such as kirging identified hot spots of metal contamination in road areas influenced mainly by presence of natural rocks. Copyright © 2014 Elsevier B.V. All rights reserved.
Water flow in high-speed handpieces.
Cavalcanti, Bruno Neves; Serairdarian, Paulo Isaías; Rode, Sigmar Mello
2005-05-01
This study measured the water flow commonly used in high-speed handpieces to evaluate the water flow's influence on temperature generation. Different flow speeds were evaluated between turbines that had different numbers of cooling apertures. Two water samples were collected from each high-speed handpiece at private practices and at the School of Dentistry at São José dos Campos. The first sample was collected at the customary flow and the second was collected with the terminal opened for maximum flow. The two samples were collected into weighed glass receptacles after 15 seconds of turbine operation. The glass receptacles were reweighed and the difference between weights was recorded to calculate the water flow in mL/min and for further statistical analysis. The average water flow for 137 samples was 29.48 mL/min. The flow speeds obtained were 42.38 mL/min for turbines with one coolant aperture; 34.31 mL/min for turbines with two coolant apertures; and 30.44 mL/min for turbines with three coolant apertures. There were statistical differences between turbines with one and three coolant apertures (Tukey-Kramer multiple comparisons test with P < .05). Turbine handpieces with one cooling aperture distributed more water for the burs than high-speed handpieces with more than one aperture.
Advanced LIGO low-latency searches
NASA Astrophysics Data System (ADS)
Kanner, Jonah; LIGO Scientific Collaboration, Virgo Collaboration
2016-06-01
Advanced LIGO recently made the first detection of gravitational waves from merging binary black holes. The signal was first identified by a low-latency analysis, which identifies gravitational-wave transients within a few minutes of data collection. More generally, Advanced LIGO transients are sought with a suite of automated tools, which collectively identify events, evaluate statistical significance, estimate source position, and attempt to characterize source properties. This low-latency effort is enabling a broad multi-messenger approach to the science of compact object mergers and other transients. This talk will give an overview of the low-latency methodology and recent results.
A study protocol to evaluate the relationship between outdoor air pollution and pregnancy outcomes
2010-01-01
Background The present study protocol is designed to assess the relationship between outdoor air pollution and low birth weight and preterm births outcomes performing a semi-ecological analysis. Semi-ecological design studies are widely used to assess effects of air pollution in humans. In this type of analysis, health outcomes and covariates are measured in individuals and exposure assignments are usually based on air quality monitor stations. Therefore, estimating individual exposures are one of the major challenges when investigating these relationships with a semi-ecologic design. Methods/Design Semi-ecologic study consisting of a retrospective cohort study with ecologic assignment of exposure is applied. Health outcomes and covariates are collected at Primary Health Care Center. Data from pregnant registry, clinical record and specific questionnaire administered orally to the mothers of children born in period 2007-2010 in Portuguese Alentejo Litoral region, are collected by the research team. Outdoor air pollution data are collected with a lichen diversity biomonitoring program, and individual pregnancy exposures are assessed with spatial geostatistical simulation, which provides the basis for uncertainty analysis of individual exposures. Awareness of outdoor air pollution uncertainty will improve validity of individual exposures assignments for further statistical analysis with multivariate regression models. Discussion Exposure misclassification is an issue of concern in semi-ecological design. In this study, personal exposures are assigned to each pregnant using geocoded addresses data. A stochastic simulation method is applied to lichen diversity values index measured at biomonitoring survey locations, in order to assess spatial uncertainty of lichen diversity value index at each geocoded address. These methods assume a model for spatial autocorrelation of exposure and provide a distribution of exposures in each study location. We believe that variability of simulated exposure values at geocoded addresses will improve knowledge on variability of exposures, improving therefore validity of individual exposures to input in posterior statistical analysis. PMID:20950449
A study protocol to evaluate the relationship between outdoor air pollution and pregnancy outcomes.
Ribeiro, Manuel C; Pereira, Maria J; Soares, Amílcar; Branquinho, Cristina; Augusto, Sofia; Llop, Esteve; Fonseca, Susana; Nave, Joaquim G; Tavares, António B; Dias, Carlos M; Silva, Ana; Selemane, Ismael; de Toro, Joaquin; Santos, Mário J; Santos, Fernanda
2010-10-15
The present study protocol is designed to assess the relationship between outdoor air pollution and low birth weight and preterm births outcomes performing a semi-ecological analysis. Semi-ecological design studies are widely used to assess effects of air pollution in humans. In this type of analysis, health outcomes and covariates are measured in individuals and exposure assignments are usually based on air quality monitor stations. Therefore, estimating individual exposures are one of the major challenges when investigating these relationships with a semi-ecologic design. Semi-ecologic study consisting of a retrospective cohort study with ecologic assignment of exposure is applied. Health outcomes and covariates are collected at Primary Health Care Center. Data from pregnant registry, clinical record and specific questionnaire administered orally to the mothers of children born in period 2007-2010 in Portuguese Alentejo Litoral region, are collected by the research team. Outdoor air pollution data are collected with a lichen diversity biomonitoring program, and individual pregnancy exposures are assessed with spatial geostatistical simulation, which provides the basis for uncertainty analysis of individual exposures. Awareness of outdoor air pollution uncertainty will improve validity of individual exposures assignments for further statistical analysis with multivariate regression models. Exposure misclassification is an issue of concern in semi-ecological design. In this study, personal exposures are assigned to each pregnant using geocoded addresses data. A stochastic simulation method is applied to lichen diversity values index measured at biomonitoring survey locations, in order to assess spatial uncertainty of lichen diversity value index at each geocoded address. These methods assume a model for spatial autocorrelation of exposure and provide a distribution of exposures in each study location. We believe that variability of simulated exposure values at geocoded addresses will improve knowledge on variability of exposures, improving therefore validity of individual exposures to input in posterior statistical analysis.
Spatial variability and long-term analysis of groundwater quality of Faisalabad industrial zone
NASA Astrophysics Data System (ADS)
Nasir, Muhammad Salman; Nasir, Abdul; Rashid, Haroon; Shah, Syed Hamid Hussain
2017-10-01
Water is the basic necessity of life and is essential for healthy society. In this study, groundwater quality analysis was carried out for the industrial zone of Faisalabad city. Sixty samples of groundwater were collected from the study area. The quality maps of deliberately analyzed results were prepared in GIS. The collected samples were analyzed for chemical parameters and heavy metals, such as total hardness, alkalinity, cadmium, arsenic, nickel, lead, and fluoride, and then, the results were compared with the WHO guidelines. The values of these results were represented by a mapping of quality parameters using the ArcView GIS v9.3, and IDW was used for raster interpolation. The long-term analysis of these parameters has been carried out using the `R Statistical' software. It was concluded that water is partially not fit for drinking, and direct use of this groundwater may cause health issues.
Ging, Patricia B.
1999-01-01
Surface-water sampling protocols of the U.S. Geological Survey National Water-Quality Assessment (NAWQA) Program specify samples for most properties and constituents to be collected manually in equal-width increments across a stream channel and composited for analysis. Single-point sampling with an automated sampler (autosampler) during storms was proposed in the upper part of the South-Central Texas NAWQA study unit, raising the question of whether property and constituent concentrations from automatically collected samples differ significantly from those in samples collected manually. Statistical (Wilcoxon signed-rank test) analyses of 3 to 16 paired concentrations for each of 26 properties and constituents from water samples collected using both methods at eight sites in the upper part of the study unit indicated that there were no significant differences in concentrations for dissolved constituents, other than calcium and organic carbon.
Indoor Location Sensing with Invariant Wi-Fi Received Signal Strength Fingerprinting
Husen, Mohd Nizam; Lee, Sukhan
2016-01-01
A method of location fingerprinting based on the Wi-Fi received signal strength (RSS) in an indoor environment is presented. The method aims to overcome the RSS instability due to varying channel disturbances in time by introducing the concept of invariant RSS statistics. The invariant RSS statistics represent here the RSS distributions collected at individual calibration locations under minimal random spatiotemporal disturbances in time. The invariant RSS statistics thus collected serve as the reference pattern classes for fingerprinting. Fingerprinting is carried out at an unknown location by identifying the reference pattern class that maximally supports the spontaneous RSS sensed from individual Wi-Fi sources. A design guideline is also presented as a rule of thumb for estimating the number of Wi-Fi signal sources required to be available for any given number of calibration locations under a certain level of random spatiotemporal disturbances. Experimental results show that the proposed method not only provides 17% higher success rate than conventional ones but also removes the need for recalibration. Furthermore, the resolution is shown finer by 40% with the execution time more than an order of magnitude faster than the conventional methods. These results are also backed up by theoretical analysis. PMID:27845711
Indoor Location Sensing with Invariant Wi-Fi Received Signal Strength Fingerprinting.
Husen, Mohd Nizam; Lee, Sukhan
2016-11-11
A method of location fingerprinting based on the Wi-Fi received signal strength (RSS) in an indoor environment is presented. The method aims to overcome the RSS instability due to varying channel disturbances in time by introducing the concept of invariant RSS statistics. The invariant RSS statistics represent here the RSS distributions collected at individual calibration locations under minimal random spatiotemporal disturbances in time. The invariant RSS statistics thus collected serve as the reference pattern classes for fingerprinting. Fingerprinting is carried out at an unknown location by identifying the reference pattern class that maximally supports the spontaneous RSS sensed from individual Wi-Fi sources. A design guideline is also presented as a rule of thumb for estimating the number of Wi-Fi signal sources required to be available for any given number of calibration locations under a certain level of random spatiotemporal disturbances. Experimental results show that the proposed method not only provides 17% higher success rate than conventional ones but also removes the need for recalibration. Furthermore, the resolution is shown finer by 40% with the execution time more than an order of magnitude faster than the conventional methods. These results are also backed up by theoretical analysis.
Staging Liver Fibrosis with Statistical Observers
NASA Astrophysics Data System (ADS)
Brand, Jonathan Frieman
Chronic liver disease is a worldwide health problem, and hepatic fibrosis (HF) is one of the hallmarks of the disease. Pathology diagnosis of HF is based on textural change in the liver as a lobular collagen network that develops within portal triads. The scale of collagen lobules is characteristically on order of 1mm, which close to the resolution limit of in vivo Gd-enhanced MRI. In this work the methods to collect training and testing images for a Hotelling observer are covered. An observer based on local texture analysis is trained and tested using wet-tissue phantoms. The technique is used to optimize the MRI sequence based on task performance. The final method developed is a two stage model observer to classify fibrotic and healthy tissue in both phantoms and in vivo MRI images. The first stage observer tests for the presence of local texture. Test statistics from the first observer are used to train the second stage observer to globally sample the local observer results. A decision of the disease class is made for an entire MRI image slice using test statistics collected from the second observer. The techniques are tested on wet-tissue phantoms and in vivo clinical patient data.
ERIC Educational Resources Information Center
Duy, Joanna; Vaughan, Liwen
2003-01-01
Vendor-provided electronic resource usage statistics are not currently standardized across vendors. This study investigates the feasibility of using locally collected data to check the reliability of vendor-provided data. Vendor-provided data were compared with local data collected from North Carolina State University (NCSU) Libraries' Web…
A Cyber-Attack Detection Model Based on Multivariate Analyses
NASA Astrophysics Data System (ADS)
Sakai, Yuto; Rinsaka, Koichiro; Dohi, Tadashi
In the present paper, we propose a novel cyber-attack detection model based on two multivariate-analysis methods to the audit data observed on a host machine. The statistical techniques used here are the well-known Hayashi's quantification method IV and cluster analysis method. We quantify the observed qualitative audit event sequence via the quantification method IV, and collect similar audit event sequence in the same groups based on the cluster analysis. It is shown in simulation experiments that our model can improve the cyber-attack detection accuracy in some realistic cases where both normal and attack activities are intermingled.
An Introduction to MAMA (Meta-Analysis of MicroArray data) System.
Zhang, Zhe; Fenstermacher, David
2005-01-01
Analyzing microarray data across multiple experiments has been proven advantageous. To support this kind of analysis, we are developing a software system called MAMA (Meta-Analysis of MicroArray data). MAMA utilizes a client-server architecture with a relational database on the server-side for the storage of microarray datasets collected from various resources. The client-side is an application running on the end user's computer that allows the user to manipulate microarray data and analytical results locally. MAMA implementation will integrate several analytical methods, including meta-analysis within an open-source framework offering other developers the flexibility to plug in additional statistical algorithms.
Assessment of levels of bacterial contamination of large wild game meat in Europe.
Membré, Jeanne-Marie; Laroche, Michel; Magras, Catherine
2011-08-01
The variations in prevalence and levels of pathogens and fecal contamination indicators in large wild game meat were studied to assess their potential impact on consumers. This analysis was based on hazard analysis, data generation and statistical analysis. A total of 2919 meat samples from three species (red deer, roe deer, wild boar) were collected at French game meat traders' facilities using two sampling protocols. Information was gathered on the types of meat cuts (forequarter or haunch; first sampling protocol) or type of retail-ready meat (stewing meat or roasting meat; second protocol), and also on the meat storage conditions (frozen or chilled), country of origin (eight countries) and shooting season (autumn, winter, spring). The samples were analyzed in both protocols for detection and enumeration of Escherichia coli, coagulase+staphylococci and Clostridium perfringens. In addition, detection and enumeration of thermotolerant coliforms and Listeria monocytogenes were performed for samples collected in the first and second protocols, respectively. The levels of bacterial contamination of the raw meat were determined by performing statistical analysis involving probabilistic techniques and Bayesian inference. C. perfringens was found in the highest numbers for the three indicators of microbial quality, hygiene and good handling, and L. monocytogenes in the lowest. Differences in contamination levels between game species and between meats distributed as chilled or frozen products were not significant. These results might be included in quantitative exposure assessments. Copyright © 2011 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Reddy, A.; Attaelmanan, A. G.; Mouton, M.
2012-07-01
The identification of more than 25% of the pottery sherds from the late PIR.D period (ca. 2nd - mid. 3rd c. AD) assemblage from the recently excavated building H at Mleiha as Indian is based on form and fabric, but using only visual assessment. Petrographic analysis of the fabrics can provide more precise indicators of the geographical origin of the wares. In this study, a total of 21 sherds from various key sites in Western India were compared with 7 different 'Indian' coarse-ware vessels sampled at Mleiha using X-ray fluorescence (XRF) spectrometry. The analyses were conducted on powdered samples collected from the core of each sherd. Each sample was irradiated for 1000 seconds using a 1.2 mm diameter X-ray beam. The resulting spectra were used for quantification of the X-ray intensity and elemental concentration. Levels of correlation in the elemental ratios of the sherds were statistically tested using an F-test as well as a Chi-test. Initial review of the XRF results indicates that the Maharashtra and Gujarat regions of India are probable source areas for at least two of the types of wares. Collection of additional samples from these areas and other regions of India, and further statistical analysis through methods such as Principal Component Analysis will help to isolate groups of wares from India and correlate them with types of vessels imported into the Oman peninsula in antiquity.
The extent and consequences of p-hacking in science.
Head, Megan L; Holman, Luke; Lanfear, Rob; Kahn, Andrew T; Jennions, Michael D
2015-03-01
A focus on novel, confirmatory, and statistically significant results leads to substantial bias in the scientific literature. One type of bias, known as "p-hacking," occurs when researchers collect or select data or statistical analyses until nonsignificant results become significant. Here, we use text-mining to demonstrate that p-hacking is widespread throughout science. We then illustrate how one can test for p-hacking when performing a meta-analysis and show that, while p-hacking is probably common, its effect seems to be weak relative to the real effect sizes being measured. This result suggests that p-hacking probably does not drastically alter scientific consensuses drawn from meta-analyses.
Demystification of Bell inequality
NASA Astrophysics Data System (ADS)
Khrennikov, Andrei
2009-08-01
The main aim of this review is to show that the common conclusion that Bell's argument implies that any attempt to proceed beyond quantum mechanics induces a nonlocal model was not totally justified. Our analysis of Bell's argument demonstrates that violation of Bell's inequality implies neither "death of realism" nor nonlocality. This violation is just a sign of non-Kolmogorovness of statistical data - impossibility to put statistical data collected in a few different experiments (corresponding to incompatible settings of polarization beam splitters) in one probability space. This inequality was well known in theoretical probability since 19th century (from works of Boole). We couple non-Kolmogorovness of data with design of modern detectors of photons.
Cho, Gun-Sang; Kim, Dae-Sung; Yi, Eun-Surk
2015-12-01
The purpose of this study is to verification of relationship model between Korean new elderly class's recovery resilience and productive aging. As of 2013, this study sampled preliminary elderly people in Gyeonggi-do and other provinces nationwide. Data from a total of effective 484 subjects was analyzed. The collected data was processed using the IBM SPSS 20.0 and AMOS 20.0, and underwent descriptive statistical analysis, confirmatory factor analysis, and structure model verification. The path coefficient associated with model fitness was examined. The standardization path coefficient between recovery resilience and productive aging is β=0.975 (t=14.790), revealing a statistically significant positive effect. Thus, it was found that the proposed basic model on the direct path of recovery resilience and productive aging was fit for the model.
Cho, Gun-Sang; Kim, Dae-Sung; Yi, Eun-Surk
2015-01-01
The purpose of this study is to verification of relationship model between Korean new elderly class’s recovery resilience and productive aging. As of 2013, this study sampled preliminary elderly people in Gyeonggi-do and other provinces nationwide. Data from a total of effective 484 subjects was analyzed. The collected data was processed using the IBM SPSS 20.0 and AMOS 20.0, and underwent descriptive statistical analysis, confirmatory factor analysis, and structure model verification. The path coefficient associated with model fitness was examined. The standardization path coefficient between recovery resilience and productive aging is β=0.975 (t=14.790), revealing a statistically significant positive effect. Thus, it was found that the proposed basic model on the direct path of recovery resilience and productive aging was fit for the model. PMID:26730383
Power analysis on the time effect for the longitudinal Rasch model.
Feddag, M L; Blanchin, M; Hardouin, J B; Sebille, V
2014-01-01
Statistics literature in the social, behavioral, and biomedical sciences typically stress the importance of power analysis. Patient Reported Outcomes (PRO) such as quality of life and other perceived health measures (pain, fatigue, stress,...) are increasingly used as important health outcomes in clinical trials or in epidemiological studies. They cannot be directly observed nor measured as other clinical or biological data and they are often collected through questionnaires with binary or polytomous items. The Rasch model is the well known model in the item response theory (IRT) for binary data. The article proposes an approach to evaluate the statistical power of the time effect for the longitudinal Rasch model with two time points. The performance of this method is compared to the one obtained by simulation study. Finally, the proposed approach is illustrated on one subscale of the SF-36 questionnaire.
Torres, Carlos Henrique; Czeresnia, Dina
2003-01-01
This article aimed at studying the teaching of epidemiology in medical school undergraduate courses. Medical books have been analyzed in order to understand the constitution of epidemiology as a scientific subject, as well as its relations with medical science and secondary school teaching. The introduction and development of the subject in UFRJ Medical School were studied through the analysis of internal regulations, courses brochures and programs. Professors were interviewed. The analysis revealed that epidemiology concepts spread to health services and practice, as well as to medical research. At UFRJ Medical School, starting at the end of the 1980's, epidemiology teaching began to be emphasized and valued. That was the time when collective health began to develop. Among collective health subjects, both epidemiology and statistics developed the most.
Data Acquisition and Preprocessing in Studies on Humans: What Is Not Taught in Statistics Classes?
Zhu, Yeyi; Hernandez, Ladia M; Mueller, Peter; Dong, Yongquan; Forman, Michele R
2013-01-01
The aim of this paper is to address issues in research that may be missing from statistics classes and important for (bio-)statistics students. In the context of a case study, we discuss data acquisition and preprocessing steps that fill the gap between research questions posed by subject matter scientists and statistical methodology for formal inference. Issues include participant recruitment, data collection training and standardization, variable coding, data review and verification, data cleaning and editing, and documentation. Despite the critical importance of these details in research, most of these issues are rarely discussed in an applied statistics program. One reason for the lack of more formal training is the difficulty in addressing the many challenges that can possibly arise in the course of a study in a systematic way. This article can help to bridge this gap between research questions and formal statistical inference by using an illustrative case study for a discussion. We hope that reading and discussing this paper and practicing data preprocessing exercises will sensitize statistics students to these important issues and achieve optimal conduct, quality control, analysis, and interpretation of a study.
Dynamics of essential collective motions in proteins: Theory
NASA Astrophysics Data System (ADS)
Stepanova, Maria
2007-11-01
A general theoretical background is introduced for characterization of conformational motions in protein molecules, and for building reduced coarse-grained models of proteins, based on the statistical analysis of their phase trajectories. Using the projection operator technique, a system of coupled generalized Langevin equations is derived for essential collective coordinates, which are generated by principal component analysis of molecular dynamic trajectories. The number of essential degrees of freedom is not limited in the theory. An explicit analytic relation is established between the generalized Langevin equation for essential collective coordinates and that for the all-atom phase trajectory projected onto the subspace of essential collective degrees of freedom. The theory introduced is applied to identify correlated dynamic domains in a macromolecule and to construct coarse-grained models representing the conformational motions in a protein through a few interacting domains embedded in a dissipative medium. A rigorous theoretical background is provided for identification of dynamic correlated domains in a macromolecule. Examples of domain identification in protein G are given and employed to interpret NMR experiments. Challenges and potential outcomes of the theory are discussed.
Persistence of space radiation induced cytogenetic damage in the blood lymphocytes of astronauts.
George, K; Chappell, L J; Cucinotta, F A
2010-08-14
Cytogenetic damage was assessed in blood lymphocytes from 16 astronauts before and after they participated in long-duration space missions of 3 months or more. The frequency of chromosome damage was measured by fluorescence in situ hybridization (FISH) chromosome painting before flight and at various intervals from a few days to many months after return from the mission. For all individuals, the frequency of chromosome exchanges measured within a month of return from space was higher than their preflight yield. However, some individuals showed a temporal decline in chromosome damage with time after flight. Statistical analysis using combined data for all astronauts indicated a significant overall decreasing trend in total chromosome exchanges with time after flight, although this trend was not seen for all astronauts and the yield of chromosome damage in some individuals actually increased with time after flight. The decreasing trend in total exchanges was slightly more significant when statistical analysis was restricted to data collected more than 220 days after return from flight. When analysis was restricted to data collected within 220 days of return from the mission there was no relationship between total exchanges and time. Translocation yields varied more between astronauts and there was only a slight non-significant decrease with time after flight that was similar for both later and earlier sampling times. Copyright (c) 2010. Published by Elsevier B.V.
Spatio-temporal Genetic Structuring of Leishmania major in Tunisia by Microsatellite Analysis
Harrabi, Myriam; Bettaieb, Jihène; Ghawar, Wissem; Toumi, Amine; Zaâtour, Amor; Yazidi, Rihab; Chaâbane, Sana; Chalghaf, Bilel; Hide, Mallorie; Bañuls, Anne-Laure; Ben Salah, Afif
2015-01-01
In Tunisia, cases of zoonotic cutaneous leishmaniasis caused by Leishmania major are increasing and spreading from the south-west to new areas in the center. To improve the current knowledge on L. major evolution and population dynamics, we performed multi-locus microsatellite typing of human isolates from Tunisian governorates where the disease is endemic (Gafsa, Kairouan and Sidi Bouzid governorates) and collected during two periods: 1991–1992 and 2008–2012. Analysis (F-statistics and Bayesian model-based approach) of the genotyping results of isolates collected in Sidi Bouzid in 1991–1992 and 2008–2012 shows that, over two decades, in the same area, Leishmania parasites evolved by generating genetically differentiated populations. The genetic patterns of 2008–2012 isolates from the three governorates indicate that L. major populations did not spread gradually from the south to the center of Tunisia, according to a geographical gradient, suggesting that human activities might be the source of the disease expansion. The genotype analysis also suggests previous (Bayesian model-based approach) and current (F-statistics) flows of genotypes between governorates and districts. Human activities as well as reservoir dynamics and the effects of environmental changes could explain how the disease progresses. This study provides new insights into the evolution and spread of L. major in Tunisia that might improve our understanding of the parasite flow between geographically and temporally distinct populations. PMID:26302440
Analysis techniques for residual acceleration data
NASA Technical Reports Server (NTRS)
Rogers, Melissa J. B.; Alexander, J. Iwan D.; Snyder, Robert S.
1990-01-01
Various aspects of residual acceleration data are of interest to low-gravity experimenters. Maximum and mean values and various other statistics can be obtained from data as collected in the time domain. Additional information may be obtained through manipulation of the data. Fourier analysis is discussed as a means of obtaining information about dominant frequency components of a given data window. Transformation of data into different coordinate axes is useful in the analysis of experiments with different orientations and can be achieved by the use of a transformation matrix. Application of such analysis techniques to residual acceleration data provides additional information than what is provided in a time history and increases the effectiveness of post-flight analysis of low-gravity experiments.
Developing web-based data analysis tools for precision farming using R and Shiny
NASA Astrophysics Data System (ADS)
Jahanshiri, Ebrahim; Mohd Shariff, Abdul Rashid
2014-06-01
Technologies that are set to increase the productivity of agricultural practices require more and more data. Nevertheless, farming data is also being increasingly cheap to collect and maintain. Bulk of data that are collected by the sensors and samples need to be analysed in an efficient and transparent manner. Web technologies have long being used to develop applications that can assist the farmers and managers. However until recently, analysing the data in an online environment has not been an easy task especially in the eyes of data analysts. This barrier is now overcome by the availability of new application programming interfaces that can provide real-time web based data analysis. In this paper developing a prototype web based application for data analysis using new facilities in R statistical package and its web development facility, Shiny is explored. The pros and cons of this type of data analysis environment for precision farming are enumerated and future directions in web application development for agricultural data are discussed.
Geographic information systems, remote sensing, and spatial analysis activities in Texas, 2002-07
Pearson, D.K.; Gary, R.H.; Wilson, Z.D.
2007-01-01
Geographic information system (GIS) technology has become an important tool for scientific investigation, resource management, and environmental planning. A GIS is a computer-aided system capable of collecting, storing, analyzing, and displaying spatially referenced digital data. GIS technology is particularly useful when analyzing a wide variety of spatial data such as with remote sensing and spatial analysis. Remote sensing involves collecting remotely sensed data, such as satellite imagery, aerial photography, or radar images, and analyzing the data to gather information or investigate trends about the environment or the Earth's surface. Spatial analysis combines remotely sensed, thematic, statistical, quantitative, and geographical data through overlay, modeling, and other analytical techniques to investigate specific research questions. It is the combination of data formats and analysis techniques that has made GIS an essential tool in scientific investigations. This document presents information about the technical capabilities and project activities of the U.S. Geological Survey (USGS) Texas Water Science Center (TWSC) GIS Workgroup from 2002 through 2007.
Johnson, Quentin R; Lindsay, Richard J; Shen, Tongye
2018-02-21
A computational method which extracts the dominant motions from an ensemble of biomolecular conformations via a correlation analysis of residue-residue contacts is presented. The algorithm first renders the structural information into contact matrices, then constructs the collective modes based on the correlated dynamics of a selected set of dynamic contacts. Associated programs can bridge the results for further visualization using graphics software. The aim of this method is to provide an analysis of conformations of biopolymers from the contact viewpoint. It may assist a systematical uncovering of conformational switching mechanisms existing in proteins and biopolymer systems in general by statistical analysis of simulation snapshots. In contrast to conventional correlation analyses of Cartesian coordinates (such as distance covariance analysis and Cartesian principal component analysis), this program also provides an alternative way to locate essential collective motions in general. Herein, we detail the algorithm in a stepwise manner and comment on the importance of the method as applied to decoding allosteric mechanisms. © 2018 Wiley Periodicals, Inc. © 2018 Wiley Periodicals, Inc.
Trajectory selection for the Mariner Jupiter/Saturn 1977 project
NASA Technical Reports Server (NTRS)
Dyer, J. S.; Miles, R. F., Jr.
1974-01-01
The use of decision analysis to facilitate a group decision-making problem in the selection of trajectories for the two spacecraft of the Mariner Jupiter/Saturn 1977 Project. A set of 32 candidate trajectory pairs was developed. Cardinal utility function values were assigned to the trajectory pairs, and the data and statistics derived from collective choice rules were used in selecting the science-preferred trajectory pair.
ERIC Educational Resources Information Center
Palmer, Mary E.
2013-01-01
This study investigated the direct impact of team cohesiveness and student-athletes' perceptions of coaching behavior/leadership functions on the success of NCAA Division I Women's basketball, based on the teams' win/loss records. The research collection was quantitative in nature. Statistical design and analysis provided justification for the use…
James S. Han; Theodore Mianowski; Yi-yu Lin
1999-01-01
The efficacy of fiber length measurement techniques such as digitizing, the Kajaani procedure, and NIH Image are compared in order to determine the optimal tool. Kenaf bast fibers, aspen, and red pine fibers were collected from different anatomical parts, and the fiber lengths were compared using various analytical tools. A statistical analysis on the validity of the...
Defense Safety Oversight Council (DSOC) Reducing Vehicular Vibration and Impact
2013-10-10
Data Collection Analysis/Writing Total Funding = Planned Completion Medical Research and Materiel Command U.S. Army Aeromedical Research...newly introduced to the UK in 2000-2001 • Little was known about long term health effects of monocular helmet mounted displays Purpose • Analyze data ...Collate data then analyze questionnaires and examinations for statistical differences Product/Payoff • Increased knowledge of risks to Apache
ERIC Educational Resources Information Center
Froelich, Amy G.; Nettleton, Dan
2013-01-01
In this article, we present a study to test whether neutral observers perceive a resemblance between a parent and a child. We demonstrate the general approach for two separate parent/ child pairs using survey data collected from introductory statistics students serving as neutral observers. We then present ideas for incorporating the study design…