Evidence for a Global Sampling Process in Extraction of Summary Statistics of Item Sizes in a Set.
Tokita, Midori; Ueda, Sachiyo; Ishiguchi, Akira
2016-01-01
Several studies have shown that our visual system may construct a "summary statistical representation" over groups of visual objects. Although there is a general understanding that human observers can accurately represent sets of a variety of features, many questions on how summary statistics, such as an average, are computed remain unanswered. This study investigated sampling properties of visual information used by human observers to extract two types of summary statistics of item sets, average and variance. We presented three models of ideal observers to extract the summary statistics: a global sampling model without sampling noise, global sampling model with sampling noise, and limited sampling model. We compared the performance of an ideal observer of each model with that of human observers using statistical efficiency analysis. Results suggest that summary statistics of items in a set may be computed without representing individual items, which makes it possible to discard the limited sampling account. Moreover, the extraction of summary statistics may not necessarily require the representation of individual objects with focused attention when the sets of items are larger than 4.
Explorations in statistics: the log transformation.
Curran-Everett, Douglas
2018-06-01
Learning about statistics is a lot like learning about science: the learning is more meaningful if you can actively explore. This thirteenth installment of Explorations in Statistics explores the log transformation, an established technique that rescales the actual observations from an experiment so that the assumptions of some statistical analysis are better met. A general assumption in statistics is that the variability of some response Y is homogeneous across groups or across some predictor variable X. If the variability-the standard deviation-varies in rough proportion to the mean value of Y, a log transformation can equalize the standard deviations. Moreover, if the actual observations from an experiment conform to a skewed distribution, then a log transformation can make the theoretical distribution of the sample mean more consistent with a normal distribution. This is important: the results of a one-sample t test are meaningful only if the theoretical distribution of the sample mean is roughly normal. If we log-transform our observations, then we want to confirm the transformation was useful. We can do this if we use the Box-Cox method, if we bootstrap the sample mean and the statistic t itself, and if we assess the residual plots from the statistical model of the actual and transformed sample observations.
Statistical scaling of geometric characteristics in stochastically generated pore microstructures
Hyman, Jeffrey D.; Guadagnini, Alberto; Winter, C. Larrabee
2015-05-21
In this study, we analyze the statistical scaling of structural attributes of virtual porous microstructures that are stochastically generated by thresholding Gaussian random fields. Characterization of the extent at which randomly generated pore spaces can be considered as representative of a particular rock sample depends on the metrics employed to compare the virtual sample against its physical counterpart. Typically, comparisons against features and/patterns of geometric observables, e.g., porosity and specific surface area, flow-related macroscopic parameters, e.g., permeability, or autocorrelation functions are used to assess the representativeness of a virtual sample, and thereby the quality of the generation method. Here, wemore » rely on manifestations of statistical scaling of geometric observables which were recently observed in real millimeter scale rock samples [13] as additional relevant metrics by which to characterize a virtual sample. We explore the statistical scaling of two geometric observables, namely porosity (Φ) and specific surface area (SSA), of porous microstructures generated using the method of Smolarkiewicz and Winter [42] and Hyman and Winter [22]. Our results suggest that the method can produce virtual pore space samples displaying the symptoms of statistical scaling observed in real rock samples. Order q sample structure functions (statistical moments of absolute increments) of Φ and SSA scale as a power of the separation distance (lag) over a range of lags, and extended self-similarity (linear relationship between log structure functions of successive orders) appears to be an intrinsic property of the generated media. The width of the range of lags where power-law scaling is observed and the Hurst coefficient associated with the variables we consider can be controlled by the generation parameters of the method.« less
The large sample size fallacy.
Lantz, Björn
2013-06-01
Significance in the statistical sense has little to do with significance in the common practical sense. Statistical significance is a necessary but not a sufficient condition for practical significance. Hence, results that are extremely statistically significant may be highly nonsignificant in practice. The degree of practical significance is generally determined by the size of the observed effect, not the p-value. The results of studies based on large samples are often characterized by extreme statistical significance despite small or even trivial effect sizes. Interpreting such results as significant in practice without further analysis is referred to as the large sample size fallacy in this article. The aim of this article is to explore the relevance of the large sample size fallacy in contemporary nursing research. Relatively few nursing articles display explicit measures of observed effect sizes or include a qualitative discussion of observed effect sizes. Statistical significance is often treated as an end in itself. Effect sizes should generally be calculated and presented along with p-values for statistically significant results, and observed effect sizes should be discussed qualitatively through direct and explicit comparisons with the effects in related literature. © 2012 Nordic College of Caring Science.
Olives, Casey; Valadez, Joseph J; Brooker, Simon J; Pagano, Marcello
2012-01-01
Originally a binary classifier, Lot Quality Assurance Sampling (LQAS) has proven to be a useful tool for classification of the prevalence of Schistosoma mansoni into multiple categories (≤10%, >10 and <50%, ≥50%), and semi-curtailed sampling has been shown to effectively reduce the number of observations needed to reach a decision. To date the statistical underpinnings for Multiple Category-LQAS (MC-LQAS) have not received full treatment. We explore the analytical properties of MC-LQAS, and validate its use for the classification of S. mansoni prevalence in multiple settings in East Africa. We outline MC-LQAS design principles and formulae for operating characteristic curves. In addition, we derive the average sample number for MC-LQAS when utilizing semi-curtailed sampling and introduce curtailed sampling in this setting. We also assess the performance of MC-LQAS designs with maximum sample sizes of n=15 and n=25 via a weighted kappa-statistic using S. mansoni data collected in 388 schools from four studies in East Africa. Overall performance of MC-LQAS classification was high (kappa-statistic of 0.87). In three of the studies, the kappa-statistic for a design with n=15 was greater than 0.75. In the fourth study, where these designs performed poorly (kappa-statistic less than 0.50), the majority of observations fell in regions where potential error is known to be high. Employment of semi-curtailed and curtailed sampling further reduced the sample size by as many as 0.5 and 3.5 observations per school, respectively, without increasing classification error. This work provides the needed analytics to understand the properties of MC-LQAS for assessing the prevalance of S. mansoni and shows that in most settings a sample size of 15 children provides a reliable classification of schools.
Chaibub Neto, Elias
2015-01-01
In this paper we propose a vectorized implementation of the non-parametric bootstrap for statistics based on sample moments. Basically, we adopt the multinomial sampling formulation of the non-parametric bootstrap, and compute bootstrap replications of sample moment statistics by simply weighting the observed data according to multinomial counts instead of evaluating the statistic on a resampled version of the observed data. Using this formulation we can generate a matrix of bootstrap weights and compute the entire vector of bootstrap replications with a few matrix multiplications. Vectorization is particularly important for matrix-oriented programming languages such as R, where matrix/vector calculations tend to be faster than scalar operations implemented in a loop. We illustrate the application of the vectorized implementation in real and simulated data sets, when bootstrapping Pearson’s sample correlation coefficient, and compared its performance against two state-of-the-art R implementations of the non-parametric bootstrap, as well as a straightforward one based on a for loop. Our investigations spanned varying sample sizes and number of bootstrap replications. The vectorized bootstrap compared favorably against the state-of-the-art implementations in all cases tested, and was remarkably/considerably faster for small/moderate sample sizes. The same results were observed in the comparison with the straightforward implementation, except for large sample sizes, where the vectorized bootstrap was slightly slower than the straightforward implementation due to increased time expenditures in the generation of weight matrices via multinomial sampling. PMID:26125965
Statistical scaling of pore-scale Lagrangian velocities in natural porous media.
Siena, M; Guadagnini, A; Riva, M; Bijeljic, B; Pereira Nunes, J P; Blunt, M J
2014-08-01
We investigate the scaling behavior of sample statistics of pore-scale Lagrangian velocities in two different rock samples, Bentheimer sandstone and Estaillades limestone. The samples are imaged using x-ray computer tomography with micron-scale resolution. The scaling analysis relies on the study of the way qth-order sample structure functions (statistical moments of order q of absolute increments) of Lagrangian velocities depend on separation distances, or lags, traveled along the mean flow direction. In the sandstone block, sample structure functions of all orders exhibit a power-law scaling within a clearly identifiable intermediate range of lags. Sample structure functions associated with the limestone block display two diverse power-law regimes, which we infer to be related to two overlapping spatially correlated structures. In both rocks and for all orders q, we observe linear relationships between logarithmic structure functions of successive orders at all lags (a phenomenon that is typically known as extended power scaling, or extended self-similarity). The scaling behavior of Lagrangian velocities is compared with the one exhibited by porosity and specific surface area, which constitute two key pore-scale geometric observables. The statistical scaling of the local velocity field reflects the behavior of these geometric observables, with the occurrence of power-law-scaling regimes within the same range of lags for sample structure functions of Lagrangian velocity, porosity, and specific surface area.
ERIC Educational Resources Information Center
Ali, Usama S.; Walker, Michael E.
2014-01-01
Two methods are currently in use at Educational Testing Service (ETS) for equating observed item difficulty statistics. The first method involves the linear equating of item statistics in an observed sample to reference statistics on the same items. The second method, or the item response curve (IRC) method, involves the summation of conditional…
Olives, Casey; Valadez, Joseph J.; Brooker, Simon J.; Pagano, Marcello
2012-01-01
Background Originally a binary classifier, Lot Quality Assurance Sampling (LQAS) has proven to be a useful tool for classification of the prevalence of Schistosoma mansoni into multiple categories (≤10%, >10 and <50%, ≥50%), and semi-curtailed sampling has been shown to effectively reduce the number of observations needed to reach a decision. To date the statistical underpinnings for Multiple Category-LQAS (MC-LQAS) have not received full treatment. We explore the analytical properties of MC-LQAS, and validate its use for the classification of S. mansoni prevalence in multiple settings in East Africa. Methodology We outline MC-LQAS design principles and formulae for operating characteristic curves. In addition, we derive the average sample number for MC-LQAS when utilizing semi-curtailed sampling and introduce curtailed sampling in this setting. We also assess the performance of MC-LQAS designs with maximum sample sizes of n = 15 and n = 25 via a weighted kappa-statistic using S. mansoni data collected in 388 schools from four studies in East Africa. Principle Findings Overall performance of MC-LQAS classification was high (kappa-statistic of 0.87). In three of the studies, the kappa-statistic for a design with n = 15 was greater than 0.75. In the fourth study, where these designs performed poorly (kappa-statistic less than 0.50), the majority of observations fell in regions where potential error is known to be high. Employment of semi-curtailed and curtailed sampling further reduced the sample size by as many as 0.5 and 3.5 observations per school, respectively, without increasing classification error. Conclusion/Significance This work provides the needed analytics to understand the properties of MC-LQAS for assessing the prevalance of S. mansoni and shows that in most settings a sample size of 15 children provides a reliable classification of schools. PMID:22970333
The effect of sampling rate on observed statistics in a correlated random walk
Rosser, G.; Fletcher, A. G.; Maini, P. K.; Baker, R. E.
2013-01-01
Tracking the movement of individual cells or animals can provide important information about their motile behaviour, with key examples including migrating birds, foraging mammals and bacterial chemotaxis. In many experimental protocols, observations are recorded with a fixed sampling interval and the continuous underlying motion is approximated as a series of discrete steps. The size of the sampling interval significantly affects the tracking measurements, the statistics computed from observed trajectories, and the inferences drawn. Despite the widespread use of tracking data to investigate motile behaviour, many open questions remain about these effects. We use a correlated random walk model to study the variation with sampling interval of two key quantities of interest: apparent speed and angle change. Two variants of the model are considered, in which reorientations occur instantaneously and with a stationary pause, respectively. We employ stochastic simulations to study the effect of sampling on the distributions of apparent speeds and angle changes, and present novel mathematical analysis in the case of rapid sampling. Our investigation elucidates the complex nature of sampling effects for sampling intervals ranging over many orders of magnitude. Results show that inclusion of a stationary phase significantly alters the observed distributions of both quantities. PMID:23740484
Exploring Tree Age & Diameter to Illustrate Sample Design & Inference in Observational Ecology
ERIC Educational Resources Information Center
Casady, Grant M.
2015-01-01
Undergraduate biology labs often explore the techniques of data collection but neglect the statistical framework necessary to express findings. Students can be confused about how to use their statistical knowledge to address specific biological questions. Growth in the area of observational ecology requires that students gain experience in…
NASA Astrophysics Data System (ADS)
Agus, M.; Hitchcott, P. K.; Penna, M. P.; Peró-Cebollero, M.; Guàrdia-Olmos, J.
2016-11-01
Many studies have investigated the features of probabilistic reasoning developed in relation to different formats of problem presentation, showing that it is affected by various individual and contextual factors. Incomplete understanding of the identity and role of these factors may explain the inconsistent evidence concerning the effect of problem presentation format. Thus, superior performance has sometimes been observed for graphically, rather than verbally, presented problems. The present study was undertaken to address this issue. Psychology undergraduates without any statistical expertise (N = 173 in Italy; N = 118 in Spain; N = 55 in England) were administered statistical problems in two formats (verbal-numerical and graphical-pictorial) under a condition of time pressure. Students also completed additional measures indexing several potentially relevant individual dimensions (statistical ability, statistical anxiety, attitudes towards statistics and confidence). Interestingly, a facilitatory effect of graphical presentation was observed in the Italian and Spanish samples but not in the English one. Significantly, the individual dimensions predicting statistical performance also differed between the samples, highlighting a different role of confidence. Hence, these findings confirm previous observations concerning problem presentation format while simultaneously highlighting the importance of individual dimensions.
Hagell, Peter; Westergren, Albert
Sample size is a major factor in statistical null hypothesis testing, which is the basis for many approaches to testing Rasch model fit. Few sample size recommendations for testing fit to the Rasch model concern the Rasch Unidimensional Measurement Models (RUMM) software, which features chi-square and ANOVA/F-ratio based fit statistics, including Bonferroni and algebraic sample size adjustments. This paper explores the occurrence of Type I errors with RUMM fit statistics, and the effects of algebraic sample size adjustments. Data with simulated Rasch model fitting 25-item dichotomous scales and sample sizes ranging from N = 50 to N = 2500 were analysed with and without algebraically adjusted sample sizes. Results suggest the occurrence of Type I errors with N less then or equal to 500, and that Bonferroni correction as well as downward algebraic sample size adjustment are useful to avoid such errors, whereas upward adjustment of smaller samples falsely signal misfit. Our observations suggest that sample sizes around N = 250 to N = 500 may provide a good balance for the statistical interpretation of the RUMM fit statistics studied here with respect to Type I errors and under the assumption of Rasch model fit within the examined frame of reference (i.e., about 25 item parameters well targeted to the sample).
Micro-organism distribution sampling for bioassays
NASA Technical Reports Server (NTRS)
Nelson, B. A.
1975-01-01
Purpose of sampling distribution is to characterize sample-to-sample variation so statistical tests may be applied, to estimate error due to sampling (confidence limits) and to evaluate observed differences between samples. Distribution could be used for bioassays taken in hospitals, breweries, food-processing plants, and pharmaceutical plants.
Why weight? Modelling sample and observational level variability improves power in RNA-seq analyses
Liu, Ruijie; Holik, Aliaksei Z.; Su, Shian; Jansz, Natasha; Chen, Kelan; Leong, Huei San; Blewitt, Marnie E.; Asselin-Labat, Marie-Liesse; Smyth, Gordon K.; Ritchie, Matthew E.
2015-01-01
Variations in sample quality are frequently encountered in small RNA-sequencing experiments, and pose a major challenge in a differential expression analysis. Removal of high variation samples reduces noise, but at a cost of reducing power, thus limiting our ability to detect biologically meaningful changes. Similarly, retaining these samples in the analysis may not reveal any statistically significant changes due to the higher noise level. A compromise is to use all available data, but to down-weight the observations from more variable samples. We describe a statistical approach that facilitates this by modelling heterogeneity at both the sample and observational levels as part of the differential expression analysis. At the sample level this is achieved by fitting a log-linear variance model that includes common sample-specific or group-specific parameters that are shared between genes. The estimated sample variance factors are then converted to weights and combined with observational level weights obtained from the mean–variance relationship of the log-counts-per-million using ‘voom’. A comprehensive analysis involving both simulations and experimental RNA-sequencing data demonstrates that this strategy leads to a universally more powerful analysis and fewer false discoveries when compared to conventional approaches. This methodology has wide application and is implemented in the open-source ‘limma’ package. PMID:25925576
Statistical and sampling issues when using multiple particle tracking
NASA Astrophysics Data System (ADS)
Savin, Thierry; Doyle, Patrick S.
2007-08-01
Video microscopy can be used to simultaneously track several microparticles embedded in a complex material. The trajectories are used to extract a sample of displacements at random locations in the material. From this sample, averaged quantities characterizing the dynamics of the probes are calculated to evaluate structural and/or mechanical properties of the assessed material. However, the sampling of measured displacements in heterogeneous systems is singular because the volume of observation with video microscopy is finite. By carefully characterizing the sampling design in the experimental output of the multiple particle tracking technique, we derive estimators for the mean and variance of the probes’ dynamics that are independent of the peculiar statistical characteristics. We expose stringent tests of these estimators using simulated and experimental complex systems with a known heterogeneous structure. Up to a certain fundamental limitation, which we characterize through a material degree of sampling by the embedded probe tracking, these estimators can be applied to quantify the heterogeneity of a material, providing an original and intelligible kind of information on complex fluid properties. More generally, we show that the precise assessment of the statistics in the multiple particle tracking output sample of observations is essential in order to provide accurate unbiased measurements.
NASA Technical Reports Server (NTRS)
Bell, Thomas L.; Abdullah, A.; Martin, Russell L.; North, Gerald R.
1990-01-01
Estimates of monthly average rainfall based on satellite observations from a low earth orbit will differ from the true monthly average because the satellite observes a given area only intermittently. This sampling error inherent in satellite monitoring of rainfall would occur even if the satellite instruments could measure rainfall perfectly. The size of this error is estimated for a satellite system being studied at NASA, the Tropical Rainfall Measuring Mission (TRMM). First, the statistical description of rainfall on scales from 1 to 1000 km is examined in detail, based on rainfall data from the Global Atmospheric Research Project Atlantic Tropical Experiment (GATE). A TRMM-like satellite is flown over a two-dimensional time-evolving simulation of rainfall using a stochastic model with statistics tuned to agree with GATE statistics. The distribution of sampling errors found from many months of simulated observations is found to be nearly normal, even though the distribution of area-averaged rainfall is far from normal. For a range of orbits likely to be employed in TRMM, sampling error is found to be less than 10 percent of the mean for rainfall averaged over a 500 x 500 sq km area.
A Third Moment Adjusted Test Statistic for Small Sample Factor Analysis.
Lin, Johnny; Bentler, Peter M
2012-01-01
Goodness of fit testing in factor analysis is based on the assumption that the test statistic is asymptotically chi-square; but this property may not hold in small samples even when the factors and errors are normally distributed in the population. Robust methods such as Browne's asymptotically distribution-free method and Satorra Bentler's mean scaling statistic were developed under the presumption of non-normality in the factors and errors. This paper finds new application to the case where factors and errors are normally distributed in the population but the skewness of the obtained test statistic is still high due to sampling error in the observed indicators. An extension of Satorra Bentler's statistic is proposed that not only scales the mean but also adjusts the degrees of freedom based on the skewness of the obtained test statistic in order to improve its robustness under small samples. A simple simulation study shows that this third moment adjusted statistic asymptotically performs on par with previously proposed methods, and at a very small sample size offers superior Type I error rates under a properly specified model. Data from Mardia, Kent and Bibby's study of students tested for their ability in five content areas that were either open or closed book were used to illustrate the real-world performance of this statistic.
Wu, Robert; Glen, Peter; Ramsay, Tim; Martel, Guillaume
2014-06-28
Observational studies dominate the surgical literature. Statistical adjustment is an important strategy to account for confounders in observational studies. Research has shown that published articles are often poor in statistical quality, which may jeopardize their conclusions. The Statistical Analyses and Methods in the Published Literature (SAMPL) guidelines have been published to help establish standards for statistical reporting.This study will seek to determine whether the quality of statistical adjustment and the reporting of these methods are adequate in surgical observational studies. We hypothesize that incomplete reporting will be found in all surgical observational studies, and that the quality and reporting of these methods will be of lower quality in surgical journals when compared with medical journals. Finally, this work will seek to identify predictors of high-quality reporting. This work will examine the top five general surgical and medical journals, based on a 5-year impact factor (2007-2012). All observational studies investigating an intervention related to an essential component area of general surgery (defined by the American Board of Surgery), with an exposure, outcome, and comparator, will be included in this systematic review. Essential elements related to statistical reporting and quality were extracted from the SAMPL guidelines and include domains such as intent of analysis, primary analysis, multiple comparisons, numbers and descriptive statistics, association and correlation analyses, linear regression, logistic regression, Cox proportional hazard analysis, analysis of variance, survival analysis, propensity analysis, and independent and correlated analyses. Each article will be scored as a proportion based on fulfilling criteria in relevant analyses used in the study. A logistic regression model will be built to identify variables associated with high-quality reporting. A comparison will be made between the scores of surgical observational studies published in medical versus surgical journals. Secondary outcomes will pertain to individual domains of analysis. Sensitivity analyses will be conducted. This study will explore the reporting and quality of statistical analyses in surgical observational studies published in the most referenced surgical and medical journals in 2013 and examine whether variables (including the type of journal) can predict high-quality reporting.
Statistical Inference for Data Adaptive Target Parameters.
Hubbard, Alan E; Kherad-Pajouh, Sara; van der Laan, Mark J
2016-05-01
Consider one observes n i.i.d. copies of a random variable with a probability distribution that is known to be an element of a particular statistical model. In order to define our statistical target we partition the sample in V equal size sub-samples, and use this partitioning to define V splits in an estimation sample (one of the V subsamples) and corresponding complementary parameter-generating sample. For each of the V parameter-generating samples, we apply an algorithm that maps the sample to a statistical target parameter. We define our sample-split data adaptive statistical target parameter as the average of these V-sample specific target parameters. We present an estimator (and corresponding central limit theorem) of this type of data adaptive target parameter. This general methodology for generating data adaptive target parameters is demonstrated with a number of practical examples that highlight new opportunities for statistical learning from data. This new framework provides a rigorous statistical methodology for both exploratory and confirmatory analysis within the same data. Given that more research is becoming "data-driven", the theory developed within this paper provides a new impetus for a greater involvement of statistical inference into problems that are being increasingly addressed by clever, yet ad hoc pattern finding methods. To suggest such potential, and to verify the predictions of the theory, extensive simulation studies, along with a data analysis based on adaptively determined intervention rules are shown and give insight into how to structure such an approach. The results show that the data adaptive target parameter approach provides a general framework and resulting methodology for data-driven science.
Why weight? Modelling sample and observational level variability improves power in RNA-seq analyses.
Liu, Ruijie; Holik, Aliaksei Z; Su, Shian; Jansz, Natasha; Chen, Kelan; Leong, Huei San; Blewitt, Marnie E; Asselin-Labat, Marie-Liesse; Smyth, Gordon K; Ritchie, Matthew E
2015-09-03
Variations in sample quality are frequently encountered in small RNA-sequencing experiments, and pose a major challenge in a differential expression analysis. Removal of high variation samples reduces noise, but at a cost of reducing power, thus limiting our ability to detect biologically meaningful changes. Similarly, retaining these samples in the analysis may not reveal any statistically significant changes due to the higher noise level. A compromise is to use all available data, but to down-weight the observations from more variable samples. We describe a statistical approach that facilitates this by modelling heterogeneity at both the sample and observational levels as part of the differential expression analysis. At the sample level this is achieved by fitting a log-linear variance model that includes common sample-specific or group-specific parameters that are shared between genes. The estimated sample variance factors are then converted to weights and combined with observational level weights obtained from the mean-variance relationship of the log-counts-per-million using 'voom'. A comprehensive analysis involving both simulations and experimental RNA-sequencing data demonstrates that this strategy leads to a universally more powerful analysis and fewer false discoveries when compared to conventional approaches. This methodology has wide application and is implemented in the open-source 'limma' package. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.
Mechanical properties of silicate glasses exposed to a low-Earth orbit
NASA Technical Reports Server (NTRS)
Wiedlocher, David E.; Tucker, Dennis S.; Nichols, Ron; Kinser, Donald L.
1992-01-01
The effects of a 5.8 year exposure to low earth orbit environment upon the mechanical properties of commercial optical fused silica, low iron soda-lime-silica, Pyrex 7740, Vycor 7913, BK-7, and the glass ceramic Zerodur were examined. Mechanical testing employed the ASTM-F-394 piston on 3-ball method in a liquid nitrogen environment. Samples were exposed on the Long Duration Exposure Facility (LDEF) in two locations. Impacts were observed on all specimens except Vycor. Weibull analysis as well as a standard statistical evaluation were conducted. The Weibull analysis revealed no differences between control samples and the two exposed samples. We thus concluded that radiation components of the Earth orbital environment did not degrade the mechanical strength of the samples examined within the limits of experimental error. The upper bound of strength degradation for meteorite impacted samples based upon statistical analysis and observation was 50 percent.
Statistics 101 for Radiologists.
Anvari, Arash; Halpern, Elkan F; Samir, Anthony E
2015-10-01
Diagnostic tests have wide clinical applications, including screening, diagnosis, measuring treatment effect, and determining prognosis. Interpreting diagnostic test results requires an understanding of key statistical concepts used to evaluate test efficacy. This review explains descriptive statistics and discusses probability, including mutually exclusive and independent events and conditional probability. In the inferential statistics section, a statistical perspective on study design is provided, together with an explanation of how to select appropriate statistical tests. Key concepts in recruiting study samples are discussed, including representativeness and random sampling. Variable types are defined, including predictor, outcome, and covariate variables, and the relationship of these variables to one another. In the hypothesis testing section, we explain how to determine if observed differences between groups are likely to be due to chance. We explain type I and II errors, statistical significance, and study power, followed by an explanation of effect sizes and how confidence intervals can be used to generalize observed effect sizes to the larger population. Statistical tests are explained in four categories: t tests and analysis of variance, proportion analysis tests, nonparametric tests, and regression techniques. We discuss sensitivity, specificity, accuracy, receiver operating characteristic analysis, and likelihood ratios. Measures of reliability and agreement, including κ statistics, intraclass correlation coefficients, and Bland-Altman graphs and analysis, are introduced. © RSNA, 2015.
NASA Technical Reports Server (NTRS)
Jasperson, W. H.; Nastrom, G. D.; Davis, R. E.; Holdeman, J. D.
1984-01-01
Summary studies are presented for the entire cloud observation archieve from the NASA Global Atmospheric Sampling Program (GASP). Studies are also presented for GASP particle concentration data gathered concurrently with the cloud observations. Cloud encounters are shown on about 15 percent of the data samples overall, but the probability of cloud encounter is shown to vary significantly with altitude, latitude, and distance from the tropopause. Several meteorological circulation features are apparent in the latitudinal distribution of cloud cover, and the cloud encounter statistics are shown to be consistent with the classical mid-latitude cyclone model. Observations of clouds spaced more closely than 90 minutes are shown to be statistically dependent. The statistics for cloud and particle encounter are utilized to estimate the frequency of cloud encounter on long range airline routes, and to assess the probability and extent of laminar flow loss due to cloud or particle encounter by aircraft utilizing laminar flow control (LFC). It is shown that the probability of extended cloud encounter is too low, of itself, to make LFC impractical.
A Third Moment Adjusted Test Statistic for Small Sample Factor Analysis
Lin, Johnny; Bentler, Peter M.
2012-01-01
Goodness of fit testing in factor analysis is based on the assumption that the test statistic is asymptotically chi-square; but this property may not hold in small samples even when the factors and errors are normally distributed in the population. Robust methods such as Browne’s asymptotically distribution-free method and Satorra Bentler’s mean scaling statistic were developed under the presumption of non-normality in the factors and errors. This paper finds new application to the case where factors and errors are normally distributed in the population but the skewness of the obtained test statistic is still high due to sampling error in the observed indicators. An extension of Satorra Bentler’s statistic is proposed that not only scales the mean but also adjusts the degrees of freedom based on the skewness of the obtained test statistic in order to improve its robustness under small samples. A simple simulation study shows that this third moment adjusted statistic asymptotically performs on par with previously proposed methods, and at a very small sample size offers superior Type I error rates under a properly specified model. Data from Mardia, Kent and Bibby’s study of students tested for their ability in five content areas that were either open or closed book were used to illustrate the real-world performance of this statistic. PMID:23144511
Saini, Komal; Singh, Parminder; Bajwa, Bikramjit Singh
2016-12-01
LED flourimeter has been used for microanalysis of uranium concentration in groundwater samples collected from six districts of South West (SW), West (W) and North East (NE) Punjab, India. Average value of uranium content in water samples of SW Punjab is observed to be higher than WHO, USEPA recommended safe limit of 30µgl -1 as well as AERB proposed limit of 60µgl -1 . Whereas, for W and NE region of Punjab, average level of uranium concentration was within AERB recommended limit of 60µgl -1 . Average value observed in SW Punjab is around 3-4 times the value observed in W Punjab, whereas its value is more than 17 times the average value observed in NE region of Punjab. Statistical analysis of carcinogenic as well as non carcinogenic risks due to uranium have been evaluated for each studied district. Copyright © 2016 Elsevier Ltd. All rights reserved.
Valid statistical inference methods for a case-control study with missing data.
Tian, Guo-Liang; Zhang, Chi; Jiang, Xuejun
2018-04-01
The main objective of this paper is to derive the valid sampling distribution of the observed counts in a case-control study with missing data under the assumption of missing at random by employing the conditional sampling method and the mechanism augmentation method. The proposed sampling distribution, called the case-control sampling distribution, can be used to calculate the standard errors of the maximum likelihood estimates of parameters via the Fisher information matrix and to generate independent samples for constructing small-sample bootstrap confidence intervals. Theoretical comparisons of the new case-control sampling distribution with two existing sampling distributions exhibit a large difference. Simulations are conducted to investigate the influence of the three different sampling distributions on statistical inferences. One finding is that the conclusion by the Wald test for testing independency under the two existing sampling distributions could be completely different (even contradictory) from the Wald test for testing the equality of the success probabilities in control/case groups under the proposed distribution. A real cervical cancer data set is used to illustrate the proposed statistical methods.
Pearson-type goodness-of-fit test with bootstrap maximum likelihood estimation.
Yin, Guosheng; Ma, Yanyuan
2013-01-01
The Pearson test statistic is constructed by partitioning the data into bins and computing the difference between the observed and expected counts in these bins. If the maximum likelihood estimator (MLE) of the original data is used, the statistic generally does not follow a chi-squared distribution or any explicit distribution. We propose a bootstrap-based modification of the Pearson test statistic to recover the chi-squared distribution. We compute the observed and expected counts in the partitioned bins by using the MLE obtained from a bootstrap sample. This bootstrap-sample MLE adjusts exactly the right amount of randomness to the test statistic, and recovers the chi-squared distribution. The bootstrap chi-squared test is easy to implement, as it only requires fitting exactly the same model to the bootstrap data to obtain the corresponding MLE, and then constructs the bin counts based on the original data. We examine the test size and power of the new model diagnostic procedure using simulation studies and illustrate it with a real data set.
[Application of statistics on chronic-diseases-relating observational research papers].
Hong, Zhi-heng; Wang, Ping; Cao, Wei-hua
2012-09-01
To study the application of statistics on Chronic-diseases-relating observational research papers which were recently published in the Chinese Medical Association Magazines, with influential index above 0.5. Using a self-developed criterion, two investigators individually participated in assessing the application of statistics on Chinese Medical Association Magazines, with influential index above 0.5. Different opinions reached an agreement through discussion. A total number of 352 papers from 6 magazines, including the Chinese Journal of Epidemiology, Chinese Journal of Oncology, Chinese Journal of Preventive Medicine, Chinese Journal of Cardiology, Chinese Journal of Internal Medicine and Chinese Journal of Endocrinology and Metabolism, were reviewed. The rate of clear statement on the following contents as: research objectives, t target audience, sample issues, objective inclusion criteria and variable definitions were 99.43%, 98.57%, 95.43%, 92.86% and 96.87%. The correct rates of description on quantitative and qualitative data were 90.94% and 91.46%, respectively. The rates on correctly expressing the results, on statistical inference methods related to quantitative, qualitative data and modeling were 100%, 95.32% and 87.19%, respectively. 89.49% of the conclusions could directly response to the research objectives. However, 69.60% of the papers did not mention the exact names of the study design, statistically, that the papers were using. 11.14% of the papers were in lack of further statement on the exclusion criteria. Percentage of the papers that could clearly explain the sample size estimation only taking up as 5.16%. Only 24.21% of the papers clearly described the variable value assignment. Regarding the introduction on statistical conduction and on database methods, the rate was only 24.15%. 18.75% of the papers did not express the statistical inference methods sufficiently. A quarter of the papers did not use 'standardization' appropriately. As for the aspect of statistical inference, the rate of description on statistical testing prerequisite was only 24.12% while 9.94% papers did not even employ the statistical inferential method that should be used. The main deficiencies on the application of Statistics used in papers related to Chronic-diseases-related observational research were as follows: lack of sample-size determination, variable value assignment description not sufficient, methods on statistics were not introduced clearly or properly, lack of consideration for pre-requisition regarding the use of statistical inferences.
Searching for the 3.5 keV Line in the Stacked Suzaku Observations of Galaxy Clusters
NASA Technical Reports Server (NTRS)
Bulbul, Esra; Markevitch, Maxim; Foster, Adam; Miller, Eric; Bautz, Mark; Lowenstein, Mike; Randall, Scott W.; Smith, Randall K.
2016-01-01
We perform a detailed study of the stacked Suzaku observations of 47 galaxy clusters, spanning a redshift range of 0.01-0.45, to search for the unidentified 3.5 keV line. This sample provides an independent test for the previously detected line. We detect a 2sigma-significant spectral feature at 3.5 keV in the spectrum of the full sample. When the sample is divided into two subsamples (cool-core and non-cool core clusters), the cool-core subsample shows no statistically significant positive residuals at the line energy. A very weak (approx. 2sigma confidence) spectral feature at 3.5 keV is permitted by the data from the non-cool-core clusters sample. The upper limit on a neutrino decay mixing angle of sin(sup 2)(2theta) = 6.1 x 10(exp -11) from the full Suzaku sample is consistent with the previous detections in the stacked XMM-Newton sample of galaxy clusters (which had a higher statistical sensitivity to faint lines), M31, and Galactic center, at a 90% confidence level. However, the constraint from the present sample, which does not include the Perseus cluster, is in tension with previously reported line flux observed in the core of the Perseus cluster with XMM-Newton and Suzaku.
NASA Technical Reports Server (NTRS)
Jasperson, W. H.; Nastron, G. D.; Davis, R. E.; Holdeman, J. D.
1984-01-01
Summary studies are presented for the entire cloud observation archive from the NASA Global Atmospheric Sampling Program (GASP). Studies are also presented for GASP particle-concentration data gathered concurrently with the cloud observations. Cloud encounters are shown on about 15 percent of the data samples overall, but the probability of cloud encounter is shown to vary significantly with altitude, latitude, and distance from the tropopause. Several meteorological circulation features are apparent in the latitudinal distribution of cloud cover, and the cloud-encounter statistics are shown to be consistent with the classical mid-latitude cyclone model. Observations of clouds spaced more closely than 90 minutes are shown to be statistically dependent. The statistics for cloud and particle encounter are utilized to estimate the frequency of cloud encounter on long-range airline routes, and to assess the probability and extent of laminaar flow loss due to cloud or particle encounter by aircraft utilizing laminar flow control (LFC). It is shown that the probability of extended cloud encounter is too low, of itself, to make LFC impractical. This report is presented in two volumes. Volume I contains the narrative, analysis, and conclusions. Volume II contains five supporting appendixes.
Order statistics applied to the most massive and most distant galaxy clusters
NASA Astrophysics Data System (ADS)
Waizmann, J.-C.; Ettori, S.; Bartelmann, M.
2013-06-01
In this work, we present an analytic framework for calculating the individual and joint distributions of the nth most massive or nth highest redshift galaxy cluster for a given survey characteristic allowing us to formulate Λ cold dark matter (ΛCDM) exclusion criteria. We show that the cumulative distribution functions steepen with increasing order, giving them a higher constraining power with respect to the extreme value statistics. Additionally, we find that the order statistics in mass (being dominated by clusters at lower redshifts) is sensitive to the matter density and the normalization of the matter fluctuations, whereas the order statistics in redshift is particularly sensitive to the geometric evolution of the Universe. For a fixed cosmology, both order statistics are efficient probes of the functional shape of the mass function at the high-mass end. To allow a quick assessment of both order statistics, we provide fits as a function of the survey area that allow percentile estimation with an accuracy better than 2 per cent. Furthermore, we discuss the joint distributions in the two-dimensional case and find that for the combination of the largest and the second largest observation, it is most likely to find them to be realized with similar values with a broadly peaked distribution. When combining the largest observation with higher orders, it is more likely to find a larger gap between the observations and when combining higher orders in general, the joint probability density function peaks more strongly. Having introduced the theory, we apply the order statistical analysis to the Southpole Telescope (SPT) massive cluster sample and metacatalogue of X-ray detected clusters of galaxies catalogue and find that the 10 most massive clusters in the sample are consistent with ΛCDM and the Tinker mass function. For the order statistics in redshift, we find a discrepancy between the data and the theoretical distributions, which could in principle indicate a deviation from the standard cosmology. However, we attribute this deviation to the uncertainty in the modelling of the SPT survey selection function. In turn, by assuming the ΛCDM reference cosmology, order statistics can also be utilized for consistency checks of the completeness of the observed sample and of the modelling of the survey selection function.
Sample sizes and model comparison metrics for species distribution models
B.B. Hanberry; H.S. He; D.C. Dey
2012-01-01
Species distribution models use small samples to produce continuous distribution maps. The question of how small a sample can be to produce an accurate model generally has been answered based on comparisons to maximum sample sizes of 200 observations or fewer. In addition, model comparisons often are made with the kappa statistic, which has become controversial....
NASA Technical Reports Server (NTRS)
Alberts, J. R.; Burden, H. W.; Hawes, N.; Ronca, A. E.
1996-01-01
To assess prenatal and postnatal developmental status in the offspring of a group of animals, it is typical to examine fetuses from some of the dams as well as infants born to the remaining dams. Statistical limitations often arise, particularly when the animals are rare or especially precious, because all offspring of the dam represent only a single statistical observation; littermates are not independent observations (biologically or statistically). We describe a study in which pregnant laboratory rats were laparotomized on day 7 of gestation (GD7) to ascertain the number and distribution of uterine implantation sites and were subjected to a simulated experience on a 10-day space shuttle flight. After the simulated landing on GD18, rats were unilaterally hysterectomized, thus providing a sample of fetuses from 10 independent uteruses, followed by successful vaginal delivery on GD22, yielding postnatal samples from 10 uteruses. A broad profile of maternal and offspring morphologic and physiologic measures indicated that these novel sampling procedures did not compromise maternal well-being and maintained normal offspring development and function. Measures included maternal organ weights and hormone concentrations, offspring body size, growth, organ weights, sexual differentiation, and catecholamine concentrations.
Austin, Peter C; Schuster, Tibor; Platt, Robert W
2015-10-15
Estimating statistical power is an important component of the design of both randomized controlled trials (RCTs) and observational studies. Methods for estimating statistical power in RCTs have been well described and can be implemented simply. In observational studies, statistical methods must be used to remove the effects of confounding that can occur due to non-random treatment assignment. Inverse probability of treatment weighting (IPTW) using the propensity score is an attractive method for estimating the effects of treatment using observational data. However, sample size and power calculations have not been adequately described for these methods. We used an extensive series of Monte Carlo simulations to compare the statistical power of an IPTW analysis of an observational study with time-to-event outcomes with that of an analysis of a similarly-structured RCT. We examined the impact of four factors on the statistical power function: number of observed events, prevalence of treatment, the marginal hazard ratio, and the strength of the treatment-selection process. We found that, on average, an IPTW analysis had lower statistical power compared to an analysis of a similarly-structured RCT. The difference in statistical power increased as the magnitude of the treatment-selection model increased. The statistical power of an IPTW analysis tended to be lower than the statistical power of a similarly-structured RCT.
NASA Technical Reports Server (NTRS)
Melbourne, William G.
1986-01-01
In double differencing a regression system obtained from concurrent Global Positioning System (GPS) observation sequences, one either undersamples the system to avoid introducing colored measurement statistics, or one fully samples the system incurring the resulting non-diagonal covariance matrix for the differenced measurement errors. A suboptimal estimation result will be obtained in the undersampling case and will also be obtained in the fully sampled case unless the color noise statistics are taken into account. The latter approach requires a least squares weighting matrix derived from inversion of a non-diagonal covariance matrix for the differenced measurement errors instead of inversion of the customary diagonal one associated with white noise processes. Presented is the so-called fully redundant double differencing algorithm for generating a weighted double differenced regression system that yields equivalent estimation results, but features for certain cases a diagonal weighting matrix even though the differenced measurement error statistics are highly colored.
Min and Max Exponential Extreme Interval Values and Statistics
ERIC Educational Resources Information Center
Jance, Marsha; Thomopoulos, Nick
2009-01-01
The extreme interval values and statistics (expected value, median, mode, standard deviation, and coefficient of variation) for the smallest (min) and largest (max) values of exponentially distributed variables with parameter ? = 1 are examined for different observation (sample) sizes. An extreme interval value g[subscript a] is defined as a…
Analysis of Variance with Summary Statistics in Microsoft® Excel®
ERIC Educational Resources Information Center
Larson, David A.; Hsu, Ko-Cheng
2010-01-01
Students regularly are asked to solve Single Factor Analysis of Variance problems given only the sample summary statistics (number of observations per category, category means, and corresponding category standard deviations). Most undergraduate students today use Excel for data analysis of this type. However, Excel, like all other statistical…
A random-sum Wilcoxon statistic and its application to analysis of ROC and LROC data.
Tang, Liansheng Larry; Balakrishnan, N
2011-01-01
The Wilcoxon-Mann-Whitney statistic is commonly used for a distribution-free comparison of two groups. One requirement for its use is that the sample sizes of the two groups are fixed. This is violated in some of the applications such as medical imaging studies and diagnostic marker studies; in the former, the violation occurs since the number of correctly localized abnormal images is random, while in the latter the violation is due to some subjects not having observable measurements. For this reason, we propose here a random-sum Wilcoxon statistic for comparing two groups in the presence of ties, and derive its variance as well as its asymptotic distribution for large sample sizes. The proposed statistic includes the regular Wilcoxon rank-sum statistic. Finally, we apply the proposed statistic for summarizing location response operating characteristic data from a liver computed tomography study, and also for summarizing diagnostic accuracy of biomarker data.
Robust functional statistics applied to Probability Density Function shape screening of sEMG data.
Boudaoud, S; Rix, H; Al Harrach, M; Marin, F
2014-01-01
Recent studies pointed out possible shape modifications of the Probability Density Function (PDF) of surface electromyographical (sEMG) data according to several contexts like fatigue and muscle force increase. Following this idea, criteria have been proposed to monitor these shape modifications mainly using High Order Statistics (HOS) parameters like skewness and kurtosis. In experimental conditions, these parameters are confronted with small sample size in the estimation process. This small sample size induces errors in the estimated HOS parameters restraining real-time and precise sEMG PDF shape monitoring. Recently, a functional formalism, the Core Shape Model (CSM), has been used to analyse shape modifications of PDF curves. In this work, taking inspiration from CSM method, robust functional statistics are proposed to emulate both skewness and kurtosis behaviors. These functional statistics combine both kernel density estimation and PDF shape distances to evaluate shape modifications even in presence of small sample size. Then, the proposed statistics are tested, using Monte Carlo simulations, on both normal and Log-normal PDFs that mimic observed sEMG PDF shape behavior during muscle contraction. According to the obtained results, the functional statistics seem to be more robust than HOS parameters to small sample size effect and more accurate in sEMG PDF shape screening applications.
Testing for independence in J×K contingency tables with complex sample survey data.
Lipsitz, Stuart R; Fitzmaurice, Garrett M; Sinha, Debajyoti; Hevelone, Nathanael; Giovannucci, Edward; Hu, Jim C
2015-09-01
The test of independence of row and column variables in a (J×K) contingency table is a widely used statistical test in many areas of application. For complex survey samples, use of the standard Pearson chi-squared test is inappropriate due to correlation among units within the same cluster. Rao and Scott (1981, Journal of the American Statistical Association 76, 221-230) proposed an approach in which the standard Pearson chi-squared statistic is multiplied by a design effect to adjust for the complex survey design. Unfortunately, this test fails to exist when one of the observed cell counts equals zero. Even with the large samples typical of many complex surveys, zero cell counts can occur for rare events, small domains, or contingency tables with a large number of cells. Here, we propose Wald and score test statistics for independence based on weighted least squares estimating equations. In contrast to the Rao-Scott test statistic, the proposed Wald and score test statistics always exist. In simulations, the score test is found to perform best with respect to type I error. The proposed method is motivated by, and applied to, post surgical complications data from the United States' Nationwide Inpatient Sample (NIS) complex survey of hospitals in 2008. © 2015, The International Biometric Society.
Modified Distribution-Free Goodness-of-Fit Test Statistic.
Chun, So Yeon; Browne, Michael W; Shapiro, Alexander
2018-03-01
Covariance structure analysis and its structural equation modeling extensions have become one of the most widely used methodologies in social sciences such as psychology, education, and economics. An important issue in such analysis is to assess the goodness of fit of a model under analysis. One of the most popular test statistics used in covariance structure analysis is the asymptotically distribution-free (ADF) test statistic introduced by Browne (Br J Math Stat Psychol 37:62-83, 1984). The ADF statistic can be used to test models without any specific distribution assumption (e.g., multivariate normal distribution) of the observed data. Despite its advantage, it has been shown in various empirical studies that unless sample sizes are extremely large, this ADF statistic could perform very poorly in practice. In this paper, we provide a theoretical explanation for this phenomenon and further propose a modified test statistic that improves the performance in samples of realistic size. The proposed statistic deals with the possible ill-conditioning of the involved large-scale covariance matrices.
Szyda, Joanna; Liu, Zengting; Zatoń-Dobrowolska, Magdalena; Wierzbicki, Heliodor; Rzasa, Anna
2008-01-01
We analysed data from a selective DNA pooling experiment with 130 individuals of the arctic fox (Alopex lagopus), which originated from 2 different types regarding body size. The association between alleles of 6 selected unlinked molecular markers and body size was tested by using univariate and multinomial logistic regression models, applying odds ratio and test statistics from the power divergence family. Due to the small sample size and the resulting sparseness of the data table, in hypothesis testing we could not rely on the asymptotic distributions of the tests. Instead, we tried to account for data sparseness by (i) modifying confidence intervals of odds ratio; (ii) using a normal approximation of the asymptotic distribution of the power divergence tests with different approaches for calculating moments of the statistics; and (iii) assessing P values empirically, based on bootstrap samples. As a result, a significant association was observed for 3 markers. Furthermore, we used simulations to assess the validity of the normal approximation of the asymptotic distribution of the test statistics under the conditions of small and sparse samples.
Analysis of statistical misconception in terms of statistical reasoning
NASA Astrophysics Data System (ADS)
Maryati, I.; Priatna, N.
2018-05-01
Reasoning skill is needed for everyone to face globalization era, because every person have to be able to manage and use information from all over the world which can be obtained easily. Statistical reasoning skill is the ability to collect, group, process, interpret, and draw conclusion of information. Developing this skill can be done through various levels of education. However, the skill is low because many people assume that statistics is just the ability to count and using formulas and so do students. Students still have negative attitude toward course which is related to research. The purpose of this research is analyzing students’ misconception in descriptive statistic course toward the statistical reasoning skill. The observation was done by analyzing the misconception test result and statistical reasoning skill test; observing the students’ misconception effect toward statistical reasoning skill. The sample of this research was 32 students of math education department who had taken descriptive statistic course. The mean value of misconception test was 49,7 and standard deviation was 10,6 whereas the mean value of statistical reasoning skill test was 51,8 and standard deviation was 8,5. If the minimal value is 65 to state the standard achievement of a course competence, students’ mean value is lower than the standard competence. The result of students’ misconception study emphasized on which sub discussion that should be considered. Based on the assessment result, it was found that students’ misconception happen on this: 1) writing mathematical sentence and symbol well, 2) understanding basic definitions, 3) determining concept that will be used in solving problem. In statistical reasoning skill, the assessment was done to measure reasoning from: 1) data, 2) representation, 3) statistic format, 4) probability, 5) sample, and 6) association.
Influence of light curing and sample thickness on microhardness of a composite resin
Aguiar, Flávio HB; Andrade, Kelly RM; Leite Lima, Débora AN; Ambrosano, Gláucia MB; Lovadino, José R
2009-01-01
The aim of this in vitro study was to evaluate the influence of light-curing units and different sample thicknesses on the microhardness of a composite resin. Composite resin specimens were randomly prepared and assigned to nine experimental groups (n = 5): considering three light-curing units (conventional quartz tungsten halogen [QTH]: 550 mW/cm2 – 20 s; high irradiance QTH: 1160 mW/cm2 – 10 s; and light-emitting diode [LED]: 360 mW/cm2 – 40 s) and three sample thicknesses (0.5 mm, 1 mm, and 2 mm). All samples were polymerized with the light tip 8 mm away from the specimen. Knoop microhardness was then measured on the top and bottom surfaces of each sample. The top surfaces, with some exceptions, were almost similar; however, in relation to the bottom surfaces, statistical differences were found between curing units and thicknesses. In all experimental groups, the 0.5-mm-thick increments showed microhardness values statistically higher than those observed for 1- and -2-mm increments. The conventional and LED units showed higher hardness mean values and were statistically different from the high irradiance unit. In all experimental groups, microhardness mean values obtained for the top surface were higher than those observed for the bottom surface. In conclusion, higher levels of irradiance or thinner increments would help improve hybrid composite resin polymerization. PMID:23674901
Liang, Si; Xu, Feng; Tang, Weibiao; Zhang, Zheng; Zhang, Wei; Liu, Lili; Wang, Junxia; Lin, Kuangfei
2016-08-01
Hair samples and paired serum samples were collected from e-waste and urban areas in Wenling of Zhejiang Province, China. The PBDE and DBDPE concentrations in hair and serum samples from e-waste workers were significantly higher than those of non-occupational residents and urban residents. BDE209 was the dominating BFRs in hair and serum samples from the e-waste area, while DBDPE was the major BFRs from the urban area. Statistically significant correlations were observed between hair level and serum level for some substances (BDE209, DBDPE, BDE99, BDE47, BDE28, and BDE17), although the PBDE congener profiles in hair were different from those in the serum. A statistically significant positive correlation between the PBDE concentrations and the working age, as well as gender difference, was observed in e-waste workers. Different sources of PBDEs and DBDPE in three groups were identified by principal component analysis and spearman correlation coefficient. Hair is suggested to be a useful matrix for biomonitoring the PBDE exposure in humans.
Thomas, Elaine
2005-01-01
This article is the second in a series of three that will give health care professionals (HCPs) a sound introduction to medical statistics (Thomas, 2004). The objective of research is to find out about the population at large. However, it is generally not possible to study the whole of the population and research questions are addressed in an appropriate study sample. The next crucial step is then to use the information from the sample of individuals to make statements about the wider population of like individuals. This procedure of drawing conclusions about the population, based on study data, is known as inferential statistics. The findings from the study give us the best estimate of what is true for the relevant population, given the sample is representative of the population. It is important to consider how accurate this best estimate is, based on a single sample, when compared to the unknown population figure. Any difference between the observed sample result and the population characteristic is termed the sampling error. This article will cover the two main forms of statistical inference (hypothesis tests and estimation) along with issues that need to be addressed when considering the implications of the study results. Copyright (c) 2005 Whurr Publishers Ltd.
NASA Technical Reports Server (NTRS)
Hughes, William O.; McNelis, Anne M.
2010-01-01
The Earth Observing System (EOS) Terra spacecraft was launched on an Atlas IIAS launch vehicle on its mission to observe planet Earth in late 1999. Prior to launch, the new design of the spacecraft's pyroshock separation system was characterized by a series of 13 separation ground tests. The analysis methods used to evaluate this unusually large amount of shock data will be discussed in this paper, with particular emphasis on population distributions and finding statistically significant families of data, leading to an overall shock separation interface level. The wealth of ground test data also allowed a derivation of a Mission Assurance level for the flight. All of the flight shock measurements were below the EOS Terra Mission Assurance level thus contributing to the overall success of the EOS Terra mission. The effectiveness of the statistical methodology for characterizing the shock interface level and for developing a flight Mission Assurance level from a large sample size of shock data is demonstrated in this paper.
Moorman, J. Randall; Delos, John B.; Flower, Abigail A.; Cao, Hanqing; Kovatchev, Boris P.; Richman, Joshua S.; Lake, Douglas E.
2014-01-01
We have applied principles of statistical signal processing and non-linear dynamics to analyze heart rate time series from premature newborn infants in order to assist in the early diagnosis of sepsis, a common and potentially deadly bacterial infection of the bloodstream. We began with the observation of reduced variability and transient decelerations in heart rate interval time series for hours up to days prior to clinical signs of illness. We find that measurements of standard deviation, sample asymmetry and sample entropy are highly related to imminent clinical illness. We developed multivariable statistical predictive models, and an interface to display the real-time results to clinicians. Using this approach, we have observed numerous cases in which incipient neonatal sepsis was diagnosed and treated without any clinical illness at all. This review focuses on the mathematical and statistical time series approaches used to detect these abnormal heart rate characteristics and present predictive monitoring information to the clinician. PMID:22026974
Trutschel, Diana; Palm, Rebecca; Holle, Bernhard; Simon, Michael
2017-11-01
Because not every scientific question on effectiveness can be answered with randomised controlled trials, research methods that minimise bias in observational studies are required. Two major concerns influence the internal validity of effect estimates: selection bias and clustering. Hence, to reduce the bias of the effect estimates, more sophisticated statistical methods are needed. To introduce statistical approaches such as propensity score matching and mixed models into representative real-world analysis and to conduct the implementation in statistical software R to reproduce the results. Additionally, the implementation in R is presented to allow the results to be reproduced. We perform a two-level analytic strategy to address the problems of bias and clustering: (i) generalised models with different abilities to adjust for dependencies are used to analyse binary data and (ii) the genetic matching and covariate adjustment methods are used to adjust for selection bias. Hence, we analyse the data from two population samples, the sample produced by the matching method and the full sample. The different analysis methods in this article present different results but still point in the same direction. In our example, the estimate of the probability of receiving a case conference is higher in the treatment group than in the control group. Both strategies, genetic matching and covariate adjustment, have their limitations but complement each other to provide the whole picture. The statistical approaches were feasible for reducing bias but were nevertheless limited by the sample used. For each study and obtained sample, the pros and cons of the different methods have to be weighted. Copyright © 2017 The Author(s). Published by Elsevier Ltd.. All rights reserved.
Bolin, Jocelyn Holden; Finch, W Holmes
2014-01-01
Statistical classification of phenomena into observed groups is very common in the social and behavioral sciences. Statistical classification methods, however, are affected by the characteristics of the data under study. Statistical classification can be further complicated by initial misclassification of the observed groups. The purpose of this study is to investigate the impact of initial training data misclassification on several statistical classification and data mining techniques. Misclassification conditions in the three group case will be simulated and results will be presented in terms of overall as well as subgroup classification accuracy. Results show decreased classification accuracy as sample size, group separation and group size ratio decrease and as misclassification percentage increases with random forests demonstrating the highest accuracy across conditions.
Kappa statistic to measure agreement beyond chance in free-response assessments.
Carpentier, Marc; Combescure, Christophe; Merlini, Laura; Perneger, Thomas V
2017-04-19
The usual kappa statistic requires that all observations be enumerated. However, in free-response assessments, only positive (or abnormal) findings are notified, but negative (or normal) findings are not. This situation occurs frequently in imaging or other diagnostic studies. We propose here a kappa statistic that is suitable for free-response assessments. We derived the equivalent of Cohen's kappa statistic for two raters under the assumption that the number of possible findings for any given patient is very large, as well as a formula for sampling variance that is applicable to independent observations (for clustered observations, a bootstrap procedure is proposed). The proposed statistic was applied to a real-life dataset, and compared with the common practice of collapsing observations within a finite number of regions of interest. The free-response kappa is computed from the total numbers of discordant (b and c) and concordant positive (d) observations made in all patients, as 2d/(b + c + 2d). In 84 full-body magnetic resonance imaging procedures in children that were evaluated by 2 independent raters, the free-response kappa statistic was 0.820. Aggregation of results within regions of interest resulted in overestimation of agreement beyond chance. The free-response kappa provides an estimate of agreement beyond chance in situations where only positive findings are reported by raters.
Tarescavage, Anthony M; Alosco, Michael L; Ben-Porath, Yossef S; Wood, Arcangela; Luna-Jones, Lynn
2015-04-01
We investigated the internal structure comparability of Minnesota Multiphasic Personality Inventory-2-Restructured Form (MMPI-2-RF) scores derived from the MMPI-2 and MMPI-2-RF booklets in a sample of 320 criminal defendants (229 males and 54 females). After exclusion of invalid protocols, the final sample consisted of 96 defendants who were administered the MMPI-2-RF booklet and 83 who completed the MMPI-2. No statistically significant differences in MMPI-2-RF invalidity rates were observed between the two forms. Individuals in the final sample who completed the MMPI-2-RF did not statistically differ on demographics or referral question from those who were administered the MMPI-2 booklet. Independent t tests showed no statistically significant differences between MMPI-2-RF scores generated with the MMPI-2 and MMPI-2-RF booklets on the test's substantive scales. Statistically significant small differences were observed on the revised Variable Response Inconsistency (VRIN-r) and True Response Inconsistency (TRIN-r) scales. Cronbach's alpha and standard errors of measurement were approximately equal between the booklets for all MMPI-2-RF scales. Finally, MMPI-2-RF intercorrelations produced from the two forms yielded mostly small and a few medium differences, indicating that discriminant validity and test structure are maintained. Overall, our findings reflect the internal structure comparability of MMPI-2-RF scale scores generated from MMPI-2 and MMPI-2-RF booklets. Implications of these results and limitations of these findings are discussed. © The Author(s) 2014.
Latent spatial models and sampling design for landscape genetics
Hanks, Ephraim M.; Hooten, Mevin B.; Knick, Steven T.; Oyler-McCance, Sara J.; Fike, Jennifer A.; Cross, Todd B.; Schwartz, Michael K.
2016-01-01
We propose a spatially-explicit approach for modeling genetic variation across space and illustrate how this approach can be used to optimize spatial prediction and sampling design for landscape genetic data. We propose a multinomial data model for categorical microsatellite allele data commonly used in landscape genetic studies, and introduce a latent spatial random effect to allow for spatial correlation between genetic observations. We illustrate how modern dimension reduction approaches to spatial statistics can allow for efficient computation in landscape genetic statistical models covering large spatial domains. We apply our approach to propose a retrospective spatial sampling design for greater sage-grouse (Centrocercus urophasianus) population genetics in the western United States.
NASA Technical Reports Server (NTRS)
Bell, Thomas L.; Kundu, Prasun K.; Einaudi, Franco (Technical Monitor)
2000-01-01
Estimates from TRMM satellite data of monthly total rainfall over an area are subject to substantial sampling errors due to the limited number of visits to the area by the satellite during the month. Quantitative comparisons of TRMM averages with data collected by other satellites and by ground-based systems require some estimate of the size of this sampling error. A method of estimating this sampling error based on the actual statistics of the TRMM observations and on some modeling work has been developed. "Sampling error" in TRMM monthly averages is defined here relative to the monthly total a hypothetical satellite permanently stationed above the area would have reported. "Sampling error" therefore includes contributions from the random and systematic errors introduced by the satellite remote sensing system. As part of our long-term goal of providing error estimates for each grid point accessible to the TRMM instruments, sampling error estimates for TRMM based on rain retrievals from TRMM microwave (TMI) data are compared for different times of the year and different oceanic areas (to minimize changes in the statistics due to algorithmic differences over land and ocean). Changes in sampling error estimates due to changes in rain statistics due 1) to evolution of the official algorithms used to process the data, and 2) differences from other remote sensing systems such as the Defense Meteorological Satellite Program (DMSP) Special Sensor Microwave/Imager (SSM/I), are analyzed.
The Detection and Statistics of Giant Arcs behind CLASH Clusters
NASA Astrophysics Data System (ADS)
Xu, Bingxiao; Postman, Marc; Meneghetti, Massimo; Seitz, Stella; Zitrin, Adi; Merten, Julian; Maoz, Dani; Frye, Brenda; Umetsu, Keiichi; Zheng, Wei; Bradley, Larry; Vega, Jesus; Koekemoer, Anton
2016-02-01
We developed an algorithm to find and characterize gravitationally lensed galaxies (arcs) to perform a comparison of the observed and simulated arc abundance. Observations are from the Cluster Lensing And Supernova survey with Hubble (CLASH). Simulated CLASH images are created using the MOKA package and also clusters selected from the high-resolution, hydrodynamical simulations, MUSIC, over the same mass and redshift range as the CLASH sample. The algorithm's arc elongation accuracy, completeness, and false positive rate are determined and used to compute an estimate of the true arc abundance. We derive a lensing efficiency of 4 ± 1 arcs (with length ≥6″ and length-to-width ratio ≥7) per cluster for the X-ray-selected CLASH sample, 4 ± 1 arcs per cluster for the MOKA-simulated sample, and 3 ± 1 arcs per cluster for the MUSIC-simulated sample. The observed and simulated arc statistics are in full agreement. We measure the photometric redshifts of all detected arcs and find a median redshift zs = 1.9 with 33% of the detected arcs having zs > 3. We find that the arc abundance does not depend strongly on the source redshift distribution but is sensitive to the mass distribution of the dark matter halos (e.g., the c-M relation). Our results show that consistency between the observed and simulated distributions of lensed arc sizes and axial ratios can be achieved by using cluster-lensing simulations that are carefully matched to the selection criteria used in the observations.
Implications of Satellite Swath Width on Global Aerosol Optical Thickness Statistics
NASA Technical Reports Server (NTRS)
Colarco, Peter; Kahn, Ralph; Remer, Lorraine; Levy, Robert; Welton, Ellsworth
2012-01-01
We assess the impact of swath width on the statistics of aerosol optical thickness (AOT) retrieved by satellite as inferred from observations made by the Moderate Resolution Imaging Spectroradiometer (MODIS). We sub-sample the year 2009 MODIS data from both the Terra and Aqua spacecraft along several candidate swaths of various widths. We find that due to spatial sampling there is an uncertainty of approximately 0.01 in the global, annual mean AOT. The sub-sampled monthly mean gridded AOT are within +/- 0.01 of the full swath AOT about 20% of the time for the narrow swath sub-samples, about 30% of the time for the moderate width sub-samples, and about 45% of the time for the widest swath considered. These results suggest that future aerosol satellite missions with only a narrow swath view may not sample the true AOT distribution sufficiently to reduce significantly the uncertainty in aerosol direct forcing of climate.
Analysis models for the estimation of oceanic fields
NASA Technical Reports Server (NTRS)
Carter, E. F.; Robinson, A. R.
1987-01-01
A general model for statistically optimal estimates is presented for dealing with scalar, vector and multivariate datasets. The method deals with anisotropic fields and treats space and time dependence equivalently. Problems addressed include the analysis, or the production of synoptic time series of regularly gridded fields from irregular and gappy datasets, and the estimate of fields by compositing observations from several different instruments and sampling schemes. Technical issues are discussed, including the convergence of statistical estimates, the choice of representation of the correlations, the influential domain of an observation, and the efficiency of numerical computations.
Conservative Tests under Satisficing Models of Publication Bias.
McCrary, Justin; Christensen, Garret; Fanelli, Daniele
2016-01-01
Publication bias leads consumers of research to observe a selected sample of statistical estimates calculated by producers of research. We calculate critical values for statistical significance that could help to adjust after the fact for the distortions created by this selection effect, assuming that the only source of publication bias is file drawer bias. These adjusted critical values are easy to calculate and differ from unadjusted critical values by approximately 50%-rather than rejecting a null hypothesis when the t-ratio exceeds 2, the analysis suggests rejecting a null hypothesis when the t-ratio exceeds 3. Samples of published social science research indicate that on average, across research fields, approximately 30% of published t-statistics fall between the standard and adjusted cutoffs.
Conservative Tests under Satisficing Models of Publication Bias
McCrary, Justin; Christensen, Garret; Fanelli, Daniele
2016-01-01
Publication bias leads consumers of research to observe a selected sample of statistical estimates calculated by producers of research. We calculate critical values for statistical significance that could help to adjust after the fact for the distortions created by this selection effect, assuming that the only source of publication bias is file drawer bias. These adjusted critical values are easy to calculate and differ from unadjusted critical values by approximately 50%—rather than rejecting a null hypothesis when the t-ratio exceeds 2, the analysis suggests rejecting a null hypothesis when the t-ratio exceeds 3. Samples of published social science research indicate that on average, across research fields, approximately 30% of published t-statistics fall between the standard and adjusted cutoffs. PMID:26901834
Guo, Ying; Little, Roderick J; McConnell, Daniel S
2012-01-01
Covariate measurement error is common in epidemiologic studies. Current methods for correcting measurement error with information from external calibration samples are insufficient to provide valid adjusted inferences. We consider the problem of estimating the regression of an outcome Y on covariates X and Z, where Y and Z are observed, X is unobserved, but a variable W that measures X with error is observed. Information about measurement error is provided in an external calibration sample where data on X and W (but not Y and Z) are recorded. We describe a method that uses summary statistics from the calibration sample to create multiple imputations of the missing values of X in the regression sample, so that the regression coefficients of Y on X and Z and associated standard errors can be estimated using simple multiple imputation combining rules, yielding valid statistical inferences under the assumption of a multivariate normal distribution. The proposed method is shown by simulation to provide better inferences than existing methods, namely the naive method, classical calibration, and regression calibration, particularly for correction for bias and achieving nominal confidence levels. We also illustrate our method with an example using linear regression to examine the relation between serum reproductive hormone concentrations and bone mineral density loss in midlife women in the Michigan Bone Health and Metabolism Study. Existing methods fail to adjust appropriately for bias due to measurement error in the regression setting, particularly when measurement error is substantial. The proposed method corrects this deficiency.
ADAPTIVE MATCHING IN RANDOMIZED TRIALS AND OBSERVATIONAL STUDIES
van der Laan, Mark J.; Balzer, Laura B.; Petersen, Maya L.
2014-01-01
SUMMARY In many randomized and observational studies the allocation of treatment among a sample of n independent and identically distributed units is a function of the covariates of all sampled units. As a result, the treatment labels among the units are possibly dependent, complicating estimation and posing challenges for statistical inference. For example, cluster randomized trials frequently sample communities from some target population, construct matched pairs of communities from those included in the sample based on some metric of similarity in baseline community characteristics, and then randomly allocate a treatment and a control intervention within each matched pair. In this case, the observed data can neither be represented as the realization of n independent random variables, nor, contrary to current practice, as the realization of n/2 independent random variables (treating the matched pair as the independent sampling unit). In this paper we study estimation of the average causal effect of a treatment under experimental designs in which treatment allocation potentially depends on the pre-intervention covariates of all units included in the sample. We define efficient targeted minimum loss based estimators for this general design, present a theorem that establishes the desired asymptotic normality of these estimators and allows for asymptotically valid statistical inference, and discuss implementation of these estimators. We further investigate the relative asymptotic efficiency of this design compared with a design in which unit-specific treatment assignment depends only on the units’ covariates. Our findings have practical implications for the optimal design and analysis of pair matched cluster randomized trials, as well as for observational studies in which treatment decisions may depend on characteristics of the entire sample. PMID:25097298
The Statistics of Radio Astronomical Polarimetry: Disjoint, Superposed, and Composite Samples
DOE Office of Scientific and Technical Information (OSTI.GOV)
Straten, W. van; Tiburzi, C., E-mail: willem.van.straten@aut.ac.nz
2017-02-01
A statistical framework is presented for the study of the orthogonally polarized modes of radio pulsar emission via the covariances between the Stokes parameters. To accommodate the typically heavy-tailed distributions of single-pulse radio flux density, the fourth-order joint cumulants of the electric field are used to describe the superposition of modes with arbitrary probability distributions. The framework is used to consider the distinction between superposed and disjoint modes, with particular attention to the effects of integration over finite samples. If the interval over which the polarization state is estimated is longer than the timescale for switching between two or moremore » disjoint modes of emission, then the modes are unresolved by the instrument. The resulting composite sample mean exhibits properties that have been attributed to mode superposition, such as depolarization. Because the distinction between disjoint modes and a composite sample of unresolved disjoint modes depends on the temporal resolution of the observing instrumentation, the arguments in favor of superposed modes of pulsar emission are revisited, and observational evidence for disjoint modes is described. In principle, the four-dimensional covariance matrix that describes the distribution of sample mean Stokes parameters can be used to distinguish between disjoint modes, superposed modes, and a composite sample of unresolved disjoint modes. More comprehensive and conclusive interpretation of the covariance matrix requires more detailed consideration of various relevant phenomena, including temporally correlated subpulse modulation (e.g., jitter), statistical dependence between modes (e.g., covariant intensities and partial coherence), and multipath propagation effects (e.g., scintillation and scattering).« less
A General Linear Method for Equating with Small Samples
ERIC Educational Resources Information Center
Albano, Anthony D.
2015-01-01
Research on equating with small samples has shown that methods with stronger assumptions and fewer statistical estimates can lead to decreased error in the estimated equating function. This article introduces a new approach to linear observed-score equating, one which provides flexible control over how form difficulty is assumed versus estimated…
Zhu, Hong; Xu, Xiaohan; Ahn, Chul
2017-01-01
Paired experimental design is widely used in clinical and health behavioral studies, where each study unit contributes a pair of observations. Investigators often encounter incomplete observations of paired outcomes in the data collected. Some study units contribute complete pairs of observations, while the others contribute either pre- or post-intervention observations. Statistical inference for paired experimental design with incomplete observations of continuous outcomes has been extensively studied in literature. However, sample size method for such study design is sparsely available. We derive a closed-form sample size formula based on the generalized estimating equation approach by treating the incomplete observations as missing data in a linear model. The proposed method properly accounts for the impact of mixed structure of observed data: a combination of paired and unpaired outcomes. The sample size formula is flexible to accommodate different missing patterns, magnitude of missingness, and correlation parameter values. We demonstrate that under complete observations, the proposed generalized estimating equation sample size estimate is the same as that based on the paired t-test. In the presence of missing data, the proposed method would lead to a more accurate sample size estimate comparing with the crude adjustment. Simulation studies are conducted to evaluate the finite-sample performance of the generalized estimating equation sample size formula. A real application example is presented for illustration.
CALIFA: a diameter-selected sample for an integral field spectroscopy galaxy survey
NASA Astrophysics Data System (ADS)
Walcher, C. J.; Wisotzki, L.; Bekeraité, S.; Husemann, B.; Iglesias-Páramo, J.; Backsmann, N.; Barrera Ballesteros, J.; Catalán-Torrecilla, C.; Cortijo, C.; del Olmo, A.; Garcia Lorenzo, B.; Falcón-Barroso, J.; Jilkova, L.; Kalinova, V.; Mast, D.; Marino, R. A.; Méndez-Abreu, J.; Pasquali, A.; Sánchez, S. F.; Trager, S.; Zibetti, S.; Aguerri, J. A. L.; Alves, J.; Bland-Hawthorn, J.; Boselli, A.; Castillo Morales, A.; Cid Fernandes, R.; Flores, H.; Galbany, L.; Gallazzi, A.; García-Benito, R.; Gil de Paz, A.; González-Delgado, R. M.; Jahnke, K.; Jungwiert, B.; Kehrig, C.; Lyubenova, M.; Márquez Perez, I.; Masegosa, J.; Monreal Ibero, A.; Pérez, E.; Quirrenbach, A.; Rosales-Ortega, F. F.; Roth, M. M.; Sanchez-Blazquez, P.; Spekkens, K.; Tundo, E.; van de Ven, G.; Verheijen, M. A. W.; Vilchez, J. V.; Ziegler, B.
2014-09-01
We describe and discuss the selection procedure and statistical properties of the galaxy sample used by the Calar Alto Legacy Integral Field Area (CALIFA) survey, a public legacy survey of 600 galaxies using integral field spectroscopy. The CALIFA "mother sample" was selected from the Sloan Digital Sky Survey (SDSS) DR7 photometric catalogue to include all galaxies with an r-band isophotal major axis between 45'' and 79.2'' and with a redshift 0.005 < z < 0.03. The mother sample contains 939 objects, 600 of which will be observed in the course of the CALIFA survey. The selection of targets for observations is based solely on visibility and thus keeps the statistical properties of the mother sample. By comparison with a large set of SDSS galaxies, we find that the CALIFA sample is representative of galaxies over a luminosity range of -19 > Mr > -23.1 and over a stellar mass range between 109.7 and 1011.4 M⊙. In particular, within these ranges, the diameter selection does not lead to any significant bias against - or in favour of - intrinsically large or small galaxies. Only below luminosities of Mr = -19 (or stellar masses <109.7 M⊙) is there a prevalence of galaxies with larger isophotal sizes, especially of nearly edge-on late-type galaxies, but such galaxies form <10% of the full sample. We estimate volume-corrected distribution functions in luminosities and sizes and show that these are statistically fully compatible with estimates from the full SDSS when accounting for large-scale structure. For full characterization of the sample, we also present a number of value-added quantities determined for the galaxies in the CALIFA sample. These include consistent multi-band photometry based on growth curve analyses; stellar masses; distances and quantities derived from these; morphological classifications; and an overview of available multi-wavelength photometric measurements. We also explore different ways of characterizing the environments of CALIFA galaxies, finding that the sample covers environmental conditions from the field to genuine clusters. We finally consider the expected incidence of active galactic nuclei among CALIFA galaxies given the existing pre-CALIFA data, finding that the final observed CALIFA sample will contain approximately 30 Sey2 galaxies. Based on observations collected at the Centro Astronómico Hispano Alemán (CAHA) at Calar Alto, operated jointly by the Max Planck Institute for Astronomy and the Instituto de Astrofísica de Andalucía (CSIC). Publically released data products from CALIFA are made available on the webpage http://www.caha.es/CALIFA
On-line estimation of error covariance parameters for atmospheric data assimilation
NASA Technical Reports Server (NTRS)
Dee, Dick P.
1995-01-01
A simple scheme is presented for on-line estimation of covariance parameters in statistical data assimilation systems. The scheme is based on a maximum-likelihood approach in which estimates are produced on the basis of a single batch of simultaneous observations. Simple-sample covariance estimation is reasonable as long as the number of available observations exceeds the number of tunable parameters by two or three orders of magnitude. Not much is known at present about model error associated with actual forecast systems. Our scheme can be used to estimate some important statistical model error parameters such as regionally averaged variances or characteristic correlation length scales. The advantage of the single-sample approach is that it does not rely on any assumptions about the temporal behavior of the covariance parameters: time-dependent parameter estimates can be continuously adjusted on the basis of current observations. This is of practical importance since it is likely to be the case that both model error and observation error strongly depend on the actual state of the atmosphere. The single-sample estimation scheme can be incorporated into any four-dimensional statistical data assimilation system that involves explicit calculation of forecast error covariances, including optimal interpolation (OI) and the simplified Kalman filter (SKF). The computational cost of the scheme is high but not prohibitive; on-line estimation of one or two covariance parameters in each analysis box of an operational bozed-OI system is currently feasible. A number of numerical experiments performed with an adaptive SKF and an adaptive version of OI, using a linear two-dimensional shallow-water model and artificially generated model error are described. The performance of the nonadaptive versions of these methods turns out to depend rather strongly on correct specification of model error parameters. These parameters are estimated under a variety of conditions, including uniformly distributed model error and time-dependent model error statistics.
Exact intervals and tests for median when one sample value possibly an outliner
NASA Technical Reports Server (NTRS)
Keller, G. J.; Walsh, J. E.
1973-01-01
Available are independent observations (continuous data) that are believed to be a random sample. Desired are distribution-free confidence intervals and significance tests for the population median. However, there is the possibility that either the smallest or the largest observation is an outlier. Then, use of a procedure for rejection of an outlying observation might seem appropriate. Such a procedure would consider that two alternative situations are possible and would select one of them. Either (1) the n observations are truly a random sample, or (2) an outlier exists and its removal leaves a random sample of size n-1. For either situation, confidence intervals and tests are desired for the median of the population yielding the random sample. Unfortunately, satisfactory rejection procedures of a distribution-free nature do not seem to be available. Moreover, all rejection procedures impose undesirable conditional effects on the observations, and also, can select the wrong one of the two above situations. It is found that two-sided intervals and tests based on two symmetrically located order statistics (not the largest and smallest) of the n observations have this property.
Xu, Stanley; Clarke, Christina L; Newcomer, Sophia R; Daley, Matthew F; Glanz, Jason M
2018-05-16
Vaccine safety studies are often electronic health record (EHR)-based observational studies. These studies often face significant methodological challenges, including confounding and misclassification of adverse event. Vaccine safety researchers use self-controlled case series (SCCS) study design to handle confounding effect and employ medical chart review to ascertain cases that are identified using EHR data. However, for common adverse events, limited resources often make it impossible to adjudicate all adverse events observed in electronic data. In this paper, we considered four approaches for analyzing SCCS data with confirmation rates estimated from an internal validation sample: (1) observed cases, (2) confirmed cases only, (3) known confirmation rate, and (4) multiple imputation (MI). We conducted a simulation study to evaluate these four approaches using type I error rates, percent bias, and empirical power. Our simulation results suggest that when misclassification of adverse events is present, approaches such as observed cases, confirmed case only, and known confirmation rate may inflate the type I error, yield biased point estimates, and affect statistical power. The multiple imputation approach considers the uncertainty of estimated confirmation rates from an internal validation sample, yields a proper type I error rate, largely unbiased point estimate, proper variance estimate, and statistical power. © 2018 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Staging Liver Fibrosis with Statistical Observers
NASA Astrophysics Data System (ADS)
Brand, Jonathan Frieman
Chronic liver disease is a worldwide health problem, and hepatic fibrosis (HF) is one of the hallmarks of the disease. Pathology diagnosis of HF is based on textural change in the liver as a lobular collagen network that develops within portal triads. The scale of collagen lobules is characteristically on order of 1mm, which close to the resolution limit of in vivo Gd-enhanced MRI. In this work the methods to collect training and testing images for a Hotelling observer are covered. An observer based on local texture analysis is trained and tested using wet-tissue phantoms. The technique is used to optimize the MRI sequence based on task performance. The final method developed is a two stage model observer to classify fibrotic and healthy tissue in both phantoms and in vivo MRI images. The first stage observer tests for the presence of local texture. Test statistics from the first observer are used to train the second stage observer to globally sample the local observer results. A decision of the disease class is made for an entire MRI image slice using test statistics collected from the second observer. The techniques are tested on wet-tissue phantoms and in vivo clinical patient data.
A new u-statistic with superior design sensitivity in matched observational studies.
Rosenbaum, Paul R
2011-09-01
In an observational or nonrandomized study of treatment effects, a sensitivity analysis indicates the magnitude of bias from unmeasured covariates that would need to be present to alter the conclusions of a naïve analysis that presumes adjustments for observed covariates suffice to remove all bias. The power of sensitivity analysis is the probability that it will reject a false hypothesis about treatment effects allowing for a departure from random assignment of a specified magnitude; in particular, if this specified magnitude is "no departure" then this is the same as the power of a randomization test in a randomized experiment. A new family of u-statistics is proposed that includes Wilcoxon's signed rank statistic but also includes other statistics with substantially higher power when a sensitivity analysis is performed in an observational study. Wilcoxon's statistic has high power to detect small effects in large randomized experiments-that is, it often has good Pitman efficiency-but small effects are invariably sensitive to small unobserved biases. Members of this family of u-statistics that emphasize medium to large effects can have substantially higher power in a sensitivity analysis. For example, in one situation with 250 pair differences that are Normal with expectation 1/2 and variance 1, the power of a sensitivity analysis that uses Wilcoxon's statistic is 0.08 while the power of another member of the family of u-statistics is 0.66. The topic is examined by performing a sensitivity analysis in three observational studies, using an asymptotic measure called the design sensitivity, and by simulating power in finite samples. The three examples are drawn from epidemiology, clinical medicine, and genetic toxicology. © 2010, The International Biometric Society.
Comparing geological and statistical approaches for element selection in sediment tracing research
NASA Astrophysics Data System (ADS)
Laceby, J. Patrick; McMahon, Joe; Evrard, Olivier; Olley, Jon
2015-04-01
Elevated suspended sediment loads reduce reservoir capacity and significantly increase the cost of operating water treatment infrastructure, making the management of sediment supply to reservoirs of increasingly importance. Sediment fingerprinting techniques can be used to determine the relative contributions of different sources of sediment accumulating in reservoirs. The objective of this research is to compare geological and statistical approaches to element selection for sediment fingerprinting modelling. Time-integrated samplers (n=45) were used to obtain source samples from four major subcatchments flowing into the Baroon Pocket Dam in South East Queensland, Australia. The geochemistry of potential sources were compared to the geochemistry of sediment cores (n=12) sampled in the reservoir. The geochemical approach selected elements for modelling that provided expected, observed and statistical discrimination between sediment sources. Two statistical approaches selected elements for modelling with the Kruskal-Wallis H-test and Discriminatory Function Analysis (DFA). In particular, two different significance levels (0.05 & 0.35) for the DFA were included to investigate the importance of element selection on modelling results. A distribution model determined the relative contributions of different sources to sediment sampled in the Baroon Pocket Dam. Elemental discrimination was expected between one subcatchment (Obi Obi Creek) and the remaining subcatchments (Lexys, Falls and Bridge Creek). Six major elements were expected to provide discrimination. Of these six, only Fe2O3 and SiO2 provided expected, observed and statistical discrimination. Modelling results with this geological approach indicated 36% (+/- 9%) of sediment sampled in the reservoir cores were from mafic-derived sources and 64% (+/- 9%) were from felsic-derived sources. The geological and the first statistical approach (DFA0.05) differed by only 1% (σ 5%) for 5 out of 6 model groupings with only the Lexys Creek modelling results differing significantly (35%). The statistical model with expanded elemental selection (DFA0.35) differed from the geological model by an average of 30% for all 6 models. Elemental selection for sediment fingerprinting therefore has the potential to impact modeling results. Accordingly is important to incorporate both robust geological and statistical approaches when selecting elements for sediment fingerprinting. For the Baroon Pocket Dam, management should focus on reducing the supply of sediments derived from felsic sources in each of the subcatchments.
Rosenberg, Matthew B; Dockery, Christopher R
2008-11-01
Laser-induced breakdown spectroscopy (LIBS) has been used to determine the period of time that a shooter will test positive for gunshot residue (GSR) after firing a revolver. Multiple rounds of primer were fired and samples collected at multiple hour intervals using an adhesive tape pressed against the skin. Samples were analyzed directly using a commercially available laser-induced breakdown spectrometer where barium emission (originating from barium nitrate in the primer) was observed. Population statistics were used to compare suspected GSR to a library of blank samples from which a threshold value was established. Statistically significant results, positive for GSR, are obtained 5.27 days after a firearm discharge using these techniques.
ERIC Educational Resources Information Center
Videtich, Patricia E.; Neal, William J.
2012-01-01
Using sieving and sample "unknowns" for instructional grain-size analysis and interpretation of sands in undergraduate sedimentology courses has advantages over other techniques. Students (1) learn to calculate and use statistics; (2) visually observe differences in the grain-size fractions, thereby developing a sense of specific size…
Statistical Analysis of Hubble /WFC3 Transit Spectroscopy of Extrasolar Planets
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fu, Guangwei; Deming, Drake; Knutson, Heather
2017-10-01
Transmission spectroscopy provides a window to study exoplanetary atmospheres, but that window is fogged by clouds and hazes. Clouds and haze introduce a degeneracy between the strength of gaseous absorption features and planetary physical parameters such as abundances. One way to break that degeneracy is via statistical studies. We collect all published HST /WFC3 transit spectra for 1.1–1.65 μ m water vapor absorption and perform a statistical study on potential correlations between the water absorption feature and planetary parameters. We fit the observed spectra with a template calculated for each planet using the Exo-transmit code. We express the magnitude ofmore » the water absorption in scale heights, thereby removing the known dependence on temperature, surface gravity, and mean molecular weight. We find that the absorption in scale heights has a positive baseline correlation with planetary equilibrium temperature; our hypothesis is that decreasing cloud condensation with increasing temperature is responsible for this baseline slope. However, the observed sample is also intrinsically degenerate in the sense that equilibrium temperature correlates with planetary mass. We compile the distribution of absorption in scale heights, and we find that this distribution is closer to log-normal than Gaussian. However, we also find that the distribution of equilibrium temperatures for the observed planets is similarly log-normal. This indicates that the absorption values are affected by observational bias, whereby observers have not yet targeted a sufficient sample of the hottest planets.« less
Statistical Analysis of Hubble/WFC3 Transit Spectroscopy of Extrasolar Planets
NASA Astrophysics Data System (ADS)
Fu, Guangwei; Deming, Drake; Knutson, Heather; Madhusudhan, Nikku; Mandell, Avi; Fraine, Jonathan
2018-01-01
Transmission spectroscopy provides a window to study exoplanetary atmospheres, but that window is fogged by clouds and hazes. Clouds and haze introduce a degeneracy between the strength of gaseous absorption features and planetary physical parameters such as abundances. One way to break that degeneracy is via statistical studies. We collect all published HST/WFC3 transit spectra for 1.1-1.65 micron water vapor absorption, and perform a statistical study on potential correlations between the water absorption feature and planetary parameters. We fit the observed spectra with a template calculated for each planet using the Exo-Transmit code. We express the magnitude of the water absorption in scale heights, thereby removing the known dependence on temperature, surface gravity, and mean molecular weight. We find that the absorption in scale heights has a positive baseline correlation with planetary equilibrium temperature; our hypothesis is that decreasing cloud condensation with increasing temperature is responsible for this baseline slope. However, the observed sample is also intrinsically degenerate in the sense that equilibrium temperature correlates with planetary mass. We compile the distribution of absorption in scale heights, and we find that this distribution is closer to log-normal than Gaussian. However, we also find that the distribution of equilibrium temperatures for the observed planets is similarly log-normal. This indicates that the absorption values are affected by observational bias, whereby observers have not yet targeted a sufficient sample of the hottest planets.
Statistical Analysis of Hubble/WFC3 Transit Spectroscopy of Extrasolar Planets
NASA Astrophysics Data System (ADS)
Fu, Guangwei; Deming, Drake; Knutson, Heather; Madhusudhan, Nikku; Mandell, Avi; Fraine, Jonathan
2017-10-01
Transmission spectroscopy provides a window to study exoplanetary atmospheres, but that window is fogged by clouds and hazes. Clouds and haze introduce a degeneracy between the strength of gaseous absorption features and planetary physical parameters such as abundances. One way to break that degeneracy is via statistical studies. We collect all published HST/WFC3 transit spectra for 1.1-1.65 μm water vapor absorption and perform a statistical study on potential correlations between the water absorption feature and planetary parameters. We fit the observed spectra with a template calculated for each planet using the Exo-transmit code. We express the magnitude of the water absorption in scale heights, thereby removing the known dependence on temperature, surface gravity, and mean molecular weight. We find that the absorption in scale heights has a positive baseline correlation with planetary equilibrium temperature; our hypothesis is that decreasing cloud condensation with increasing temperature is responsible for this baseline slope. However, the observed sample is also intrinsically degenerate in the sense that equilibrium temperature correlates with planetary mass. We compile the distribution of absorption in scale heights, and we find that this distribution is closer to log-normal than Gaussian. However, we also find that the distribution of equilibrium temperatures for the observed planets is similarly log-normal. This indicates that the absorption values are affected by observational bias, whereby observers have not yet targeted a sufficient sample of the hottest planets.
Control of oral malodour by dentifrices measured by gas chromatography.
Newby, Evelyn E; Hickling, Jenneth M; Hughes, Francis J; Proskin, Howard M; Bosma, Marylynn P
2008-04-01
To evaluate the effect of toothpaste treatments on levels of oral volatile sulphur compounds (VSCs) measured by gas chromatography in two clinical studies. These were blinded, randomised, controlled, crossover studies with 16 (study A) or 20 (study B) healthy volunteers between the ages of 19-54. Study A: breath samples were collected at baseline, immediately and lhr after brushing. Four dentifrices (Zinc A, Zinc B, commercially available triclosan dentifrice and zinc free control) were evaluated. Study B: breath samples were collected at baseline, immediately, 1, 2, 3 and 7 hours after treatment. Subjects consumed a light breakfast then provided an additional breath sample between baseline assessment and treatment. Two dentifrices (gel-to-foam and a commercially available triclosan dentrifrice) were evaluated. Breath samples were collected in syringes and analysed for VSCs (hydrogen sulphide, methyl mercaptan and Total VSCs) utilising gas chromatography (GC) with flame photometric detection. Study A: immediately after treatment, a statistically significant reduction in VSCs from baseline was observed for Zinc A product only. A statistically significant reduction in VSCs from baseline was observed after 1 hour for all products. Both zinc products exhibited a significantly greater reduction from baseline VSCs than Colgate Total and Control at all time points. Study B: a statistically significant reduction in VSCs from baseline was observed at all time points for both products. The gel-to-foam product exhibited significantly greater reduction from baseline Total VSC concentration than Colgate Total at all time points from 1 hour post-treatment. Control of oral malodour by toothpaste treatment, evaluated as VSC levels using GC, has been demonstrated. Zinc is effective at reducing VSCs and the efficacy of zinc is formulation dependent. A gel-to-foam dentifrice was more effective at reducing VSCs than Colgate Total up to 7 hours.
Pointwise probability reinforcements for robust statistical inference.
Frénay, Benoît; Verleysen, Michel
2014-02-01
Statistical inference using machine learning techniques may be difficult with small datasets because of abnormally frequent data (AFDs). AFDs are observations that are much more frequent in the training sample that they should be, with respect to their theoretical probability, and include e.g. outliers. Estimates of parameters tend to be biased towards models which support such data. This paper proposes to introduce pointwise probability reinforcements (PPRs): the probability of each observation is reinforced by a PPR and a regularisation allows controlling the amount of reinforcement which compensates for AFDs. The proposed solution is very generic, since it can be used to robustify any statistical inference method which can be formulated as a likelihood maximisation. Experiments show that PPRs can be easily used to tackle regression, classification and projection: models are freed from the influence of outliers. Moreover, outliers can be filtered manually since an abnormality degree is obtained for each observation. Copyright © 2013 Elsevier Ltd. All rights reserved.
Mycology of chronic suppurative otitis media-cholesteatoma disease: An evaluative study.
Singh, Gautam Bir; Solo, Medozhanuo; Kaur, Ravinder; Arora, Rubeena; Kumar, Sunil
To detect the prevalence of fungus in chronic suppurative otitis media-cholesteatoma disease and to evaluate its clinical significance. Prospective observational study conducted in a sample size of 46 patients at a tertiary care university teaching hospital. Forty six patients suffering from chronic suppurative otitis media-cholesteatoma disease were recruited in this prospective study. Data was duly recorded. Cholesteatoma sample was procured at the time of mastoid surgery and microbiologically analysed for fungal infestation. Clinical correlation to fungus infestation of cholesteatoma was statistically analysed. Out of the recruited 46 patients, post-operatively cholesteatoma was seen in 40 cases only. Seventeen i.e. 42.5% of these cases had fungal colonization of cholesteatoma. Further a statistically significant correlation between persistent otorrhoea and fungal infestation of cholesteatoma was observed. Three cases of fungal otomastoiditis were also recorded in this study, but a statistically significant correlation between complications and fungus infestation of cholesteatoma could not be clearly established. There is fungal colonization of cholesteatoma which is pathogenic and can cause persistent otorrhoea. Copyright © 2017 Elsevier Inc. All rights reserved.
The problem of pseudoreplication in neuroscientific studies: is it affecting your analysis?
2010-01-01
Background Pseudoreplication occurs when observations are not statistically independent, but treated as if they are. This can occur when there are multiple observations on the same subjects, when samples are nested or hierarchically organised, or when measurements are correlated in time or space. Analysis of such data without taking these dependencies into account can lead to meaningless results, and examples can easily be found in the neuroscience literature. Results A single issue of Nature Neuroscience provided a number of examples and is used as a case study to highlight how pseudoreplication arises in neuroscientific studies, why the analyses in these papers are incorrect, and appropriate analytical methods are provided. 12% of papers had pseudoreplication and a further 36% were suspected of having pseudoreplication, but it was not possible to determine for certain because insufficient information was provided. Conclusions Pseudoreplication can undermine the conclusions of a statistical analysis, and it would be easier to detect if the sample size, degrees of freedom, the test statistic, and precise p-values are reported. This information should be a requirement for all publications. PMID:20074371
Hsiao, Chiaowen; Liu, Mengya; Stanton, Rick; McGee, Monnie; Qian, Yu
2015-01-01
Abstract Flow cytometry (FCM) is a fluorescence‐based single‐cell experimental technology that is routinely applied in biomedical research for identifying cellular biomarkers of normal physiological responses and abnormal disease states. While many computational methods have been developed that focus on identifying cell populations in individual FCM samples, very few have addressed how the identified cell populations can be matched across samples for comparative analysis. This article presents FlowMap‐FR, a novel method for cell population mapping across FCM samples. FlowMap‐FR is based on the Friedman–Rafsky nonparametric test statistic (FR statistic), which quantifies the equivalence of multivariate distributions. As applied to FCM data by FlowMap‐FR, the FR statistic objectively quantifies the similarity between cell populations based on the shapes, sizes, and positions of fluorescence data distributions in the multidimensional feature space. To test and evaluate the performance of FlowMap‐FR, we simulated the kinds of biological and technical sample variations that are commonly observed in FCM data. The results show that FlowMap‐FR is able to effectively identify equivalent cell populations between samples under scenarios of proportion differences and modest position shifts. As a statistical test, FlowMap‐FR can be used to determine whether the expression of a cellular marker is statistically different between two cell populations, suggesting candidates for new cellular phenotypes by providing an objective statistical measure. In addition, FlowMap‐FR can indicate situations in which inappropriate splitting or merging of cell populations has occurred during gating procedures. We compared the FR statistic with the symmetric version of Kullback–Leibler divergence measure used in a previous population matching method with both simulated and real data. The FR statistic outperforms the symmetric version of KL‐distance in distinguishing equivalent from nonequivalent cell populations. FlowMap‐FR was also employed as a distance metric to match cell populations delineated by manual gating across 30 FCM samples from a benchmark FlowCAP data set. An F‐measure of 0.88 was obtained, indicating high precision and recall of the FR‐based population matching results. FlowMap‐FR has been implemented as a standalone R/Bioconductor package so that it can be easily incorporated into current FCM data analytical workflows. © 2015 International Society for Advancement of Cytometry PMID:26274018
Hsiao, Chiaowen; Liu, Mengya; Stanton, Rick; McGee, Monnie; Qian, Yu; Scheuermann, Richard H
2016-01-01
Flow cytometry (FCM) is a fluorescence-based single-cell experimental technology that is routinely applied in biomedical research for identifying cellular biomarkers of normal physiological responses and abnormal disease states. While many computational methods have been developed that focus on identifying cell populations in individual FCM samples, very few have addressed how the identified cell populations can be matched across samples for comparative analysis. This article presents FlowMap-FR, a novel method for cell population mapping across FCM samples. FlowMap-FR is based on the Friedman-Rafsky nonparametric test statistic (FR statistic), which quantifies the equivalence of multivariate distributions. As applied to FCM data by FlowMap-FR, the FR statistic objectively quantifies the similarity between cell populations based on the shapes, sizes, and positions of fluorescence data distributions in the multidimensional feature space. To test and evaluate the performance of FlowMap-FR, we simulated the kinds of biological and technical sample variations that are commonly observed in FCM data. The results show that FlowMap-FR is able to effectively identify equivalent cell populations between samples under scenarios of proportion differences and modest position shifts. As a statistical test, FlowMap-FR can be used to determine whether the expression of a cellular marker is statistically different between two cell populations, suggesting candidates for new cellular phenotypes by providing an objective statistical measure. In addition, FlowMap-FR can indicate situations in which inappropriate splitting or merging of cell populations has occurred during gating procedures. We compared the FR statistic with the symmetric version of Kullback-Leibler divergence measure used in a previous population matching method with both simulated and real data. The FR statistic outperforms the symmetric version of KL-distance in distinguishing equivalent from nonequivalent cell populations. FlowMap-FR was also employed as a distance metric to match cell populations delineated by manual gating across 30 FCM samples from a benchmark FlowCAP data set. An F-measure of 0.88 was obtained, indicating high precision and recall of the FR-based population matching results. FlowMap-FR has been implemented as a standalone R/Bioconductor package so that it can be easily incorporated into current FCM data analytical workflows. © The Authors. Published by Wiley Periodicals, Inc. on behalf of ISAC.
THE DETECTION AND STATISTICS OF GIANT ARCS BEHIND CLASH CLUSTERS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Xu, Bingxiao; Zheng, Wei; Postman, Marc
We developed an algorithm to find and characterize gravitationally lensed galaxies (arcs) to perform a comparison of the observed and simulated arc abundance. Observations are from the Cluster Lensing And Supernova survey with Hubble (CLASH). Simulated CLASH images are created using the MOKA package and also clusters selected from the high-resolution, hydrodynamical simulations, MUSIC, over the same mass and redshift range as the CLASH sample. The algorithm's arc elongation accuracy, completeness, and false positive rate are determined and used to compute an estimate of the true arc abundance. We derive a lensing efficiency of 4 ± 1 arcs (with length ≥6″ andmore » length-to-width ratio ≥7) per cluster for the X-ray-selected CLASH sample, 4 ± 1 arcs per cluster for the MOKA-simulated sample, and 3 ± 1 arcs per cluster for the MUSIC-simulated sample. The observed and simulated arc statistics are in full agreement. We measure the photometric redshifts of all detected arcs and find a median redshift z{sub s} = 1.9 with 33% of the detected arcs having z{sub s} > 3. We find that the arc abundance does not depend strongly on the source redshift distribution but is sensitive to the mass distribution of the dark matter halos (e.g., the c–M relation). Our results show that consistency between the observed and simulated distributions of lensed arc sizes and axial ratios can be achieved by using cluster-lensing simulations that are carefully matched to the selection criteria used in the observations.« less
Planetary mass function and planetary systems
NASA Astrophysics Data System (ADS)
Dominik, M.
2011-02-01
With planets orbiting stars, a planetary mass function should not be seen as a low-mass extension of the stellar mass function, but a proper formalism needs to take care of the fact that the statistical properties of planet populations are linked to the properties of their respective host stars. This can be accounted for by describing planet populations by means of a differential planetary mass-radius-orbit function, which together with the fraction of stars with given properties that are orbited by planets and the stellar mass function allows the derivation of all statistics for any considered sample. These fundamental functions provide a framework for comparing statistics that result from different observing techniques and campaigns which all have their very specific selection procedures and detection efficiencies. Moreover, recent results both from gravitational microlensing campaigns and radial-velocity surveys of stars indicate that planets tend to cluster in systems rather than being the lonely child of their respective parent star. While planetary multiplicity in an observed system becomes obvious with the detection of several planets, its quantitative assessment however comes with the challenge to exclude the presence of further planets. Current exoplanet samples begin to give us first hints at the population statistics, whereas pictures of planet parameter space in its full complexity call for samples that are 2-4 orders of magnitude larger. In order to derive meaningful statistics, however, planet detection campaigns need to be designed in such a way that well-defined fully deterministic target selection, monitoring and detection criteria are applied. The probabilistic nature of gravitational microlensing makes this technique an illustrative example of all the encountered challenges and uncertainties.
Analysis of Statistical Methods Currently used in Toxicology Journals
Na, Jihye; Yang, Hyeri
2014-01-01
Statistical methods are frequently used in toxicology, yet it is not clear whether the methods employed by the studies are used consistently and conducted based on sound statistical grounds. The purpose of this paper is to describe statistical methods used in top toxicology journals. More specifically, we sampled 30 papers published in 2014 from Toxicology and Applied Pharmacology, Archives of Toxicology, and Toxicological Science and described methodologies used to provide descriptive and inferential statistics. One hundred thirteen endpoints were observed in those 30 papers, and most studies had sample size less than 10, with the median and the mode being 6 and 3 & 6, respectively. Mean (105/113, 93%) was dominantly used to measure central tendency, and standard error of the mean (64/113, 57%) and standard deviation (39/113, 34%) were used to measure dispersion, while few studies provide justifications regarding why the methods being selected. Inferential statistics were frequently conducted (93/113, 82%), with one-way ANOVA being most popular (52/93, 56%), yet few studies conducted either normality or equal variance test. These results suggest that more consistent and appropriate use of statistical method is necessary which may enhance the role of toxicology in public health. PMID:25343012
Analysis of Statistical Methods Currently used in Toxicology Journals.
Na, Jihye; Yang, Hyeri; Bae, SeungJin; Lim, Kyung-Min
2014-09-01
Statistical methods are frequently used in toxicology, yet it is not clear whether the methods employed by the studies are used consistently and conducted based on sound statistical grounds. The purpose of this paper is to describe statistical methods used in top toxicology journals. More specifically, we sampled 30 papers published in 2014 from Toxicology and Applied Pharmacology, Archives of Toxicology, and Toxicological Science and described methodologies used to provide descriptive and inferential statistics. One hundred thirteen endpoints were observed in those 30 papers, and most studies had sample size less than 10, with the median and the mode being 6 and 3 & 6, respectively. Mean (105/113, 93%) was dominantly used to measure central tendency, and standard error of the mean (64/113, 57%) and standard deviation (39/113, 34%) were used to measure dispersion, while few studies provide justifications regarding why the methods being selected. Inferential statistics were frequently conducted (93/113, 82%), with one-way ANOVA being most popular (52/93, 56%), yet few studies conducted either normality or equal variance test. These results suggest that more consistent and appropriate use of statistical method is necessary which may enhance the role of toxicology in public health.
Denwood, M J; Love, S; Innocent, G T; Matthews, L; McKendrick, I J; Hillary, N; Smith, A; Reid, S W J
2012-08-13
The faecal egg count (FEC) is the most widely used means of quantifying the nematode burden of horses, and is frequently used in clinical practice to inform treatment and prevention. The statistical process underlying the FEC is complex, comprising a Poisson counting error process for each sample, compounded with an underlying continuous distribution of means between samples. Being able to quantify the sources of variability contributing to this distribution of means is a necessary step towards providing estimates of statistical power for future FEC and FECRT studies, and may help to improve the usefulness of the FEC technique by identifying and minimising unwanted sources of variability. Obtaining such estimates require a hierarchical statistical model coupled with repeated FEC observations from a single animal over a short period of time. Here, we use this approach to provide the first comparative estimate of multiple sources of within-horse FEC variability. The results demonstrate that a substantial proportion of the observed variation in FEC between horses occurs as a result of variation in FEC within an animal, with the major sources being aggregation of eggs within faeces and variation in egg concentration between faecal piles. The McMaster procedure itself is associated with a comparatively small coefficient of variation, and is therefore highly repeatable when a sufficiently large number of eggs are observed to reduce the error associated with the counting process. We conclude that the variation between samples taken from the same animal is substantial, but can be reduced through the use of larger homogenised faecal samples. Estimates are provided for the coefficient of variation (cv) associated with each within animal source of variability in observed FEC, allowing the usefulness of individual FEC to be quantified, and providing a basis for future FEC and FECRT studies. Copyright © 2012 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Kittiwisit, Piyanat; Bowman, Judd D.; Jacobs, Daniel C.; Beardsley, Adam P.; Thyagarajan, Nithyanandan
2018-03-01
We present a baseline sensitivity analysis of the Hydrogen Epoch of Reionization Array (HERA) and its build-out stages to one-point statistics (variance, skewness, and kurtosis) of redshifted 21 cm intensity fluctuation from the Epoch of Reionization (EoR) based on realistic mock observations. By developing a full-sky 21 cm light-cone model, taking into account the proper field of view and frequency bandwidth, utilizing a realistic measurement scheme, and assuming perfect foreground removal, we show that HERA will be able to recover statistics of the sky model with high sensitivity by averaging over measurements from multiple fields. All build-out stages will be able to detect variance, while skewness and kurtosis should be detectable for HERA128 and larger. We identify sample variance as the limiting constraint of the measurements at the end of reionization. The sensitivity can also be further improved by performing frequency windowing. In addition, we find that strong sample variance fluctuation in the kurtosis measured from an individual field of observation indicates the presence of outlying cold or hot regions in the underlying fluctuations, a feature that can potentially be used as an EoR bubble indicator.
Bayesian Orbit Computation Tools for Objects on Geocentric Orbits
NASA Astrophysics Data System (ADS)
Virtanen, J.; Granvik, M.; Muinonen, K.; Oszkiewicz, D.
2013-08-01
We consider the space-debris orbital inversion problem via the concept of Bayesian inference. The methodology has been put forward for the orbital analysis of solar system small bodies in early 1990's [7] and results in a full solution of the statistical inverse problem given in terms of a posteriori probability density function (PDF) for the orbital parameters. We demonstrate the applicability of our statistical orbital analysis software to Earth orbiting objects, both using well-established Monte Carlo (MC) techniques (for a review, see e.g. [13] as well as recently developed Markov-chain MC (MCMC) techniques (e.g., [9]). In particular, we exploit the novel virtual observation MCMC method [8], which is based on the characterization of the phase-space volume of orbital solutions before the actual MCMC sampling. Our statistical methods and the resulting PDFs immediately enable probabilistic impact predictions to be carried out. Furthermore, this can be readily done also for very sparse data sets and data sets of poor quality - providing that some a priori information on the observational uncertainty is available. For asteroids, impact probabilities with the Earth from the discovery night onwards have been provided, e.g., by [11] and [10], the latter study includes the sampling of the observational-error standard deviation as a random variable.
Vexler, Albert; Tanajian, Hovig; Hutson, Alan D
In practice, parametric likelihood-ratio techniques are powerful statistical tools. In this article, we propose and examine novel and simple distribution-free test statistics that efficiently approximate parametric likelihood ratios to analyze and compare distributions of K groups of observations. Using the density-based empirical likelihood methodology, we develop a Stata package that applies to a test for symmetry of data distributions and compares K -sample distributions. Recognizing that recent statistical software packages do not sufficiently address K -sample nonparametric comparisons of data distributions, we propose a new Stata command, vxdbel, to execute exact density-based empirical likelihood-ratio tests using K samples. To calculate p -values of the proposed tests, we use the following methods: 1) a classical technique based on Monte Carlo p -value evaluations; 2) an interpolation technique based on tabulated critical values; and 3) a new hybrid technique that combines methods 1 and 2. The third, cutting-edge method is shown to be very efficient in the context of exact-test p -value computations. This Bayesian-type method considers tabulated critical values as prior information and Monte Carlo generations of test statistic values as data used to depict the likelihood function. In this case, a nonparametric Bayesian method is proposed to compute critical values of exact tests.
NASA Astrophysics Data System (ADS)
Price-Whelan, Adrian M.; Agueros, M. A.; Fournier, A.; Street, R.; Ofek, E.; Levitan, D. B.; PTF Collaboration
2013-01-01
Many current photometric, time-domain surveys are driven by specific goals such as searches for supernovae or transiting exoplanets, or studies of stellar variability. These goals in turn set the cadence with which individual fields are re-imaged. In the case of the Palomar Transient Factory (PTF), several such sub-surveys are being conducted in parallel, leading to extremely non-uniform sampling over the survey's nearly 20,000 sq. deg. footprint. While the typical 7.26 sq. deg. PTF field has been imaged 20 times in R-band, ~2300 sq. deg. have been observed more than 100 times. We use the existing PTF data 6.4x107 light curves) to study the trade-off that occurs when searching for microlensing events when one has access to a large survey footprint with irregular sampling. To examine the probability that microlensing events can be recovered in these data, we also test previous statistics used on uniformly sampled data to identify variables and transients. We find that one such statistic, the von Neumann ratio, performs best for identifying simulated microlensing events. We develop a selection method using this statistic and apply it to data from all PTF fields with >100 observations to uncover a number of interesting candidate events. This work can help constrain all-sky event rate predictions and tests microlensing signal recovery in large datasets, both of which will be useful to future wide-field, time-domain surveys such as the LSST.
Scheid, Anika; Nebel, Markus E
2012-07-09
Over the past years, statistical and Bayesian approaches have become increasingly appreciated to address the long-standing problem of computational RNA structure prediction. Recently, a novel probabilistic method for the prediction of RNA secondary structures from a single sequence has been studied which is based on generating statistically representative and reproducible samples of the entire ensemble of feasible structures for a particular input sequence. This method samples the possible foldings from a distribution implied by a sophisticated (traditional or length-dependent) stochastic context-free grammar (SCFG) that mirrors the standard thermodynamic model applied in modern physics-based prediction algorithms. Specifically, that grammar represents an exact probabilistic counterpart to the energy model underlying the Sfold software, which employs a sampling extension of the partition function (PF) approach to produce statistically representative subsets of the Boltzmann-weighted ensemble. Although both sampling approaches have the same worst-case time and space complexities, it has been indicated that they differ in performance (both with respect to prediction accuracy and quality of generated samples), where neither of these two competing approaches generally outperforms the other. In this work, we will consider the SCFG based approach in order to perform an analysis on how the quality of generated sample sets and the corresponding prediction accuracy changes when different degrees of disturbances are incorporated into the needed sampling probabilities. This is motivated by the fact that if the results prove to be resistant to large errors on the distinct sampling probabilities (compared to the exact ones), then it will be an indication that these probabilities do not need to be computed exactly, but it may be sufficient and more efficient to approximate them. Thus, it might then be possible to decrease the worst-case time requirements of such an SCFG based sampling method without significant accuracy losses. If, on the other hand, the quality of sampled structures can be observed to strongly react to slight disturbances, there is little hope for improving the complexity by heuristic procedures. We hence provide a reliable test for the hypothesis that a heuristic method could be implemented to improve the time scaling of RNA secondary structure prediction in the worst-case - without sacrificing much of the accuracy of the results. Our experiments indicate that absolute errors generally lead to the generation of useless sample sets, whereas relative errors seem to have only small negative impact on both the predictive accuracy and the overall quality of resulting structure samples. Based on these observations, we present some useful ideas for developing a time-reduced sampling method guaranteeing an acceptable predictive accuracy. We also discuss some inherent drawbacks that arise in the context of approximation. The key results of this paper are crucial for the design of an efficient and competitive heuristic prediction method based on the increasingly accepted and attractive statistical sampling approach. This has indeed been indicated by the construction of prototype algorithms.
2012-01-01
Background Over the past years, statistical and Bayesian approaches have become increasingly appreciated to address the long-standing problem of computational RNA structure prediction. Recently, a novel probabilistic method for the prediction of RNA secondary structures from a single sequence has been studied which is based on generating statistically representative and reproducible samples of the entire ensemble of feasible structures for a particular input sequence. This method samples the possible foldings from a distribution implied by a sophisticated (traditional or length-dependent) stochastic context-free grammar (SCFG) that mirrors the standard thermodynamic model applied in modern physics-based prediction algorithms. Specifically, that grammar represents an exact probabilistic counterpart to the energy model underlying the Sfold software, which employs a sampling extension of the partition function (PF) approach to produce statistically representative subsets of the Boltzmann-weighted ensemble. Although both sampling approaches have the same worst-case time and space complexities, it has been indicated that they differ in performance (both with respect to prediction accuracy and quality of generated samples), where neither of these two competing approaches generally outperforms the other. Results In this work, we will consider the SCFG based approach in order to perform an analysis on how the quality of generated sample sets and the corresponding prediction accuracy changes when different degrees of disturbances are incorporated into the needed sampling probabilities. This is motivated by the fact that if the results prove to be resistant to large errors on the distinct sampling probabilities (compared to the exact ones), then it will be an indication that these probabilities do not need to be computed exactly, but it may be sufficient and more efficient to approximate them. Thus, it might then be possible to decrease the worst-case time requirements of such an SCFG based sampling method without significant accuracy losses. If, on the other hand, the quality of sampled structures can be observed to strongly react to slight disturbances, there is little hope for improving the complexity by heuristic procedures. We hence provide a reliable test for the hypothesis that a heuristic method could be implemented to improve the time scaling of RNA secondary structure prediction in the worst-case – without sacrificing much of the accuracy of the results. Conclusions Our experiments indicate that absolute errors generally lead to the generation of useless sample sets, whereas relative errors seem to have only small negative impact on both the predictive accuracy and the overall quality of resulting structure samples. Based on these observations, we present some useful ideas for developing a time-reduced sampling method guaranteeing an acceptable predictive accuracy. We also discuss some inherent drawbacks that arise in the context of approximation. The key results of this paper are crucial for the design of an efficient and competitive heuristic prediction method based on the increasingly accepted and attractive statistical sampling approach. This has indeed been indicated by the construction of prototype algorithms. PMID:22776037
Al-Shamary, Noora M; Al-Ghouti, Mohammad A; Al-Shaikh, Ismail; Al-Meer, Saeed H; Ahmad, Talaat A
2016-03-01
The study aimed to examine the residues of organochlorines pesticides (OCPs) in vegetables and fruits in Qatar. A total of 127 samples was studied. Ninety percent of the imported samples recorded residues above the maximum residue levels (MRLs). The most frequently detected OCP in the samples was heptachlor (found in 75 samples). In the comparisons between the washed and unwashed samples, no significant differences were observed (P > 0.05). However, the effect of washing process with tap water depended on the type of vegetables and fruits.
Wente, Stephen P.
2004-01-01
Many Federal, Tribal, State, and local agencies monitor mercury in fish-tissue samples to identify sites with elevated fish-tissue mercury (fish-mercury) concentrations, track changes in fish-mercury concentrations over time, and produce fish-consumption advisories. Interpretation of such monitoring data commonly is impeded by difficulties in separating the effects of sample characteristics (species, tissues sampled, and sizes of fish) from the effects of spatial and temporal trends on fish-mercury concentrations. Without such a separation, variation in fish-mercury concentrations due to differences in the characteristics of samples collected over time or across space can be misattributed to temporal or spatial trends; and/or actual trends in fish-mercury concentration can be misattributed to differences in sample characteristics. This report describes a statistical model and national data set (31,813 samples) for calibrating the aforementioned statistical model that can separate spatiotemporal and sample characteristic effects in fish-mercury concentration data. This model could be useful for evaluating spatial and temporal trends in fishmercury concentrations and developing fish-consumption advisories. The observed fish-mercury concentration data and model predictions can be accessed, displayed geospatially, and downloaded via the World Wide Web (http://emmma.usgs.gov). This report and the associated web site may assist in the interpretation of large amounts of data from widespread fishmercury monitoring efforts.
Radiation detection method and system using the sequential probability ratio test
Nelson, Karl E [Livermore, CA; Valentine, John D [Redwood City, CA; Beauchamp, Brock R [San Ramon, CA
2007-07-17
A method and system using the Sequential Probability Ratio Test to enhance the detection of an elevated level of radiation, by determining whether a set of observations are consistent with a specified model within a given bounds of statistical significance. In particular, the SPRT is used in the present invention to maximize the range of detection, by providing processing mechanisms for estimating the dynamic background radiation, adjusting the models to reflect the amount of background knowledge at the current point in time, analyzing the current sample using the models to determine statistical significance, and determining when the sample has returned to the expected background conditions.
CALIPSO Observations of Near-Cloud Aerosol Properties as a Function of Cloud Fraction
NASA Technical Reports Server (NTRS)
Yang, Weidong; Marshak, Alexander; Varnai, Tamas; Wood, Robert
2015-01-01
This paper uses spaceborne lidar data to study how near-cloud aerosol statistics of attenuated backscatter depend on cloud fraction. The results for a large region around the Azores show that: (1) far-from-cloud aerosol statistics are dominated by samples from scenes with lower cloud fractions, while near-cloud aerosol statistics are dominated by samples from scenes with higher cloud fractions; (2) near-cloud enhancements of attenuated backscatter occur for any cloud fraction but are most pronounced for higher cloud fractions; (3) the difference in the enhancements for different cloud fractions is most significant within 5km from clouds; (4) near-cloud enhancements can be well approximated by logarithmic functions of cloud fraction and distance to clouds. These findings demonstrate that if variability in cloud fraction across the scenes used to composite aerosol statistics are not considered, a sampling artifact will affect these statistics calculated as a function of distance to clouds. For the Azores-region dataset examined here, this artifact occurs mostly within 5 km from clouds, and exaggerates the near-cloud enhancements of lidar backscatter and color ratio by about 30. This shows that for accurate characterization of the changes in aerosol properties with distance to clouds, it is important to account for the impact of changes in cloud fraction.
Analysis of the Einstein sample of early-type galaxies
NASA Technical Reports Server (NTRS)
Eskridge, Paul B.; Fabbiano, Giuseppina
1993-01-01
The EINSTEIN galaxy catalog contains x-ray data for 148 early-type (E and SO) galaxies. A detailed analysis of the global properties of this sample are studied. By comparing the x-ray properties with other tracers of the ISM, as well as with observables related to the stellar dynamics and populations of the sample, we expect to determine more clearly the physical relationships that determine the evolution of early-type galaxies. Previous studies with smaller samples have explored the relationships between x-ray luminosity (L(sub x)) and luminosities in other bands. Using our larger sample and the statistical techniques of survival analysis, a number of these earlier analyses were repeated. For our full sample, a strong statistical correlation is found between L(sub X) and L(sub B) (the probability that the null hypothesis is upheld is P less than 10(exp -4) from a variety of rank correlation tests. Regressions with several algorithms yield consistent results.
Arthur, W J; Markham, O D
1984-04-01
Polonium-210 concentrations were determined for soil, vegetation and small mammal tissues collected at a solid radioactive waste disposal area, near a phosphate ore processing plant and at two rural areas in southeastern Idaho. Polonium concentrations in media sampled near the radioactive waste disposal facility were equal to or less than values from rural area samples, indicating that disposal of solid radioactive waste at the Idaho National Engineering Laboratory Site has not resulted in increased environmental levels of polonium. Concentrations of 210Po in soils, deer mice hide and carcass samples collected near the phosphate processing plant were statistically (P less than or equal to 0.05) greater than the other sampling locations; however, the mean 210Po concentration in soils and small mammal tissues from sampling areas near the phosphate plant were only four and three times greater, respectively, than control values. No statistical (P greater than 0.05) difference was observed for 210Po concentrations in vegetation among any of the sampling locations.
NASA Technical Reports Server (NTRS)
Colarco, P. R.; Kahn, R. A.; Remer, L. A.; Levy, R. C.
2014-01-01
We use the Moderate Resolution Imaging Spectroradiometer (MODIS) satellite aerosol optical thickness (AOT) product to assess the impact of reduced swath width on global and regional AOT statistics and trends. Alongtrack and across-track sampling strategies are employed, in which the full MODIS data set is sub-sampled with various narrow-swath (approximately 400-800 km) and single pixel width (approximately 10 km) configurations. Although view-angle artifacts in the MODIS AOT retrieval confound direct comparisons between averages derived from different sub-samples, careful analysis shows that with many portions of the Earth essentially unobserved, spatial sampling introduces uncertainty in the derived seasonal-regional mean AOT. These AOT spatial sampling artifacts comprise up to 60%of the full-swath AOT value under moderate aerosol loading, and can be as large as 0.1 in some regions under high aerosol loading. Compared to full-swath observations, narrower swath and single pixel width sampling exhibits a reduced ability to detect AOT trends with statistical significance. On the other hand, estimates of the global, annual mean AOT do not vary significantly from the full-swath values as spatial sampling is reduced. Aggregation of the MODIS data at coarse grid scales (10 deg) shows consistency in the aerosol trends across sampling strategies, with increased statistical confidence, but quantitative errors in the derived trends are found even for the full-swath data when compared to high spatial resolution (0.5 deg) aggregations. Using results of a model-derived aerosol reanalysis, we find consistency in our conclusions about a seasonal-regional spatial sampling artifact in AOT Furthermore, the model shows that reduced spatial sampling can amount to uncertainty in computed shortwave top-ofatmosphere aerosol radiative forcing of 2-3 W m(sup-2). These artifacts are lower bounds, as possibly other unconsidered sampling strategies would perform less well. These results suggest that future aerosol satellite missions having significantly less than full-swath viewing are unlikely to sample the true AOT distribution well enough to obtain the statistics needed to reduce uncertainty in aerosol direct forcing of climate.
Space-Time Data fusion for Remote Sensing Applications
NASA Technical Reports Server (NTRS)
Braverman, Amy; Nguyen, H.; Cressie, N.
2011-01-01
NASA has been collecting massive amounts of remote sensing data about Earth's systems for more than a decade. Missions are selected to be complementary in quantities measured, retrieval techniques, and sampling characteristics, so these datasets are highly synergistic. To fully exploit this, a rigorous methodology for combining data with heterogeneous sampling characteristics is required. For scientific purposes, the methodology must also provide quantitative measures of uncertainty that propagate input-data uncertainty appropriately. We view this as a statistical inference problem. The true but notdirectly- observed quantities form a vector-valued field continuous in space and time. Our goal is to infer those true values or some function of them, and provide to uncertainty quantification for those inferences. We use a spatiotemporal statistical model that relates the unobserved quantities of interest at point-level to the spatially aggregated, observed data. We describe and illustrate our method using CO2 data from two NASA data sets.
NASA Technical Reports Server (NTRS)
Nastrom, G. D.; Jasperson, W. H.
1983-01-01
Temperature data obtained by the Global Atmospheric Sampling Program (GASP) during the period March 1975 to July 1979 are compiled to form flight summaries of static air temperature and a geographic temperature climatology. The flight summaries include the height and location of the coldest observed temperature and the mean flight level, temperature and the standard deviation of temperature for each flight as well as for flight segments. These summaries are ordered by route and month. The temperature climatology was computed for all statistically independent temperture data for each flight. The grid used consists of 5 deg latitude, 30 deg longitude and 2000 feet vertical resolution from FL270 to FL430 for each month of the year. The number of statistically independent observations, their mean, standard deviation and the empirical 98, 50, 16, 2 and .3 probability percentiles are presented.
The Complete Local-Volume Groups Sample (CLoGS): Early results from X-ray and radio observations
NASA Astrophysics Data System (ADS)
Vrtilek, Jan M.; O'Sullivan, Ewan; David, Laurence P.; Giacintucci, Simona; Kolokythas, Konstantinos
2017-08-01
Although the group environment is the dominant locus of galaxy evolution (in contrast to rich clusters, which contain only a few percent of galaxies), there has been a lack of reliable, representative group samples in the local Universe. In particular, X-ray selected samples are strongly biased in favor of the X-ray bright, centrally-concentrated cool-core systems. In response, we have designed the Complete Local-Volume Groups Sample (CLoGS), an optically-selected statistically-complete sample of 53 groups within 80 Mpc which is intended to overcome the limitations of X-ray selected samples and serve as a representative survey of groups in the local Universe. We have supplemented X-ray data from Chandra and XMM (70% complete to date, using both archival and new observations, with a 26-group high richness subsample 100% complete) with GMRT radio continuum observations (at 235 and 610 MHz, complete for the entire sample). CLoGS includes groups with a wide variety of properties in terms of galaxy population, hot gas content, and AGN power. We here describe early results from the survey, including the range of AGN activity observed in the dominant galaxies, the relative fraction of cool-core and non-cool-core groups in our sample, and the degree of disturbance observed in the IGM.
Shawna, Wicks; M., Taylor Christopher; Meng, Luo; Eugene, Blanchard IV; David, Ribnicky; T., Cefalu William; L., Mynatt Randall; A., Welsh David
2014-01-01
Objective The gut microbiome has been implicated in obesity and metabolic syndrome; however, most studies have focused on fecal or colonic samples. Several species of Artemisia have been reported to ameliorate insulin signaling both in vitro and in vivo. The aim of this study was to characterize the mucosal and luminal bacterial populations in the terminal ileum with or without supplementation with Artemisia extracts. Materials/Methods Following 4 weeks of supplementation with different Artemisia extracts (PMI 5011, Santa or Scopa), diet-induced obese mice were sacrificed and luminal and mucosal samples of terminal ileum were used to evaluate microbial community composition by pyrosequencing of 16S rDNA hypervariable regions. Results Significant differences in community structure and membership were observed between luminal and mucosal samples, irrespective of diet group. All Artemisia extracts increased the Bacteroidetes:Firmicutes ratio in mucosal samples. This effect was not observed in the luminal compartment. There was high inter-individual variability in the phylogenetic assessments of the ileal microbiota, limiting the statistical power of this pilot investigation. Conclusions Marked differences in bacterial communities exist dependent upon the biogeographic compartment in the terminal ileum. Future studies testing the effects of Artemisia or other botanical supplements require larger sample sizes for adequate statistical power. PMID:24985102
A statistical approach to selecting and confirming validation targets in -omics experiments
2012-01-01
Background Genomic technologies are, by their very nature, designed for hypothesis generation. In some cases, the hypotheses that are generated require that genome scientists confirm findings about specific genes or proteins. But one major advantage of high-throughput technology is that global genetic, genomic, transcriptomic, and proteomic behaviors can be observed. Manual confirmation of every statistically significant genomic result is prohibitively expensive. This has led researchers in genomics to adopt the strategy of confirming only a handful of the most statistically significant results, a small subset chosen for biological interest, or a small random subset. But there is no standard approach for selecting and quantitatively evaluating validation targets. Results Here we present a new statistical method and approach for statistically validating lists of significant results based on confirming only a small random sample. We apply our statistical method to show that the usual practice of confirming only the most statistically significant results does not statistically validate result lists. We analyze an extensively validated RNA-sequencing experiment to show that confirming a random subset can statistically validate entire lists of significant results. Finally, we analyze multiple publicly available microarray experiments to show that statistically validating random samples can both (i) provide evidence to confirm long gene lists and (ii) save thousands of dollars and hundreds of hours of labor over manual validation of each significant result. Conclusions For high-throughput -omics studies, statistical validation is a cost-effective and statistically valid approach to confirming lists of significant results. PMID:22738145
Friedman, David B
2012-01-01
All quantitative proteomics experiments measure variation between samples. When performing large-scale experiments that involve multiple conditions or treatments, the experimental design should include the appropriate number of individual biological replicates from each condition to enable the distinction between a relevant biological signal from technical noise. Multivariate statistical analyses, such as principal component analysis (PCA), provide a global perspective on experimental variation, thereby enabling the assessment of whether the variation describes the expected biological signal or the unanticipated technical/biological noise inherent in the system. Examples will be shown from high-resolution multivariable DIGE experiments where PCA was instrumental in demonstrating biologically significant variation as well as sample outliers, fouled samples, and overriding technical variation that would not be readily observed using standard univariate tests.
Emura, Takeshi; Konno, Yoshihiko; Michimae, Hirofumi
2015-07-01
Doubly truncated data consist of samples whose observed values fall between the right- and left- truncation limits. With such samples, the distribution function of interest is estimated using the nonparametric maximum likelihood estimator (NPMLE) that is obtained through a self-consistency algorithm. Owing to the complicated asymptotic distribution of the NPMLE, the bootstrap method has been suggested for statistical inference. This paper proposes a closed-form estimator for the asymptotic covariance function of the NPMLE, which is computationally attractive alternative to bootstrapping. Furthermore, we develop various statistical inference procedures, such as confidence interval, goodness-of-fit tests, and confidence bands to demonstrate the usefulness of the proposed covariance estimator. Simulations are performed to compare the proposed method with both the bootstrap and jackknife methods. The methods are illustrated using the childhood cancer dataset.
Mróz, Tomasz; Szufa, Katarzyna; Frontasyeva, Marina V; Tselmovich, Vladimir; Ostrovnaya, Tatiana; Kornaś, Andrzej; Olech, Maria A; Mietelski, Jerzy W; Brudecki, Kamil
2018-01-01
Seven lichens (Usnea antarctica and U. aurantiacoatra) and nine moss samples (Sanionia uncinata) collected in King George Island were analyzed using instrumental neutron activation analysis, and concentration of major and trace elements was calculated. For some elements, the concentrations observed in moss samples were higher than corresponding values reported from other sites in the Antarctica, but in the lichens, these were in the same range of concentrations. Scanning electron microscopy (SEM) and statistical analysis showed large influence of volcanic-origin particles. Also, the interplanetary cosmic particles (ICP) were observed in investigated samples, as mosses and lichens are good collectors of ICP and micrometeorites.
MAFsnp: A Multi-Sample Accurate and Flexible SNP Caller Using Next-Generation Sequencing Data
Hu, Jiyuan; Li, Tengfei; Xiu, Zidi; Zhang, Hong
2015-01-01
Most existing statistical methods developed for calling single nucleotide polymorphisms (SNPs) using next-generation sequencing (NGS) data are based on Bayesian frameworks, and there does not exist any SNP caller that produces p-values for calling SNPs in a frequentist framework. To fill in this gap, we develop a new method MAFsnp, a Multiple-sample based Accurate and Flexible algorithm for calling SNPs with NGS data. MAFsnp is based on an estimated likelihood ratio test (eLRT) statistic. In practical situation, the involved parameter is very close to the boundary of the parametric space, so the standard large sample property is not suitable to evaluate the finite-sample distribution of the eLRT statistic. Observing that the distribution of the test statistic is a mixture of zero and a continuous part, we propose to model the test statistic with a novel two-parameter mixture distribution. Once the parameters in the mixture distribution are estimated, p-values can be easily calculated for detecting SNPs, and the multiple-testing corrected p-values can be used to control false discovery rate (FDR) at any pre-specified level. With simulated data, MAFsnp is shown to have much better control of FDR than the existing SNP callers. Through the application to two real datasets, MAFsnp is also shown to outperform the existing SNP callers in terms of calling accuracy. An R package “MAFsnp” implementing the new SNP caller is freely available at http://homepage.fudan.edu.cn/zhangh/softwares/. PMID:26309201
The Abundance of Large Arcs From CLASH
NASA Astrophysics Data System (ADS)
Xu, Bingxiao; Postman, Marc; Meneghetti, Massimo; Coe, Dan A.; Clash Team
2015-01-01
We have developed an automated arc-finding algorithm to perform a rigorous comparison of the observed and simulated abundance of large lensed background galaxies (a.k.a arcs). We use images from the CLASH program to derive our observed arc abundance. Simulated CLASH images are created by performing ray tracing through mock clusters generated by the N-body simulation calibrated tool -- MOKA, and N-body/hydrodynamic simulations -- MUSIC, over the same mass and redshift range as the CLASH X-ray selected sample. We derive a lensing efficiency of 15 ± 3 arcs per cluster for the X-ray selected CLASH sample and 4 ± 2 arcs per cluster for the simulated sample. The marginally significant difference (3.0 σ) between the results for the observations and the simulations can be explained by the systematically smaller area with magnification larger than 3 (by a factor of ˜4) in both MOKA and MUSIC mass models relative to those derived from the CLASH data. Accounting for this difference brings the observed and simulated arc statistics into full agreement. We find that the source redshift distribution does not have big impact on the arc abundance but the arc abundance is very sensitive to the concentration of the dark matter halos. Our results suggest that the solution to the "arc statistics problem" lies primarily in matching the cluster dark matter distribution.
Cotruta, Bogdan; Gheorghe, Cristian; Iacob, Razvan; Dumbrava, Mona; Radu, Cristina; Bancila, Ion; Becheanu, Gabriel
2017-12-01
Evaluation of severity and extension of gastric atrophy and intestinal metaplasia is recommended to identify subjects with a high risk for gastric cancer. The inter-observer agreement for the assessment of gastric atrophy is reported to be low. The aim of the study was to evaluate the inter-observer agreement for the assessment of severity and extension of gastric atrophy using oriented and unoriented gastric biopsy samples. Furthermore, the quality of biopsy specimens in oriented and unoriented samples was analyzed. A total of 35 subjects with dyspeptic symptoms addressed for gastrointestinal endoscopy that agreed to enter the study were prospectively enrolled. The OLGA/OLGIM gastric biopsies protocol was used. From each subject two sets of biopsies were obtained (four from the antrum, two oriented and two unoriented, two from the gastric incisure, one oriented and one unoriented, four from the gastric body, two oriented and two unoriented). The orientation of the biopsy samples was completed using nitrocellulose filters (Endokit®, BioOptica, Milan, Italy). The samples were blindly examined by two experienced pathologists. Inter-observer agreement was evaluated using kappa statistic for inter-rater agreement. The quality of histopathology specimens taking into account the identification of lamina propria was analyzed in oriented vs. unoriented samples. The samples with detectable lamina propria mucosae were defined as good quality specimens. Categorical data was analyzed using chi-square test and a two-sided p value <0.05 was considered statistically significant. A total of 350 biopsy samples were analyzed (175 oriented / 175 unoriented). The kappa index values for oriented/unoriented OLGA 0/I/II/III and IV stages have been 0.62/0.13, 0.70/0.20, 0.61/0.06, 0.62/0.46, and 0.77/0.50, respectively. For OLGIM 0/I/II/III stages the kappa index values for oriented/unoriented samples were 0.83/0.83, 0.88/0.89, 0.70/0.88 and 0.83/1, respectively. No case of OLGIM IV stage was found in the present case series. Good quality histopathology specimens were described in 95.43% of the oriented biopsy samples, and in 89.14% of the unoriented biopsy samples, respectively (p=0.0275). The orientation of gastric biopsies specimens improves the inter-observer agreement for the assessment of gastric atrophy.
Stanley, Brett J; Antonio, Karen
2012-11-01
Several methods exist for the measurement of cyanide levels in treated wastewater,typically requiring preservation of the sample with sodium hydroxide to minimize loss of hydrogen cyanide gas (HCN). Recent reports have shown that cyanide levels may increase with chlorination or preservation. In this study, three flow injection analysis methods involving colorimetric and amperometric detection were compared within one laboratory, as well as across separate laboratories and equipment. Split wastewater samples from eight facilities and three different sampling periods were tested. An interlaboratory confidence interval of 3.5 ppb was calculated compared with the intralaboratory reporting limit of 2 ppb. The results show that free cyanide measurements are not statistically different than total cyanide levels. An artificial increase in cyanide level is observed with all methods for preserved samples relative to nonpreserved samples, with an average increase of 2.3 ppb. The possible loss of cyanide without preservation is shown to be statistically insignificant if properly stored up to 48 hours. The cyanide increase with preservation is further substantiated with the method of standard additions and is not a matrix interference. The increase appears to be correlated with the amount of cyanide observed without preservation, which appears to be greater in those facilities that disinfect their wastewater with chlorine, followed by dechlorination with sodium bisulfite.
Association of Blastocystis subtypes with diarrhea in children
NASA Astrophysics Data System (ADS)
Zulfa, F.; Sari, I. P.; Kurniawan, A.
2017-08-01
Blastocystis hominis is an intestinal zoonotic protozoa that epidemiological surveys have shown, is highly prevalent among children and may cause chronic diarrhea. This study aimed to identify Blastocystis subtypes among children and associate those subtypes to pathology. The study’s population was children aged 6-12 years old divided into asymptomatic and symptomatic (diarrhea) groups. The asymptomatic samples were obtained from primary school students in the Bukit Duri area of South Jakarta, while the symptomatic samples were obtained from patients who visited nearby primary health centers (Puskesmas). Symptomatic stool samples were examined inParasitology Laboratory FKUI. Microscopic examination of the stool samples was performed to screen for single Blastocystic infection, followed by culture, PCR of 18S rRNA, and sequencing. In the study, 53.2% of children (n = 156) harbored intestinal parasites, Blastocysts sp. A single infection of Blastocystis sp. was present in 69 (44.23%) samples, comprised of 36 symptomatic and 33 asymptomatic participants. The Blastocystis subtypes (STs) identified in this study were STs 1-4 ST3 was the most dominant and was observed with statistically significant higher frequency in the symptomatic group. ST4 was only found in one sample in the symptomatic group. While ST1 and ST2 were found more frequently in the asymptomatic group, no statistical association was observed. ST3 is more likely to be associated with clinical symptoms than ST1 and ST2.
Visual statistical learning is not reliably modulated by selective attention to isolated events
Musz, Elizabeth; Weber, Matthew J.; Thompson-Schill, Sharon L.
2014-01-01
Recent studies of visual statistical learning (VSL) indicate that the visual system can automatically extract temporal and spatial relationships between objects. We report several attempts to replicate and extend earlier work (Turk-Browne et al., 2005) in which observers performed a cover task on one of two interleaved stimulus sets, resulting in learning of temporal relationships that occur in the attended stream, but not those present in the unattended stream. Across four experiments, we exposed observers to a similar or identical familiarization protocol, directing attention to one of two interleaved stimulus sets; afterward, we assessed VSL efficacy for both sets using either implicit response-time measures or explicit familiarity judgments. In line with prior work, we observe learning for the attended stimulus set. However, unlike previous reports, we also observe learning for the unattended stimulus set. When instructed to selectively attend to only one of the stimulus sets and ignore the other set, observers could extract temporal regularities for both sets. Our efforts to experimentally decrease this effect by changing the cover task (Experiment 1) or the complexity of the statistical regularities (Experiment 3) were unsuccessful. A fourth experiment using a different assessment of learning likewise failed to show an attentional effect. Simulations drawing random samples our first three experiments (n=64) confirm that the distribution of attentional effects in our sample closely approximates the null. We offer several potential explanations for our failure to replicate earlier findings, and discuss how our results suggest limiting conditions on the relevance of attention to VSL. PMID:25172196
Torimitsu, Suguru; Nishida, Yoshifumi; Takano, Tachio; Koizumi, Yoshinori; Makino, Yohsuke; Yajima, Daisuke; Hayakawa, Mutsumi; Inokuchi, Go; Motomura, Ayumi; Chiba, Fumiko; Otsuka, Katsura; Kobayashi, Kazuhiro; Odo, Yuriko; Iwase, Hirotaro
2014-01-01
The purpose of this research was to investigate the biomechanical properties of the adult human skull and the structural changes that occur with age in both sexes. The heads of 94 Japanese cadavers (54 male cadavers, 40 female cadavers) autopsied in our department were used in this research. A total of 376 cranial samples, four from each skull, were collected. Sample fracture load was measured by a bending test. A statistically significant negative correlation between the sample fracture load and cadaver age was found. This indicates that the stiffness of cranial bones in Japanese individuals decreases with age, and the risk of skull fracture thus probably increases with age. Prior to the bending test, the sample mass, the sample thickness, the ratio of the sample thickness to cadaver stature (ST/CS), and the sample density were measured and calculated. Significant negative correlations between cadaver age and sample thickness, ST/CS, and the sample density were observed only among the female samples. Computerized tomographic (CT) images of 358 cranial samples were available. The computed tomography value (CT value) of cancellous bone which refers to a quantitative scale for describing radiodensity, cancellous bone thickness and cortical bone thickness were measured and calculated. Significant negative correlation between cadaver age and the CT value or cortical bone thickness was observed only among the female samples. These findings suggest that the skull is substantially affected by decreased bone metabolism resulting from osteoporosis. Therefore, osteoporosis prevention and treatment may increase cranial stiffness and reinforce the skull structure, leading to a decrease in the risk of skull fractures. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Kovalenko, I. D.; Doressoundiram, A.; Lellouch, E.; Vilenius, E.; Müller, T.; Stansberry, J.
2017-11-01
Context. Gravitationally bound multiple systems provide an opportunity to estimate the mean bulk density of the objects, whereas this characteristic is not available for single objects. Being a primitive population of the outer solar system, binary and multiple trans-Neptunian objects (TNOs) provide unique information about bulk density and internal structure, improving our understanding of their formation and evolution. Aims: The goal of this work is to analyse parameters of multiple trans-Neptunian systems, observed with Herschel and Spitzer space telescopes. Particularly, statistical analysis is done for radiometric size and geometric albedo, obtained from photometric observations, and for estimated bulk density. Methods: We use Monte Carlo simulation to estimate the real size distribution of TNOs. For this purpose, we expand the dataset of diameters by adopting the Minor Planet Center database list with available values of the absolute magnitude therein, and the albedo distribution derived from Herschel radiometric measurements. We use the 2-sample Anderson-Darling non-parametric statistical method for testing whether two samples of diameters, for binary and single TNOs, come from the same distribution. Additionally, we use the Spearman's coefficient as a measure of rank correlations between parameters. Uncertainties of estimated parameters together with lack of data are taken into account. Conclusions about correlations between parameters are based on statistical hypothesis testing. Results: We have found that the difference in size distributions of multiple and single TNOs is biased by small objects. The test on correlations between parameters shows that the effective diameter of binary TNOs strongly correlates with heliocentric orbital inclination and with magnitude difference between components of binary system. The correlation between diameter and magnitude difference implies that small and large binaries are formed by different mechanisms. Furthermore, the statistical test indicates, although not significant with the sample size, that a moderately strong correlation exists between diameter and bulk density. Herschel is an ESA space observatory with science instruments provided by European-led Principal Investigator consortia and with important participation from NASA.
Quality of human milk expressed in a human milk bank and at home.
Borges, Mayla S; Oliveira, Angela M de M; Hattori, Wallisen T; Abdallah, Vânia O S
2017-08-30
To evaluate the quality of the human milk expressed at home and at a human milk bank. This a retrospective, analytical, and observational study, performed by assessing titratable acidity records and the microbiological culture of 100 human milk samples expressed at home and at a human milk bank, in 2014. For the statistical analysis, generalized estimating equations (GEE) and the chi-squared test were used. When comparing the two sample groups, no significant difference was found, with 98% and 94% of the samples being approved among those collected at the milk bank and at home, respectively. No main interaction effect between local and titratable acidity records (p=0.285) was observed, and there was no statistically significant difference between the expected and observed values for the association between the collection place and the microbiological culture results (p=0.307). The quality of human milk expressed at home and at the milk bank are in agreement with the recommended standards, confirming that the expression of human milk at home is as safe as expression at the human milk bank, provided that the established hygiene, conservation, storage, and transport standards are followed. Copyright © 2017 Sociedade Brasileira de Pediatria. Published by Elsevier Editora Ltda. All rights reserved.
Testing for X-Ray–SZ Differences and Redshift Evolution in the X-Ray Morphology of Galaxy Clusters
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nurgaliev, D.; McDonald, M.; Benson, B. A.
We present a quantitative study of the X-ray morphology of galaxy clusters, as a function of their detection method and redshift. We analyze two separate samples of galaxy clusters: a sample of 36 clusters atmore » $$0.35\\lt z\\lt 0.9$$ selected in the X-ray with the ROSAT PSPC 400 deg(2) survey, and a sample of 90 clusters at $$0.25\\lt z\\lt 1.2$$ selected via the Sunyaev–Zel’dovich (SZ) effect with the South Pole Telescope. Clusters from both samples have similar-quality Chandra observations, which allow us to quantify their X-ray morphologies via two distinct methods: centroid shifts (w) and photon asymmetry ($${A}_{\\mathrm{phot}}$$). The latter technique provides nearly unbiased morphology estimates for clusters spanning a broad range of redshift and data quality. We further compare the X-ray morphologies of X-ray- and SZ-selected clusters with those of simulated clusters. We do not find a statistically significant difference in the measured X-ray morphology of X-ray and SZ-selected clusters over the redshift range probed by these samples, suggesting that the two are probing similar populations of clusters. We find that the X-ray morphologies of simulated clusters are statistically indistinguishable from those of X-ray- or SZ-selected clusters, implying that the most important physics for dictating the large-scale gas morphology (outside of the core) is well-approximated in these simulations. Finally, we find no statistically significant redshift evolution in the X-ray morphology (both for observed and simulated clusters), over the range of $$z\\sim 0.3$$ to $$z\\sim 1$$, seemingly in contradiction with the redshift-dependent halo merger rate predicted by simulations.« less
Testing for X-Ray–SZ Differences and Redshift Evolution in the X-Ray Morphology of Galaxy Clusters
Nurgaliev, D.; McDonald, M.; Benson, B. A.; ...
2017-05-16
We present a quantitative study of the X-ray morphology of galaxy clusters, as a function of their detection method and redshift. We analyze two separate samples of galaxy clusters: a sample of 36 clusters atmore » $$0.35\\lt z\\lt 0.9$$ selected in the X-ray with the ROSAT PSPC 400 deg(2) survey, and a sample of 90 clusters at $$0.25\\lt z\\lt 1.2$$ selected via the Sunyaev–Zel’dovich (SZ) effect with the South Pole Telescope. Clusters from both samples have similar-quality Chandra observations, which allow us to quantify their X-ray morphologies via two distinct methods: centroid shifts (w) and photon asymmetry ($${A}_{\\mathrm{phot}}$$). The latter technique provides nearly unbiased morphology estimates for clusters spanning a broad range of redshift and data quality. We further compare the X-ray morphologies of X-ray- and SZ-selected clusters with those of simulated clusters. We do not find a statistically significant difference in the measured X-ray morphology of X-ray and SZ-selected clusters over the redshift range probed by these samples, suggesting that the two are probing similar populations of clusters. We find that the X-ray morphologies of simulated clusters are statistically indistinguishable from those of X-ray- or SZ-selected clusters, implying that the most important physics for dictating the large-scale gas morphology (outside of the core) is well-approximated in these simulations. Finally, we find no statistically significant redshift evolution in the X-ray morphology (both for observed and simulated clusters), over the range of $$z\\sim 0.3$$ to $$z\\sim 1$$, seemingly in contradiction with the redshift-dependent halo merger rate predicted by simulations.« less
Muths, Delphine; Le Couls, Sarah; Evano, Hugues; Grewe, Peter; Bourjea, Jerome
2013-01-01
Genetic population structure of swordfish Xiphias gladius was examined based on 2231 individual samples, collected mainly between 2009 and 2010, among three major sampling areas within the Indian Ocean (IO; twelve distinct sites), Atlantic (two sites) and Pacific (one site) Oceans using analysis of nineteen microsatellite loci (n = 2146) and mitochondrial ND2 sequences (n = 2001) data. Sample collection was stratified in time and space in order to investigate the stability of the genetic structure observed with a special focus on the South West Indian Ocean. Significant AMOVA variance was observed for both markers indicating genetic population subdivision was present between oceans. Overall value of F-statistics for ND2 sequences confirmed that Atlantic and Indian Oceans swordfish represent two distinct genetic stocks. Indo-Pacific differentiation was also significant but lower than that observed between Atlantic and Indian Oceans. However, microsatellite F-statistics failed to reveal structure even at the inter-oceanic scale, indicating that resolving power of our microsatellite loci was insufficient for detecting population subdivision. At the scale of the Indian Ocean, results obtained from both markers are consistent with swordfish belonging to a single unique panmictic population. Analyses partitioned by sampling area, season, or sex also failed to identify any clear structure within this ocean. Such large spatial and temporal homogeneity of genetic structure, observed for such a large highly mobile pelagic species, suggests as satisfactory to consider swordfish as a single panmictic population in the Indian Ocean. PMID:23717447
Ten Years of Gamma-Ray Bursts Observations with BATSE
NASA Technical Reports Server (NTRS)
Fishman, Gerald J.; Rose, M. Franklin (Technical Monitor)
2000-01-01
The observed gamma-ray temporal, spectral, intensity and spatial distribution characteristics of GRBs from data obtained from BATSE/Compton Observatory, will be described. The talk will concentrate on recent studies of burst properties, correlations of GRB parameters and other statistical studies that have only recently come to light with the unprecedented sample of over 2700 GRBS. Recent studies of possible observational biases, un-triggered GRBs and threshold calculations for BATSE will also be described.
Tian, Guo-Liang; Li, Hui-Qiong
2017-08-01
Some existing confidence interval methods and hypothesis testing methods in the analysis of a contingency table with incomplete observations in both margins entirely depend on an underlying assumption that the sampling distribution of the observed counts is a product of independent multinomial/binomial distributions for complete and incomplete counts. However, it can be shown that this independency assumption is incorrect and can result in unreliable conclusions because of the under-estimation of the uncertainty. Therefore, the first objective of this paper is to derive the valid joint sampling distribution of the observed counts in a contingency table with incomplete observations in both margins. The second objective is to provide a new framework for analyzing incomplete contingency tables based on the derived joint sampling distribution of the observed counts by developing a Fisher scoring algorithm to calculate maximum likelihood estimates of parameters of interest, the bootstrap confidence interval methods, and the bootstrap testing hypothesis methods. We compare the differences between the valid sampling distribution and the sampling distribution under the independency assumption. Simulation studies showed that average/expected confidence-interval widths of parameters based on the sampling distribution under the independency assumption are shorter than those based on the new sampling distribution, yielding unrealistic results. A real data set is analyzed to illustrate the application of the new sampling distribution for incomplete contingency tables and the analysis results again confirm the conclusions obtained from the simulation studies.
On the Seasonality of Sudden Stratospheric Warmings
NASA Astrophysics Data System (ADS)
Reichler, T.; Horan, M.
2017-12-01
The downward influence of sudden stratospheric warmings (SSWs) creates significant tropospheric circulation anomalies that last for weeks. It is therefore of theoretical and practical interest to understand the time when SSWs are most likely to occur and the controlling factors for the temporal distribution of SSWs. Conceivably, the distribution between mid-winter and late-winter is controlled by the interplay between decreasing eddy convergence in the region of the polar vortex and the weakening strength of the polar vortex. General circulation models (GCMs) tend to produce SSW maxima later in winter than observations, which has been considered as a model deficiency. However, the observed record is short, suggesting that under-sampling of SSWs may contribute to this discrepancy. Here, we study the climatological frequency distribution of SSWs and related events in a long control simulation with a stratosphere resolving GCM. We also create a simple statistical model to determine the primary factors controlling the SSW distribution. The statistical model is based on the daily climatological mean, standard deviation, and autocorrelation of stratospheric winds, and assumes that the winds follow a normal distribution. We find that the null hypothesis, that model and observations stem from the same distribution, cannot be rejected, suggesting that the mid-winter SSW maximum seen in the observations is due to sampling uncertainty. We also find that the statistical model faithfully reproduces the seasonal distribution of SSWs, and that the decreasing climatological strength of the polar vortex is the primary factor for it. We conclude that the late-winter SSW maximum seen in most models is realistic and that late events will be more prominent in future observations. We further conclude that SSWs simply form the tail of normally distributed stratospheric winds, suggesting that there is a continuum of weak polar vortex states and that statistically there is nothing special about the zero-threshold used to define SSWs.
42 CFR 402.109 - Statistical sampling.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 42 Public Health 2 2011-10-01 2011-10-01 false Statistical sampling. 402.109 Section 402.109... Statistical sampling. (a) Purpose. CMS or OIG may introduce the results of a statistical sampling study to... or caused to be presented. (b) Prima facie evidence. The results of the statistical sampling study...
42 CFR 402.109 - Statistical sampling.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 42 Public Health 2 2010-10-01 2010-10-01 false Statistical sampling. 402.109 Section 402.109... Statistical sampling. (a) Purpose. CMS or OIG may introduce the results of a statistical sampling study to... or caused to be presented. (b) Prima facie evidence. The results of the statistical sampling study...
Detecting cell death with optical coherence tomography and envelope statistics
NASA Astrophysics Data System (ADS)
Farhat, Golnaz; Yang, Victor X. D.; Czarnota, Gregory J.; Kolios, Michael C.
2011-02-01
Currently no standard clinical or preclinical noninvasive method exists to monitor cell death based on morphological changes at the cellular level. In our past work we have demonstrated that quantitative high frequency ultrasound imaging can detect cell death in vitro and in vivo. In this study we apply quantitative methods previously used with high frequency ultrasound to optical coherence tomography (OCT) to detect cell death. The ultimate goal of this work is to use these methods for optically-based clinical and preclinical cancer treatment monitoring. Optical coherence tomography data were acquired from acute myeloid leukemia cells undergoing three modes of cell death. Significant increases in integrated backscatter were observed for cells undergoing apoptosis and mitotic arrest, while necrotic cells induced a decrease. These changes appear to be linked to structural changes observed in histology obtained from the cell samples. Signal envelope statistics were analyzed from fittings of the generalized gamma distribution to histograms of envelope intensities. The parameters from this distribution demonstrated sensitivities to morphological changes in the cell samples. These results indicate that OCT integrated backscatter and first order envelope statistics can be used to detect and potentially differentiate between modes of cell death in vitro.
Statistical analysis of radioimmunoassay. In comparison with bioassay (in Japanese)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nakano, R.
1973-01-01
Using the data of RIA (radioimmunoassay), statistical procedures for dealing with two problems of the linearization of dose response curve and calculation of relative potency were described. There were three methods for linearization of dose response curve of RIA. In each method, the following parameters were shown on the horizontal and vertical axis: dose x, (B/T)/sup -1/; c/x + c, B/T (C: dose which makes B/T 50%); log x, logit B/T. Among them, the last method seems to be most practical. The statistical procedures for bioassay were employed for calculating the relative potency of unknown samples compared to the standardmore » samples from dose response curves of standand and unknown samples using regression coefficient. It is desirable that relative potency is calculated by plotting more than 5 points in the standard curve and plotting more than 2 points in unknow samples. For examining the statistical limit of precision of measuremert, LH activity of gonadotropin in urine was measured and relative potency, precision coefficient and the upper and lower limits of relative potency at 95% confidence limit were calculated. On the other hand, bioassay (by the ovarian ascorbic acid reduction method and anteriol lobe of prostate weighing method) was done in the same samples, and the precision was compared with that of RIA. In these examinations, the upper and lower limits of the relative potency at 95% confidence limit were near each other, while in bioassay, a considerable difference was observed between the upper and lower limits. The necessity of standardization and systematization of the statistical procedures for increasing the precision of RIA was pointed out. (JA)« less
Robust matching for voice recognition
NASA Astrophysics Data System (ADS)
Higgins, Alan; Bahler, L.; Porter, J.; Blais, P.
1994-10-01
This paper describes an automated method of comparing a voice sample of an unknown individual with samples from known speakers in order to establish or verify the individual's identity. The method is based on a statistical pattern matching approach that employs a simple training procedure, requires no human intervention (transcription, work or phonetic marketing, etc.), and makes no assumptions regarding the expected form of the statistical distributions of the observations. The content of the speech material (vocabulary, grammar, etc.) is not assumed to be constrained in any way. An algorithm is described which incorporates frame pruning and channel equalization processes designed to achieve robust performance with reasonable computational resources. An experimental implementation demonstrating the feasibility of the concept is described.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Öztürk, Hande; Noyan, I. Cevdet
A rigorous study of sampling and intensity statistics applicable for a powder diffraction experiment as a function of crystallite size is presented. Our analysis yields approximate equations for the expected value, variance and standard deviations for both the number of diffracting grains and the corresponding diffracted intensity for a given Bragg peak. The classical formalism published in 1948 by Alexander, Klug & Kummer [J. Appl. Phys.(1948),19, 742–753] appears as a special case, limited to large crystallite sizes, here. It is observed that both the Lorentz probability expression and the statistics equations used in the classical formalism are inapplicable for nanocrystallinemore » powder samples.« less
Öztürk, Hande; Noyan, I. Cevdet
2017-08-24
A rigorous study of sampling and intensity statistics applicable for a powder diffraction experiment as a function of crystallite size is presented. Our analysis yields approximate equations for the expected value, variance and standard deviations for both the number of diffracting grains and the corresponding diffracted intensity for a given Bragg peak. The classical formalism published in 1948 by Alexander, Klug & Kummer [J. Appl. Phys.(1948),19, 742–753] appears as a special case, limited to large crystallite sizes, here. It is observed that both the Lorentz probability expression and the statistics equations used in the classical formalism are inapplicable for nanocrystallinemore » powder samples.« less
Area estimation using multiyear designs and partial crop identification
NASA Technical Reports Server (NTRS)
Sielken, R. L., Jr.
1984-01-01
Statistical procedures were developed for large area assessments using both satellite and conventional data. Crop acreages, other ground cover indices, and measures of change were the principal characteristics of interest. These characteristics are capable of being estimated from samples collected possibly from several sources at varying times, with different levels of identification. Multiyear analysis techniques were extended to include partially identified samples; the best current year sampling design corresponding to a given sampling history was determined; weights reflecting the precision or confidence in each observation were identified and utilized, and the variation in estimates incorporating partially identified samples were quantified.
Yang, Yang; DeGruttola, Victor
2016-01-01
Traditional resampling-based tests for homogeneity in covariance matrices across multiple groups resample residuals, that is, data centered by group means. These residuals do not share the same second moments when the null hypothesis is false, which makes them difficult to use in the setting of multiple testing. An alternative approach is to resample standardized residuals, data centered by group sample means and standardized by group sample covariance matrices. This approach, however, has been observed to inflate type I error when sample size is small or data are generated from heavy-tailed distributions. We propose to improve this approach by using robust estimation for the first and second moments. We discuss two statistics: the Bartlett statistic and a statistic based on eigen-decomposition of sample covariance matrices. Both statistics can be expressed in terms of standardized errors under the null hypothesis. These methods are extended to test homogeneity in correlation matrices. Using simulation studies, we demonstrate that the robust resampling approach provides comparable or superior performance, relative to traditional approaches, for single testing and reasonable performance for multiple testing. The proposed methods are applied to data collected in an HIV vaccine trial to investigate possible determinants, including vaccine status, vaccine-induced immune response level and viral genotype, of unusual correlation pattern between HIV viral load and CD4 count in newly infected patients. PMID:22740584
Yang, Yang; DeGruttola, Victor
2012-06-22
Traditional resampling-based tests for homogeneity in covariance matrices across multiple groups resample residuals, that is, data centered by group means. These residuals do not share the same second moments when the null hypothesis is false, which makes them difficult to use in the setting of multiple testing. An alternative approach is to resample standardized residuals, data centered by group sample means and standardized by group sample covariance matrices. This approach, however, has been observed to inflate type I error when sample size is small or data are generated from heavy-tailed distributions. We propose to improve this approach by using robust estimation for the first and second moments. We discuss two statistics: the Bartlett statistic and a statistic based on eigen-decomposition of sample covariance matrices. Both statistics can be expressed in terms of standardized errors under the null hypothesis. These methods are extended to test homogeneity in correlation matrices. Using simulation studies, we demonstrate that the robust resampling approach provides comparable or superior performance, relative to traditional approaches, for single testing and reasonable performance for multiple testing. The proposed methods are applied to data collected in an HIV vaccine trial to investigate possible determinants, including vaccine status, vaccine-induced immune response level and viral genotype, of unusual correlation pattern between HIV viral load and CD4 count in newly infected patients.
Sequential CFAR detectors using a dead-zone limiter
NASA Astrophysics Data System (ADS)
Tantaratana, Sawasd
1990-09-01
The performances of some proposed sequential constant-false-alarm-rate (CFAR) detectors are evaluated. The observations are passed through a dead-zone limiter, the output of which is -1, 0, or +1, depending on whether the input is less than -c, between -c and c, or greater than c, where c is a constant. The test statistic is the sum of the outputs. The test is performed on a reduced set of data (those with absolute value larger than c), with the test statistic being the sum of the signs of the reduced set of data. Both constant and linear boundaries are considered. Numerical results show a significant reduction of the average number of observations needed to achieve the same false alarm and detection probabilities as a fixed-sample-size CFAR detector using the same kind of test statistic.
NASA Technical Reports Server (NTRS)
Matney, M.; Barker, E.; Seitzer, P.; Abercromby, K. J.; Rodriquez, H. M.
2006-01-01
NASA's Orbital Debris measurements program has a goal to characterize the small debris environment in the geosynchronous Earth-orbit (GEO) region using optical telescopes ("small" refers to objects too small to catalog and track with current systems). Traditionally, observations of GEO and near-GEO objects involve following the object with the telescope long enough to obtain an orbit suitable for tracking purposes. Telescopes operating in survey mode, however, randomly observe objects that pass through their field of view. Typically, these short-arc observation are inadequate to obtain detailed orbits, but can be used to estimate approximate circular orbit elements (semimajor axis, inclination, and ascending node). From this information, it should be possible to make statistical inferences about the orbital distributions of the GEO population bright enough to be observed by the system. The Michigan Orbital Debris Survey Telescope (MODEST) has been making such statistical surveys of the GEO region for four years. During that time, the telescope has made enough observations in enough areas of the GEO belt to have had nearly complete coverage. That means that almost all objects in all possible orbits in the GEO and near- GEO region had a non-zero chance of being observed. Some regions (such as those near zero inclination) have had good coverage, while others are poorly covered. Nevertheless, it is possible to remove these statistical biases and reconstruct the orbit populations within the limits of sampling error. In this paper, these statistical techniques and assumptions are described, and the techniques are applied to the current MODEST data set to arrive at our best estimate of the GEO orbit population distribution.
Hitting Is Contagious in Baseball: Evidence from Long Hitting Streaks
Bock, Joel R.; Maewal, Akhilesh; Gough, David A.
2012-01-01
Data analysis is used to test the hypothesis that “hitting is contagious”. A statistical model is described to study the effect of a hot hitter upon his teammates’ batting during a consecutive game hitting streak. Box score data for entire seasons comprising streaks of length games, including a total observations were compiled. Treatment and control sample groups () were constructed from core lineups of players on the streaking batter’s team. The percentile method bootstrap was used to calculate confidence intervals for statistics representing differences in the mean distributions of two batting statistics between groups. Batters in the treatment group (hot streak active) showed statistically significant improvements in hitting performance, as compared against the control. Mean for the treatment group was found to be to percentage points higher during hot streaks (mean difference increased points), while the batting heat index introduced here was observed to increase by points. For each performance statistic, the null hypothesis was rejected at the significance level. We conclude that the evidence suggests the potential existence of a “statistical contagion effect”. Psychological mechanisms essential to the empirical results are suggested, as several studies from the scientific literature lend credence to contagious phenomena in sports. Causal inference from these results is difficult, but we suggest and discuss several latent variables that may contribute to the observed results, and offer possible directions for future research. PMID:23251507
VizieR Online Data Catalog: GOALS sample PACS and SPIRE fluxes (Chu+, 2017)
NASA Astrophysics Data System (ADS)
Chu, J. K.; Sanders, D. B.; Larson, K. L.; Mazzarella, J. M.; Howell, J. H.; Diaz-Santos, T.; Xu, K. C.; Paladini, R.; Schulz, B.; Shupe, D.; Appleton, P.; Armus, L.; Billot, N.; Chan, B. H. P.; Evans, A. S.; Fadda, D.; Frayer, D. T.; Haan, S.; Ishida, C. M.; Iwasawa, K.; Kim, D.-C.; Lord, S.; Murphy, E.; Petric, A.; Privon, G. C.; Surace, J. A.; Treister, E.
2017-06-01
The IRAS RBGS contains 179 LIRGs (log(LIR/Lȯ)= 22 ultra-luminous infrared galaxies (ULIRGs: log(LIR/Lȯ)>=12.0); these 201 total objects comprise the GOALS sample (Armus et al. 2009), a statistically complete flux-limited sample of infrared-luminous galaxies in the local universe. This paper presents imaging and photometry for all 201 LIRGs and LIRG systems in the IRAS RBGS that were observed during our GOALS Herschel OT1 program. (4 data files).
Effects of sterilization on poly(ethylene glycol) hydrogels.
Kanjickal, Deenu; Lopina, Stephanie; Evancho-Chapman, M Michelle; Schmidt, Steven; Donovan, Duane
2008-12-01
The past few decades have witnessed a dramatic increase in the development of polymeric biomaterials. These biomaterials have to undergo a sterilization procedure before implantation. However, many sterilization procedures have been shown to profoundly affect polymer properties. Poly(ethylene glycol) hydrogels have gained increasing importance in the controlled delivery of therapeutics and in tissue engineering. We evaluated the effect of ethylene oxide (EtO), hydrogen peroxide (H(2)O(2)), and gamma sterilization of poly(ethylene glycol) hydrogels on properties relevant to controlled drug delivery and tissue engineering. We observed that the release of cyclosporine (CyA) (an immunosuppressive drug that is effective in combating tissue rejection following organ transplantation) was significantly affected by the type of sterilization. However, that was not the case with rhodamine B, a dye. Hence, the drug release characteristics were observed to be dependent not only on the sterilization procedure but also on the type of agent that needs to be delivered. In addition, differences in the swelling ratios for the sterilized and unsterilized hydrogels were statistically significant for 1:1 crosslinked hydrogels derived from the 8000 MW polymer. Significant differences were also observed for gamma sterilization for 1:1 crosslinked hydrogels derived from the 3350 MW polymer and also the 2:1 crosslinked hydrogels derived from the 8000 MW polymer. Atomic force microscopy (AFM) studies revealed that the roughness parameter for the unsterilized and EtO-sterilized PEG hydrogels remained similar. However, a statistically significant reduction of the roughness parameter was observed for the H(2)O(2) and gamma-sterilized samples. Electron spin resonance (ESR) studies on the unsterilized and the sterilized samples revealed the presence of the peroxy and the triphenyl methyl carbon radical in the samples. The gamma and the H(2)O(2)-sterilized samples were observed to have a much higher concentration of the radical pecies when compared with the EtO and the unsterilized samples. (c) 2008 Wiley Periodicals, Inc. J Biomed Mater Res, 2008.
Code of Federal Regulations, 2014 CFR
2014-07-01
... methods employed in statistical compilations. The principal title of each exhibit should state what it... furnished: (i) Market research. (a) The following data and information shall be provided: (1) A clear and detailed description of the sample, observational, and data preparation designs, including definitions of...
Code of Federal Regulations, 2013 CFR
2013-07-01
... item of information used and the methods employed in statistical compilations. The principal title of... furnished: (i) Market research. (a) The following data and information shall be provided: (1) A clear and detailed description of the sample, observational, and data preparation designs, including definitions of...
Code of Federal Regulations, 2012 CFR
2012-07-01
... item of information used and the methods employed in statistical compilations. The principal title of... should be furnished: (i) Market research. (a) The following data and information shall be provided: (1) A clear and detailed description of the sample, observational, and data preparation designs, including...
Code of Federal Regulations, 2011 CFR
2011-07-01
... item of information used and the methods employed in statistical compilations. The principal title of... should be furnished: (i) Market research. (a) The following data and information shall be provided: (1) A clear and detailed description of the sample, observational, and data preparation designs, including...
Code of Federal Regulations, 2010 CFR
2010-07-01
... item of information used and the methods employed in statistical compilations. The principal title of... should be furnished: (i) Market research. (a) The following data and information shall be provided: (1) A clear and detailed description of the sample, observational, and data preparation designs, including...
Peitzsch, Mirko; Sulyok, Michael; Täubel, Martin; Vishwanath, Vinay; Krop, Esmeralda; Borràs-Santos, Alicia; Hyvärinen, Anne; Nevalainen, Aino; Krska, Rudolf; Larsson, Lennart
2012-08-01
Secondary metabolites produced by fungi and bacteria are among the potential agents that contribute to adverse health effects observed in occupants of buildings affected by moisture damage, dampness and associated microbial growth. However, few attempts have been made to assess the occurrence of these compounds in relation to moisture damage and dampness in buildings. This study conducted in the context of the HITEA project (Health Effects of Indoor Pollutants: Integrating microbial, toxicological and epidemiological approaches) aimed at providing systematic information on the prevalence of microbial secondary metabolites in a large number of school buildings in three European countries, considering both buildings with and without moisture damage and/or dampness observations. In order to address the multitude and diversity of secondary metabolites a large number of more than 180 analytes was targeted in settled dust and surface swab samples using liquid chromatography/mass spectrometry (LC/MS) based methodology. While 42%, 58% and 44% of all samples collected in Spanish, Dutch and Finnish schools, respectively, were positive for at least one of the metabolites analyzed, frequency of detection for the individual microbial secondary metabolites - with the exceptions of emodin, certain enniatins and physcion - was low, typically in the range of and below 10% of positive samples. In total, 30 different fungal and bacterial secondary metabolites were found in the samples. Some differences in the metabolite profiles were observed between countries and between index and reference school buildings. A major finding in this study was that settled dust derived from moisture damaged, damp schools contained larger numbers of microbial secondary metabolites at higher levels compared to respective dust samples from schools not affected by moisture damage and dampness. This observation was true for schools in each of the three countries, but became statistically significant only when combining schools from all countries and thus increasing the sample number in the statistical analyses.
Nazir, Nausheen; Jan, Muhammad Rasul; Ali, Amjad; Asif, Muhammad; Idrees, Muhammad; Nisar, Mohammad; Zahoor, Muhammad; Abd El-Salam, Naser M
2017-08-22
Hepatitis C virus (HCV) is a leading cause of chronic liver disease and frequently progresses towards liver cirrhosis and Hepatocellular Carcinoma (HCC). This study aimed to determine the prevalence of HCV genotypes and their association with possible transmission risks in the general population of Malakand Division. Sum of 570 serum samples were collected during March 2011 to January 2012 from suspected patients visited to different hospitals of Malakand. The suspected sera were tested using qualitative PCR and were then subjected to molecular genotype specific assay. Quantitative PCR was also performed for determination of pre-treatment viral load in confirmed positive patients. Out of 570 serum samples 316 sera were seen positive while 254 sera were found negative using qualitative PCR. The positive samples were then subjected to genotyping assay out of 316, type-specific PCR fragments were seen in 271 sera while 45 samples were found untypable genotypes. Genotype 3a was seen as a predominant genotype (63.3%) with a standard error of ±2.7%. Cramer's V statistic and Liklihood-Ratio statistical procedures are used to measure the strength and to test the association, respectively, between the dependent variable, genotype, and explanatory variables (e.g. gender, risk, age and area/districts). The dependent variable, genotype, is observed statistically significant association with variable risk factors. This implies that the genotype is highly dependent on how the patient was infected. In contrast, the other covariates, for example, gender, age, and district (area) no statistical significant association are observed. The association between gender-age indicates that the mean age of female was older by 10.5 ± 2.3 years with 95% confidence level using t-statistic. It was concluded from the present study that the predominant genotype was 3a in the infected population of Malakand. This study also highlights the high prevalence rate of untypable genotypes which an important issue of health care setup in Malakand and create complications in therapy of infected patients. Major mode of HCV transmission is multiple uses and re-uses of needles/injections. ISRCTN ISRCTN73824458. Registered: 28 September 2014.
Performance of digital RGB reflectance color extraction for plaque lesion
NASA Astrophysics Data System (ADS)
Hashim, Hadzli; Taib, Mohd Nasir; Jailani, Rozita; Sulaiman, Saadiah; Baba, Roshidah
2005-01-01
Several clinical psoriasis lesion groups are been studied for digital RGB color features extraction. Previous works have used samples size that included all the outliers lying beyond the standard deviation factors from the peak histograms. This paper described the statistical performances of the RGB model with and without removing these outliers. Plaque lesion is experimented with other types of psoriasis. The statistical tests are compared with respect to three samples size; the original 90 samples, the first size reduction by removing outliers from 2 standard deviation distances (2SD) and the second size reduction by removing outliers from 1 standard deviation distance (1SD). Quantification of data images through the normal/direct and differential of the conventional reflectance method is considered. Results performances are concluded by observing the error plots with 95% confidence interval and findings of the inference T-tests applied. The statistical tests outcomes have shown that B component for conventional differential method can be used to distinctively classify plaque from the other psoriasis groups in consistent with the error plots finding with an improvement in p-value greater than 0.5.
In vitro corrosion behaviour and microhardness of high-copper amalgams with platinum and indium.
Ilikli, B G; Aydin, A; Işimer, A; Alpaslan, G
1999-02-01
Samples prepared from Luxalloy, GS-80, Permite-C and Logic and polished after 24 h by traditional methods were stored in polypropylene tubes containing phosphate-buffered saline solutions (pH 3.5 and 6.5) and distilled water. The amounts of mercury, silver, tin, copper, zinc, platinum and indium in the test solutions were determined at the first, second, eighth, 52nd and 78th week by atomic absorption spectrometry. At the end of the eighth week the amalgam samples were removed from solutions and evaluated by Rockwell Super Scial Microhardness tester. Statistically significant low amounts of metal ions were measured for Permite-C containing indium and Logic containing platinum. The microhardness test results showed that there were statistically significant increases in the microhardness of Permite-C and Logic. As a result it was shown that the amalgam samples were affected from corrosion conditions to different degrees. Sample of the Logic group that was stored in distilled water, showed smoother surface properties than other amalgam samples containing high copper. However, it was observed that samples of Permite-C group had the smoothest surface properties.
Carlson, Jules C; Challis, Jonathan K; Hanson, Mark L; Wong, Charles S
2013-02-01
The stability of 24 chemicals, including pharmaceuticals and personal care products, and some agrochemicals on extraction media was evaluated by preloading them onto Oasis hydrophilic lipophilic balanced solid-phase extraction (SPE) cartridges and polar organic chemical integrative samplers (POCIS) followed by storage at -20°C over time. After 20 months, the average loss was 11% on POCIS, with only 2,4-dichlorophenoxyacetic acid, atrazine, chlorpyrifos, and gemfibrozil showing a statistically significant decline compared with initial concentrations. Losses on SPE cartridges were below 19%, with an average loss of 9%. In addition to laboratory spiked samples, multiple POCIS deployed in wastewater-impacted surface waters and SPE extracts of these waters were stored in their original coextracted matrix for nearly two years with minimal observed losses. Errors from typical sampling, handling, and concentration estimates from POCIS sampling rates were typically ± 15 to 30% relative standard deviation, so observed storage losses are minimal for most POCIS applications. While losses during storage on SPE cartridges for 20 months were small but statistically significant for many compounds, addition of labeled internal standards prior to freezing should correct for such losses. Thus, storage of processed water samples for analysis of polar organic pollutants is viable for archival purposes or studies for which samples cannot be analyzed in the short term. Copyright © 2012 SETAC.
The NASA/AFRL Meter Class Autonomous Telescope
NASA Technical Reports Server (NTRS)
Cowardin, H.; Lederer, S.; Buckalew, B.; Frith, J.; Hickson, P.; Glesne, T.; Anz-Meador, P.; Barker, E.; Stansbery, G.; Kervin, P.
2016-01-01
For the past decade, the NASA Orbital Debris Program Office (ODPO) has relied on using various ground-based telescopes in Chile to acquire statistical survey data as well as photometric and spectroscopic data of orbital debris in geosynchronous Earth orbit (GEO). The statistical survey data have been used to supply the Orbital Debris Engineering Model (ORDEM) v.3.0 with debris detections in GEO to better model the environment at altitudes where radar detections are limited. The data produced for the statistical survey ranged from 30 to 40 nights per year, which only accounted for 10% of the possible observing time. Data collection was restricted by ODPO resources and weather conditions. In order to improve the statistical sampling in GEO, as well as observe and sample other orbits, NASA's ODPO with support from the Air Force Research Laboratory (AFRL), has constructed a new observatory dedicated to orbital debris - the Meter Class Autonomous Telescope (MCAT) on Ascension Island. This location provides MCAT with the unique ability to access targets orbiting at an altitude of less than 1,000 km and low inclinations (< 20 deg). This orbital regime currently has little to no coverage by the U.S. Space Surveillance Network. Unlike previous ODPO optical assets, the ability to operate autonomously will allow rapid response observations of break-up events, an observing mode that was only available via radar tasking prior to MCAT's deployment. The primary goal of MCAT is to statistically characterize GEO via daily tasking files uploaded from ODPO. These tasking files define which operating mode to follow, providing the field center, rates, and/or targets to observe over the entire observing period. The system is also capable of tracking fast-moving targets in low Earth orbit (LEO), middle Earth orbit (MEO), as well as highly eccentric orbits like geostationary transfer orbits. On 25 August 2015, MCAT successfully acquired scientific first light, imaging the Bug Nebula and tracked objects in LEO, MEO, and GEO. NASA is working towards characterizing the system and thoroughly testing the integrated hardware and software control to achieve fully autonomous operations by late 2016. This paper will review the history and current status of the MCAT project, the details of the telescope system, and its five currently manifested operating modes.
Are great apes able to reason from multi-item samples to populations of food items?
Eckert, Johanna; Rakoczy, Hannes; Call, Josep
2017-10-01
Inductive learning from limited observations is a cognitive capacity of fundamental importance. In humans, it is underwritten by our intuitive statistics, the ability to draw systematic inferences from populations to randomly drawn samples and vice versa. According to recent research in cognitive development, human intuitive statistics develops early in infancy. Recent work in comparative psychology has produced first evidence for analogous cognitive capacities in great apes who flexibly drew inferences from populations to samples. In the present study, we investigated whether great apes (Pongo abelii, Pan troglodytes, Pan paniscus, Gorilla gorilla) also draw inductive inferences in the opposite direction, from samples to populations. In two experiments, apes saw an experimenter randomly drawing one multi-item sample from each of two populations of food items. The populations differed in their proportion of preferred to neutral items (24:6 vs. 6:24) but apes saw only the distribution of food items in the samples that reflected the distribution of the respective populations (e.g., 4:1 vs. 1:4). Based on this observation they were then allowed to choose between the two populations. Results show that apes seemed to make inferences from samples to populations and thus chose the population from which the more favorable (4:1) sample was drawn in Experiment 1. In this experiment, the more attractive sample not only contained proportionally but also absolutely more preferred food items than the less attractive sample. Experiment 2, however, revealed that when absolute and relative frequencies were disentangled, apes performed at chance level. Whether these limitations in apes' performance reflect true limits of cognitive competence or merely performance limitations due to accessory task demands is still an open question. © 2017 Wiley Periodicals, Inc.
Daigle, Courtney L; Siegford, Janice M
2014-03-01
Continuous observation is the most accurate way to determine animals' actual time budget and can provide a 'gold standard' representation of resource use, behavior frequency, and duration. Continuous observation is useful for capturing behaviors that are of short duration or occur infrequently. However, collecting continuous data is labor intensive and time consuming, making multiple individual or long-term data collection difficult. Six non-cage laying hens were video recorded for 15 h and behavioral data collected every 2 s were compared with data collected using scan sampling intervals of 5, 10, 15, 30, and 60 min and subsamples of 2 second observations performed for 10 min every 30 min, 15 min every 1 h, 30 min every 1.5 h, and 15 min every 2 h. Three statistical approaches were used to provide a comprehensive analysis to examine the quality of the data obtained via different sampling methods. General linear mixed models identified how the time budget from the sampling techniques differed from continuous observation. Correlation analysis identified how strongly results from the sampling techniques were associated with those from continuous observation. Regression analysis identified how well the results from the sampling techniques were associated with those from continuous observation, changes in magnitude, and whether a sampling technique had bias. Static behaviors were well represented with scan and time sampling techniques, while dynamic behaviors were best represented with time sampling techniques. Methods for identifying an appropriate sampling strategy based upon the type of behavior of interest are outlined and results for non-caged laying hens are presented. Copyright © 2013 Elsevier B.V. All rights reserved.
NASA Technical Reports Server (NTRS)
Shaw, R. L.
1979-01-01
A sample of 228 supernovae that occurred in galaxies with known redshifts is used to show that the mean projected linear supernova distance from the center of the parent galaxy increases with increasing redshift. This effect is interpreted as an observational bias: the discovery rate of supernovae is reduced in the inner parts of distant, poorly resolved galaxies. Even under the optimistic assumption that no selection effects work in galaxies closer than 33 Mpc, about 50% of all supernovae are lost in the inner regions of galaxies beyond 150 Mpc. This observational bias must be taken into account in the derivation of statistical properties of supernovae.
Overall voice and strain level analysis in rock singers.
Gonsalves, Aline; Amin, Elisabeth; Behlau, Mara
2010-01-01
overall voice and strain level analysis in rock singers. to analyze the voice o rock singers according to two specific parameters: overall level of vocal deviation (OLVD) and strain level (SL); to compare these parameters in three different music samples. participants were 26 male rock singers, ranging in age from 17 to 46 years (mean = 29.8 years). All of the participants answered a questionnaire for sample characterization and were submitted to the recording of three voice samples: Brazilian National Anthem (BNA), Satisfaction and self-selected repertoire song (RS). Voice samples were analyzed by five speech-language pathologists according to OLVD and SL. Statistical analysis was done using the software SPSS, version 13.0. statistically significant differences were observed for the mean values of OLVD and SL during the performance of Satisfaction (OLVD = 32.8 and SL = 0.024 / p=0.024) and during the RS performance (OLVD = 38.4 and SL = 55.8 / p=0.010). The values of OLVD and SL are directly proportional to the samples of the BNA* and RS**, i.e. the higher the strain the higher the OLVD (p,0.001*; p=0.010**). When individually analyzing the three song samples, it is observed that the OLVD does not vary significantly among them. However, the mean values present a trend to increase from non-rock to rock performances (24.0 BNA / 32.8 Satisfaction / 38.4 RS). The level of strain found during the BNA performance presents statistically significant difference when compared to the rock performances (Satisfaction and RS, p=0.008 and p=0.001). the obtained data suggest that rock style is related to the greater use of vocal strain and that this strain does not necessarily impose a negative impression to the voice, but corresponds to a common interpretative factor related to this style of music.
No association of dynamin binding protein (DNMBP) gene SNPs and Alzheimer's disease.
Minster, Ryan L; DeKosky, Steven T; Kamboh, M Ilyas
2008-10-01
A recent scan of single nucleotide polymorphisms (SNPs) on chromosome 10q found significant association of six correlated SNPs with late-onset Alzheimer's disease (AD) among Japanese. We examined the SNP with the highest statistical significance (rs3740058) in a large Caucasian American case-control cohort and the remaining five SNPs in a smaller subset of cases and controls. We observed no association of statistical significance in either the total sample or the APOE*4 non-carriers for any of the SNPs.
Lane, James D; Bigelow, Philip L; Majowicz, Shannon E; McColl, R Stephen
2016-07-01
The objectives of this study were to determine whether grid-connected industrial wind turbines (IWTs) are a risk factor for poor sleep quality, and if IWT noise is associated with sleep parameters in rural Ontarians. A daily sleep diary and actigraphy-derived measures of sleep were obtained from 12 participants from an IWT community and 10 participants from a comparison community with no wind power installations. The equivalent and maximum sound pressure levels within the bedroom were also assessed. No statistically significant differences were observed between IWT residents and non-IWT residents for any of the parameters measured in this study. Actigraphy and sleep diaries are feasible tools to understand the impact of IWTs on the quality of sleep for nearby residents. Further studies with larger sample sizes should be conducted to determine whether the lack of statistical significance observed here is a result of sample size, or reflects a true lack of association.
Measurement of air contamination in different wards of public sector hospital, Sukkur.
Memon, Badaruddin AllahDino; Bhutto, Gul Hassan; Rizvi, Wajid Hussain
2016-11-01
The aim of this study was to evaluate and assess the index of bacterial contamination in different wards of the Public Sector Hospital of Sukkur (Teaching) Pakistan; whether or not the air contamination was statistically different from the acceptable level using active and passive sampling. In addition to this main hypothesis, other investigations included: occurrence of the most common bacteria, whether or not the bacterial contamination in the wards was a persistent problem and identification of the effective antibiotics against the indentified bacteria. The evidence sought based on the One Sample T test suggests that there is a (statistically) significant difference between the observed (higher) than the acceptance level (p<0.01), the result based on One-Way ANOVA suggests that the contamination problem was persistent as there was no significant difference among observed contamination of all three visits at (p>0.01) and the result of antibiotic susceptibility test highlights sensitivity and resistance level of antibiotics for the indentified bacteria.
Bacci, Silvia; Seracini, Marco; Chiavarini, Manuela; Bartolucci, Francesco; Minelli, Liliana
2017-01-01
The aim of this study was to investigate the relationship between employment status (permanent employment, fixed-term employment, unemployment, other) and perceived health status in a sample of the Italian population. Data was obtained from the European Union Statistics on Income and Living Condition (EU-SILC) study during the period 2009 - 2012. The sample consists of 4,848 individuals, each with a complete record of observations during four years for a total of 19,392 observations. The causal relationship between perceived/self-reported health status and employment status was tested using a global logit model (STATA). Our results confirm a significant association between employment status and perceived health, as well as between perceived health status and economic status. Unemployment that was dependent on an actual lack of work opportunities and not from individual disability was found to be the most significant determinant of perceived health status; a higher educational level produces a better perceived health status.
Crans, Gerald G; Shuster, Jonathan J
2008-08-15
The debate as to which statistical methodology is most appropriate for the analysis of the two-sample comparative binomial trial has persisted for decades. Practitioners who favor the conditional methods of Fisher, Fisher's exact test (FET), claim that only experimental outcomes containing the same amount of information should be considered when performing analyses. Hence, the total number of successes should be fixed at its observed level in hypothetical repetitions of the experiment. Using conditional methods in clinical settings can pose interpretation difficulties, since results are derived using conditional sample spaces rather than the set of all possible outcomes. Perhaps more importantly from a clinical trial design perspective, this test can be too conservative, resulting in greater resource requirements and more subjects exposed to an experimental treatment. The actual significance level attained by FET (the size of the test) has not been reported in the statistical literature. Berger (J. R. Statist. Soc. D (The Statistician) 2001; 50:79-85) proposed assessing the conservativeness of conditional methods using p-value confidence intervals. In this paper we develop a numerical algorithm that calculates the size of FET for sample sizes, n, up to 125 per group at the two-sided significance level, alpha = 0.05. Additionally, this numerical method is used to define new significance levels alpha(*) = alpha+epsilon, where epsilon is a small positive number, for each n, such that the size of the test is as close as possible to the pre-specified alpha (0.05 for the current work) without exceeding it. Lastly, a sample size and power calculation example are presented, which demonstrates the statistical advantages of implementing the adjustment to FET (using alpha(*) instead of alpha) in the two-sample comparative binomial trial. 2008 John Wiley & Sons, Ltd
Wellek, Stefan
2017-02-28
In current practice, the most frequently applied approach to the handling of ties in the Mann-Whitney-Wilcoxon (MWW) test is based on the conditional distribution of the sum of mid-ranks, given the observed pattern of ties. Starting from this conditional version of the testing procedure, a sample size formula was derived and investigated by Zhao et al. (Stat Med 2008). In contrast, the approach we pursue here is a nonconditional one exploiting explicit representations for the variances of and the covariance between the two U-statistics estimators involved in the Mann-Whitney form of the test statistic. The accuracy of both ways of approximating the sample sizes required for attaining a prespecified level of power in the MWW test for superiority with arbitrarily tied data is comparatively evaluated by means of simulation. The key qualitative conclusions to be drawn from these numerical comparisons are as follows: With the sample sizes calculated by means of the respective formula, both versions of the test maintain the level and the prespecified power with about the same degree of accuracy. Despite the equivalence in terms of accuracy, the sample size estimates obtained by means of the new formula are in many cases markedly lower than that calculated for the conditional test. Perhaps, a still more important advantage of the nonconditional approach based on U-statistics is that it can be also adopted for noninferiority trials. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
Joint Adaptive Mean-Variance Regularization and Variance Stabilization of High Dimensional Data.
Dazard, Jean-Eudes; Rao, J Sunil
2012-07-01
The paper addresses a common problem in the analysis of high-dimensional high-throughput "omics" data, which is parameter estimation across multiple variables in a set of data where the number of variables is much larger than the sample size. Among the problems posed by this type of data are that variable-specific estimators of variances are not reliable and variable-wise tests statistics have low power, both due to a lack of degrees of freedom. In addition, it has been observed in this type of data that the variance increases as a function of the mean. We introduce a non-parametric adaptive regularization procedure that is innovative in that : (i) it employs a novel "similarity statistic"-based clustering technique to generate local-pooled or regularized shrinkage estimators of population parameters, (ii) the regularization is done jointly on population moments, benefiting from C. Stein's result on inadmissibility, which implies that usual sample variance estimator is improved by a shrinkage estimator using information contained in the sample mean. From these joint regularized shrinkage estimators, we derived regularized t-like statistics and show in simulation studies that they offer more statistical power in hypothesis testing than their standard sample counterparts, or regular common value-shrinkage estimators, or when the information contained in the sample mean is simply ignored. Finally, we show that these estimators feature interesting properties of variance stabilization and normalization that can be used for preprocessing high-dimensional multivariate data. The method is available as an R package, called 'MVR' ('Mean-Variance Regularization'), downloadable from the CRAN website.
Accelerating assimilation development for new observing systems using EFSO
NASA Astrophysics Data System (ADS)
Lien, Guo-Yuan; Hotta, Daisuke; Kalnay, Eugenia; Miyoshi, Takemasa; Chen, Tse-Chun
2018-03-01
To successfully assimilate data from a new observing system, it is necessary to develop appropriate data selection strategies, assimilating only the generally useful data. This development work is usually done by trial and error using observing system experiments (OSEs), which are very time and resource consuming. This study proposes a new, efficient methodology to accelerate the development using ensemble forecast sensitivity to observations (EFSO). First, non-cycled assimilation of the new observation data is conducted to compute EFSO diagnostics for each observation within a large sample. Second, the average EFSO conditionally sampled in terms of various factors is computed. Third, potential data selection criteria are designed based on the non-cycled EFSO statistics, and tested in cycled OSEs to verify the actual assimilation impact. The usefulness of this method is demonstrated with the assimilation of satellite precipitation data. It is shown that the EFSO-based method can efficiently suggest data selection criteria that significantly improve the assimilation results.
NASA Technical Reports Server (NTRS)
Bavassano, B.; Dobrowolny, H.; Fanfoni, G.; Mariani, F.; Ness, N. F.
1981-01-01
Helios 2 magnetic data were used to obtain several statistical properties of MHD fluctuations associated with the trailing edge of a given stream served in different solar rotations. Eigenvalues and eigenvectors of the variance matrix, total power and degree of compressibility of the fluctuations were derived and discussed both as a function of distance from the Sun and as a function of the frequency range included in the sample. The results obtained add new information to the picture of MHD turbulence in the solar wind. In particular, a dependence from frequency range of the radial gradients of various statistical quantities is obtained.
Probing the statistics of primordial fluctuations and their evolution
NASA Technical Reports Server (NTRS)
Gaztanaga, Enrique; Yokoyama, Jun'ichi
1993-01-01
The statistical distribution of fluctuations on various scales is analyzed in terms of the counts in cells of smoothed density fields, using volume-limited samples of galaxy redshift catalogs. It is shown that the distribution on large scales, with volume average of the two-point correlation function of the smoothed field less than about 0.05, is consistent with Gaussian. Statistics are shown to agree remarkably well with the negative binomial distribution, which has hierarchial correlations and a Gaussian behavior at large scales. If these observed properties correspond to the matter distribution, they suggest that our universe started with Gaussian fluctuations and evolved keeping hierarchial form.
Xu, Xianglong; Liu, Lingli; Sharma, Manoj; Zhao, Yong
2015-01-01
Introduction: In 2012 in China, 52.9% of men were reported to smoke while only 2.4% of women smoked. This study explored the smoking-related Knowledge, Attitudes and Practices (KAP) among young adult male smokers. Methods: A cross-sectional study was conducted in four municipal areas of Chongqing using a questionnaire administered to 536 natives young male smokers aged 18–45 years old. Results: The total score of smoking cognition, the total score of smoking attitude and the total score of positive behavior to quit smoking was significantly different among the three groups by education. Besides, 30.97% of male smokers never seriously thought about quitting smoking. Logistic regression analysis found smoking-related knowledge, attitudes, behaviors and sociodemographic factors affect having smoking cessation idea. But no statistically significant correlation was observed between smoking cognition and positive behavior to quit smoking in a sample of higher education. No statistically significant correlation was observed between smoking cognition and positive behavior to quit smoking (Pearson correlation coefficient = 0.03012, p = 0.6811), and also no statistically significant correlation was observed between smoking cognition and positive behavior to quit smoking (Pearson correlation coefficient = 0.08869, p = 0.2364) in the sample of higher education young adult males Conclusions: Young adult males with higher education have a better knowledge of smoking hazards and a more positive attitude toward smoking, however, this knowledge and attitude do not necessarily translate into health behavioral outcomes such as not smoking. Overall the present findings indicate that no statistically significant correlation between the education level and quitting smoking idea exists among young adult male smokers in China. This survey gives a snapshot of the impact of education on smoking-related KAP among young adults male smokers. PMID:25689992
Xu, Xianglong; Liu, Lingli; Sharma, Manoj; Zhao, Yong
2015-02-16
In 2012 in China, 52.9% of men were reported to smoke while only 2.4% of women smoked. This study explored the smoking-related Knowledge, Attitudes and Practices (KAP) among young adult male smokers. A cross-sectional study was conducted in four municipal areas of Chongqing using a questionnaire administered to 536 natives young male smokers aged 18-45 years old. The total score of smoking cognition, the total score of smoking attitude and the total score of positive behavior to quit smoking was significantly different among the three groups by education. Besides, 30.97% of male smokers never seriously thought about quitting smoking. Logistic regression analysis found smoking-related knowledge, attitudes, behaviors and sociodemographic factors affect having smoking cessation idea. But no statistically significant correlation was observed between smoking cognition and positive behavior to quit smoking in a sample of higher education. No statistically significant correlation was observed between smoking cognition and positive behavior to quit smoking (Pearson correlation coefficient = 0.03012, p = 0.6811), and also no statistically significant correlation was observed between smoking cognition and positive behavior to quit smoking (Pearson correlation coefficient = 0.08869, p = 0.2364) in the sample of higher education young adult males Young adult males with higher education have a better knowledge of smoking hazards and a more positive attitude toward smoking, however, this knowledge and attitude do not necessarily translate into health behavioral outcomes such as not smoking. Overall the present findings indicate that no statistically significant correlation between the education level and quitting smoking idea exists among young adult male smokers in China. This survey gives a snapshot of the impact of education on smoking-related KAP among young adults male smokers.
Perugini, Monia; Visciano, Pierina; Manera, Maurizio; Abete, Maria Cesarina; Gavinelli, Stefania; Amorena, Michele
2013-11-01
The aim of this study was to evaluate mercury and selenium distribution in different portions (exoskeleton, white meat and brown meat) of Norway lobster (Nephrops norvegicus). Some samples were also analysed as whole specimens. The same portions were also examined after boiling, in order to observe if this cooking practice could affect mercury and selenium concentrations. The highest mercury concentrations were detected in white meat, exceeding in all cases the maximum levels established by European legislation. The brown meat reported the highest selenium concentrations. In all boiled samples, mercury levels showed a statistically significant increase compared to raw portions. On the contrary, selenium concentrations detected in boiled samples of white meat, brown meat and whole specimen showed a statistically significant decrease compared to the corresponding raw samples. These results indicate that boiling modifies mercury and selenium concentrations. The high mercury levels detected represent a possible risk for consumers, and the publication and diffusion of specific advisories concerning seafood consumption is recommended.
Lockwood, S F; Bickham, J W
1991-01-01
Intraspecific variation in cellular DNA content was measured in five Coregonus autumnalis spawning populations from the Mackenzie River drainage, Canada, using flow cytometry. The rivers assayed were the Peel, Arctic Red, Mountain, Carcajou, and Liard rivers. DNA content was determined from whole blood preparations of fish from all rivers except the Carcajou, for which kidney tissue was used. DNA content measurements of kidney and blood preparations of the same fish from the Mountain River revealed statistically indistinguishable results. Mosaicism was found in blood preparations from the Peel, Arctic Red, Mountain, and Liard rivers, but was not observed in kidney tissue preparations from the Mountain or Carcajou rivers. The Liard River sample had significantly elevated mean DNA content relative to the other four samples; all other samples were statistically indistinguishable. Significant differences in mean DNA content among spawning stocks of a single species reinforces the need for adequate sample sizes of both individuals and populations when reporting "C" values for a particular species.
Teo, Guoshou; Kim, Sinae; Tsou, Chih-Chiang; Collins, Ben; Gingras, Anne-Claude; Nesvizhskii, Alexey I; Choi, Hyungwon
2015-11-03
Data independent acquisition (DIA) mass spectrometry is an emerging technique that offers more complete detection and quantification of peptides and proteins across multiple samples. DIA allows fragment-level quantification, which can be considered as repeated measurements of the abundance of the corresponding peptides and proteins in the downstream statistical analysis. However, few statistical approaches are available for aggregating these complex fragment-level data into peptide- or protein-level statistical summaries. In this work, we describe a software package, mapDIA, for statistical analysis of differential protein expression using DIA fragment-level intensities. The workflow consists of three major steps: intensity normalization, peptide/fragment selection, and statistical analysis. First, mapDIA offers normalization of fragment-level intensities by total intensity sums as well as a novel alternative normalization by local intensity sums in retention time space. Second, mapDIA removes outlier observations and selects peptides/fragments that preserve the major quantitative patterns across all samples for each protein. Last, using the selected fragments and peptides, mapDIA performs model-based statistical significance analysis of protein-level differential expression between specified groups of samples. Using a comprehensive set of simulation datasets, we show that mapDIA detects differentially expressed proteins with accurate control of the false discovery rates. We also describe the analysis procedure in detail using two recently published DIA datasets generated for 14-3-3β dynamic interaction network and prostate cancer glycoproteome. The software was written in C++ language and the source code is available for free through SourceForge website http://sourceforge.net/projects/mapdia/.This article is part of a Special Issue entitled: Computational Proteomics. Copyright © 2015 Elsevier B.V. All rights reserved.
Rapp, J.B.
1991-01-01
Q-mode factor analysis was used to quantitate the distribution of the major aliphatic hydrocarbon (n-alkanes, pristane, phytane) systems in sediments from a variety of marine environments. The compositions of the pure end members of the systems were obtained from factor scores and the distribution of the systems within each sample was obtained from factor loadings. All the data, from the diverse environments sampled (estuarine (San Francisco Bay), fresh-water (San Francisco Peninsula), polar-marine (Antarctica) and geothermal-marine (Gorda Ridge) sediments), were reduced to three major systems: a terrestrial system (mostly high molecular weight aliphatics with odd-numbered-carbon predominance), a mature system (mostly low molecular weight aliphatics without predominance) and a system containing mostly high molecular weight aliphatics with even-numbered-carbon predominance. With this statistical approach, it is possible to assign the percentage contribution from various sources to the observed distribution of aliphatic hydrocarbons in each sediment sample. ?? 1991.
Le Quellec, Sandra; Paris, Mickaël; Nougier, Christophe; Sobas, Frédéric; Rugeri, Lucia; Girard, Sandrine; Bordet, Jean-Claude; Négrier, Claude; Dargaud, Yesim
2017-05-01
Pneumatic tube system (PTS) in hospitals is commonly used for the transport of blood samples to clinical laboratories, as it is rapid and cost-effective. The aim was to compare the effects on haematology samples of a newly acquired ~2km-long PTS that links 2 hospitals with usual transport (non-pneumatic tube system, NPTS). Complete blood cell count, routine coagulation assays, platelet function tests (PFT) with light-transmission aggregometry and global coagulation assays including ROTEM® and thrombin generation assay (TGA) were performed on blood samples from 30 healthy volunteers and 9 healthy volunteers who agreed to take aspirin prior to blood sampling. The turnaround time was reduced by 31% (p<0.001) with the use of PTS. No statistically significant difference was observed for most routine haematology assays including PFT, and ROTEM® analysis. A statistically significant, but not clinically relevant, shortening of the APTT after sample transport by PTS was found (mean±SD: 30s±1.8 vs. 29.5s±2.1 for NPTS). D-dimer levels were 7.4% higher after transport through PTS but were not discordant. A statistically significant increase of thrombin generation was found in both platelet poor- and platelet rich- plasma samples after PTS transport compared to NPTS transport. PTS is suitable for the transport of samples prior to routine haematology assays including PFT, but should not be used for samples intended for thrombin generation measurement. Copyright © 2017 Elsevier Ltd. All rights reserved.
Development of Probabilistic Understanding in Fourth Grade
ERIC Educational Resources Information Center
English, Lyn D.; Watson, Jane M.
2016-01-01
The authors analyzed the development of 4th-grade students' understanding of the transition from experimental relative frequencies of outcomes to theoretical probabilities with a focus on the foundational statistical concepts of variation and expectation. After observing the decreasing variation from the theoretical probability as the sample size…
29 CFR Section 1607.16 - Definitions.
Code of Federal Regulations, 2010 CFR
2010-07-01
... action are open to users. T. Skill. A present, observable competence to perform a learned psychomoter act... criterion-related validity studies. These conditions include: (1) An adequate sample of persons available for the study to achieve findings of statistical significance; (2) having or being able to obtain a...
Climate Verification Using Running Mann Whitney Z Statistics
USDA-ARS?s Scientific Manuscript database
A robust method previously used to detect observed intra- to multi-decadal (IMD) climate regimes was adapted to test whether climate models could reproduce IMD variations in U.S. surface temperatures during 1919-2008. This procedure, called the running Mann Whitney Z (MWZ) method, samples data ranki...
The relation between statistical power and inference in fMRI
Wager, Tor D.; Yarkoni, Tal
2017-01-01
Statistically underpowered studies can result in experimental failure even when all other experimental considerations have been addressed impeccably. In fMRI the combination of a large number of dependent variables, a relatively small number of observations (subjects), and a need to correct for multiple comparisons can decrease statistical power dramatically. This problem has been clearly addressed yet remains controversial—especially in regards to the expected effect sizes in fMRI, and especially for between-subjects effects such as group comparisons and brain-behavior correlations. We aimed to clarify the power problem by considering and contrasting two simulated scenarios of such possible brain-behavior correlations: weak diffuse effects and strong localized effects. Sampling from these scenarios shows that, particularly in the weak diffuse scenario, common sample sizes (n = 20–30) display extremely low statistical power, poorly represent the actual effects in the full sample, and show large variation on subsequent replications. Empirical data from the Human Connectome Project resembles the weak diffuse scenario much more than the localized strong scenario, which underscores the extent of the power problem for many studies. Possible solutions to the power problem include increasing the sample size, using less stringent thresholds, or focusing on a region-of-interest. However, these approaches are not always feasible and some have major drawbacks. The most prominent solutions that may help address the power problem include model-based (multivariate) prediction methods and meta-analyses with related synthesis-oriented approaches. PMID:29155843
NASA Astrophysics Data System (ADS)
Price-Whelan, Adrian M.; Agüeros, Marcel A.; Fournier, Amanda P.; Street, Rachel; Ofek, Eran O.; Covey, Kevin R.; Levitan, David; Laher, Russ R.; Sesar, Branimir; Surace, Jason
2014-01-01
Many photometric time-domain surveys are driven by specific goals, such as searches for supernovae or transiting exoplanets, which set the cadence with which fields are re-imaged. In the case of the Palomar Transient Factory (PTF), several sub-surveys are conducted in parallel, leading to non-uniform sampling over its ~20,000 deg2 footprint. While the median 7.26 deg2 PTF field has been imaged ~40 times in the R band, ~2300 deg2 have been observed >100 times. We use PTF data to study the trade off between searching for microlensing events in a survey whose footprint is much larger than that of typical microlensing searches, but with far-from-optimal time sampling. To examine the probability that microlensing events can be recovered in these data, we test statistics used on uniformly sampled data to identify variables and transients. We find that the von Neumann ratio performs best for identifying simulated microlensing events in our data. We develop a selection method using this statistic and apply it to data from fields with >10 R-band observations, 1.1 × 109 light curves, uncovering three candidate microlensing events. We lack simultaneous, multi-color photometry to confirm these as microlensing events. However, their number is consistent with predictions for the event rate in the PTF footprint over the survey's three years of operations, as estimated from near-field microlensing models. This work can help constrain all-sky event rate predictions and tests microlensing signal recovery in large data sets, which will be useful to future time-domain surveys, such as that planned with the Large Synoptic Survey Telescope.
Acidity of fine sulfate particles at Great Smokey Mountains National Park
DOE Office of Scientific and Technical Information (OSTI.GOV)
Day, D.; Malm, W.C.; Kreidenweis, S.
1995-12-31
The acidity of ambient particles is of interest from the perspectives of human health, visibility, and ecology. This paper reports on the acidity of fine (< 2.5{mu}m) particles measured during August 1994 at Look Rock observation tower in Great Smokey Mountains National Park. This site is located at latitude 35{degrees} 37 feet 56 inches, longitude 83{degrees} 56 feet 32 inches, and at an elevation of 808m above sea level. All samples were collected using the IMPROVE (Interagency Monitoring of Protected Visual Environments) sampler. The sampling periods included: (1) 4-hour samples collected three times daily with starting times of 8:00 AM,more » 12:00 noon, and 4:00 PM; (2) 12-hour samples collected twice daily with starting times of 8:00 AM and 8:00 PM (all times reported are eastern daylight savings time). The IMPROVE sampler, collecting 4-hour samples, employed a citric acid/glycerol coated annular denuder to remove ammonia gas while the 12-hour sampler did not use a citric acid denuder. The intensive monitoring effort, conducted during August 1994, showed that: (1) the fine aerosol mass is generally dominated by sulfate and its associated water; (2) there was no statistically significant difference in average sulfate concentration between the 12-hour samples nor was there a statistically significant difference in average sulfate concentration between the 4-hour samples; (3) the aerosol is highly acidic, ranging from almost pure sulfuric acid to pure ammonium bisulfate, with an average molar ammonium ion to sulfate ratio of about 0.75 which suggests the ambient sulfate aerosol was a mixture of ammonium bisulfate and sulfuric acid; and (4) there was no statistically significant diurnal variation in particle acidity nor was there a statistically significant difference in particle acidity between the 4 hour samples.« less
ACCESS: The Arizona-CfA-Catolica Exoplanet Spectroscopy Survey
NASA Astrophysics Data System (ADS)
Lopez-Morales, Mercedes; Apai, Daniel; Jordan, Andres; Espinoza, Nestor; Rackham, Benjamin; Fraine, Jonathan D.; Rodler, Florian; Lewis, Nikole; Fortney, Jonathan J.; Osip, David J.
2014-06-01
The Arizona-CfA-Catolica Exoplanet Spectroscopy Survey (ACCESS) is an international, multi-institutional consortium with members from the Harvard-Smithsonian CfA, the University of Arizona, Pontificia Universidad Catolica in Chile, MIT and UC Santa Cruz and the Carnegie Institution. ACCESS' goal is to observe about two dozen planets covering a wide range of mass, radius, atmospheric temperatures and energy irradiation levels, with two main scientific goals: 1) to obtain, for the first time, a uniform sample of visible transmission spectra of exoplanets, allowing the study of their atmospheric characteristics as a statistically significant sample, and 2) to mature the technique of ground-based observations of exoplanetary atmospheres for future observations of small planets. Here we describe ACCESS and its first science results.
Farnsworth, G.L.; Nichols, J.D.; Sauer, J.R.; Fancy, S.G.; Pollock, K.H.; Shriner, S.A.; Simons, T.R.; Ralph, C. John; Rich, Terrell D.
2005-01-01
Point counts are a standard sampling procedure for many bird species, but lingering concerns still exist about the quality of information produced from the method. It is well known that variation in observer ability and environmental conditions can influence the detection probability of birds in point counts, but many biologists have been reluctant to abandon point counts in favor of more intensive approaches to counting. However, over the past few years a variety of statistical and methodological developments have begun to provide practical ways of overcoming some of the problems with point counts. We describe some of these approaches, and show how they can be integrated into standard point count protocols to greatly enhance the quality of the information. Several tools now exist for estimation of detection probability of birds during counts, including distance sampling, double observer methods, time-depletion (removal) methods, and hybrid methods that combine these approaches. Many counts are conducted in habitats that make auditory detection of birds much more likely than visual detection. As a framework for understanding detection probability during such counts, we propose separating two components of the probability a bird is detected during a count into (1) the probability a bird vocalizes during the count and (2) the probability this vocalization is detected by an observer. In addition, we propose that some measure of the area sampled during a count is necessary for valid inferences about bird populations. This can be done by employing fixed-radius counts or more sophisticated distance-sampling models. We recommend any studies employing point counts be designed to estimate detection probability and to include a measure of the area sampled.
Moyle, Phillip R.; Causey, J. Douglas
2001-01-01
This report provides chemical analyses for 31 samples collected from various phosphate mine sites in southeastern Idaho (25), northern Utah (2), and western Wyoming (4). The sampling effort was undertaken as a reconnaissance and does not constitute a characterization of mine wastes. Twenty-five samples were collected from waste rock dumps, 2 from stockpiles, and 1 each from slag, tailings, mill shale, and an outcrop. All samples were analyzed for a suite of major, minor, and trace elements. Although the analytical data set for the 31 samples is too small for detailed statistical analysis, a summary of general observations is made.
Induced earthquake magnitudes are as large as (statistically) expected
Van Der Elst, Nicholas; Page, Morgan T.; Weiser, Deborah A.; Goebel, Thomas; Hosseini, S. Mehran
2016-01-01
A major question for the hazard posed by injection-induced seismicity is how large induced earthquakes can be. Are their maximum magnitudes determined by injection parameters or by tectonics? Deterministic limits on induced earthquake magnitudes have been proposed based on the size of the reservoir or the volume of fluid injected. However, if induced earthquakes occur on tectonic faults oriented favorably with respect to the tectonic stress field, then they may be limited only by the regional tectonics and connectivity of the fault network. In this study, we show that the largest magnitudes observed at fluid injection sites are consistent with the sampling statistics of the Gutenberg-Richter distribution for tectonic earthquakes, assuming no upper magnitude bound. The data pass three specific tests: (1) the largest observed earthquake at each site scales with the log of the total number of induced earthquakes, (2) the order of occurrence of the largest event is random within the induced sequence, and (3) the injected volume controls the total number of earthquakes rather than the total seismic moment. All three tests point to an injection control on earthquake nucleation but a tectonic control on earthquake magnitude. Given that the largest observed earthquakes are exactly as large as expected from the sampling statistics, we should not conclude that these are the largest earthquakes possible. Instead, the results imply that induced earthquake magnitudes should be treated with the same maximum magnitude bound that is currently used to treat seismic hazard from tectonic earthquakes.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Callister, Stephen J.; Barry, Richard C.; Adkins, Joshua N.
2006-02-01
Central tendency, linear regression, locally weighted regression, and quantile techniques were investigated for normalization of peptide abundance measurements obtained from high-throughput liquid chromatography-Fourier transform ion cyclotron resonance mass spectrometry (LC-FTICR MS). Arbitrary abundances of peptides were obtained from three sample sets, including a standard protein sample, two Deinococcus radiodurans samples taken from different growth phases, and two mouse striatum samples from control and methamphetamine-stressed mice (strain C57BL/6). The selected normalization techniques were evaluated in both the absence and presence of biological variability by estimating extraneous variability prior to and following normalization. Prior to normalization, replicate runs from each sample setmore » were observed to be statistically different, while following normalization replicate runs were no longer statistically different. Although all techniques reduced systematic bias, assigned ranks among the techniques revealed significant trends. For most LC-FTICR MS analyses, linear regression normalization ranked either first or second among the four techniques, suggesting that this technique was more generally suitable for reducing systematic biases.« less
Chung, Sang M; Lee, David J; Hand, Austin; Young, Philip; Vaidyanathan, Jayabharathi; Sahajwalla, Chandrahas
2015-12-01
The study evaluated whether the renal function decline rate per year with age in adults varies based on two primary statistical analyses: cross-section (CS), using one observation per subject, and longitudinal (LT), using multiple observations per subject over time. A total of 16628 records (3946 subjects; age range 30-92 years) of creatinine clearance and relevant demographic data were used. On average, four samples per subject were collected for up to 2364 days (mean: 793 days). A simple linear regression and random coefficient models were selected for CS and LT analyses, respectively. The renal function decline rates per year were 1.33 and 0.95 ml/min/year for CS and LT analyses, respectively, and were slower when the repeated individual measurements were considered. The study confirms that rates are different based on statistical analyses, and that a statistically robust longitudinal model with a proper sampling design provides reliable individual as well as population estimates of the renal function decline rates per year with age in adults. In conclusion, our findings indicated that one should be cautious in interpreting the renal function decline rate with aging information because its estimation was highly dependent on the statistical analyses. From our analyses, a population longitudinal analysis (e.g. random coefficient model) is recommended if individualization is critical, such as a dose adjustment based on renal function during a chronic therapy. Copyright © 2015 John Wiley & Sons, Ltd.
Statistical Analysis of Large Scale Structure by the Discrete Wavelet Transform
NASA Astrophysics Data System (ADS)
Pando, Jesus
1997-10-01
The discrete wavelet transform (DWT) is developed as a general statistical tool for the study of large scale structures (LSS) in astrophysics. The DWT is used in all aspects of structure identification including cluster analysis, spectrum and two-point correlation studies, scale-scale correlation analysis and to measure deviations from Gaussian behavior. The techniques developed are demonstrated on 'academic' signals, on simulated models of the Lymanα (Lyα) forests, and on observational data of the Lyα forests. This technique can detect clustering in the Ly-α clouds where traditional techniques such as the two-point correlation function have failed. The position and strength of these clusters in both real and simulated data is determined and it is shown that clusters exist on scales as large as at least 20 h-1 Mpc at significance levels of 2-4 σ. Furthermore, it is found that the strength distribution of the clusters can be used to distinguish between real data and simulated samples even where other traditional methods have failed to detect differences. Second, a method for measuring the power spectrum of a density field using the DWT is developed. All common features determined by the usual Fourier power spectrum can be calculated by the DWT. These features, such as the index of a power law or typical scales, can be detected even when the samples are geometrically complex, the samples are incomplete, or the mean density on larger scales is not known (the infrared uncertainty). Using this method the spectra of Ly-α forests in both simulated and real samples is calculated. Third, a method for measuring hierarchical clustering is introduced. Because hierarchical evolution is characterized by a set of rules of how larger dark matter halos are formed by the merging of smaller halos, scale-scale correlations of the density field should be one of the most sensitive quantities in determining the merging history. We show that these correlations can be completely determined by the correlations between discrete wavelet coefficients on adjacent scales and at nearly the same spatial position, Cj,j+12/cdot2. Scale-scale correlations on two samples of the QSO Ly-α forests absorption spectra are computed. Lastly, higher order statistics are developed to detect deviations from Gaussian behavior. These higher order statistics are necessary to fully characterize the Ly-α forests because the usual 2nd order statistics, such as the two-point correlation function or power spectrum, give inconclusive results. It is shown how this technique takes advantage of the locality of the DWT to circumvent the central limit theorem. A non-Gaussian spectrum is defined and this spectrum reveals not only the magnitude, but the scales of non-Gaussianity. When applied to simulated and observational samples of the Ly-α clouds, it is found that different popular models of structure formation have different spectra while two, independent observational data sets, have the same spectra. Moreover, the non-Gaussian spectra of real data sets are significantly different from the spectra of various possible random samples. (Abstract shortened by UMI.)
Analyzing Dyadic Sequence Data—Research Questions and Implied Statistical Models
Fuchs, Peter; Nussbeck, Fridtjof W.; Meuwly, Nathalie; Bodenmann, Guy
2017-01-01
The analysis of observational data is often seen as a key approach to understanding dynamics in romantic relationships but also in dyadic systems in general. Statistical models for the analysis of dyadic observational data are not commonly known or applied. In this contribution, selected approaches to dyadic sequence data will be presented with a focus on models that can be applied when sample sizes are of medium size (N = 100 couples or less). Each of the statistical models is motivated by an underlying potential research question, the most important model results are presented and linked to the research question. The following research questions and models are compared with respect to their applicability using a hands on approach: (I) Is there an association between a particular behavior by one and the reaction by the other partner? (Pearson Correlation); (II) Does the behavior of one member trigger an immediate reaction by the other? (aggregated logit models; multi-level approach; basic Markov model); (III) Is there an underlying dyadic process, which might account for the observed behavior? (hidden Markov model); and (IV) Are there latent groups of dyads, which might account for observing different reaction patterns? (mixture Markov; optimal matching). Finally, recommendations for researchers to choose among the different models, issues of data handling, and advises to apply the statistical models in empirical research properly are given (e.g., in a new r-package “DySeq”). PMID:28443037
NASA Astrophysics Data System (ADS)
Xue, J.; Sherchan, S. P.; Lamar, F. G.; Lin, S.; Lamori, J. G.
2017-12-01
Brackish water samples from Lake Pontchartrain in Louisiana were assessed for the presence of pathogenic amoeba Naegleria fowleri, which causes primary amoebic meningoencephalitis (PAM). In our study, quantitative polymerase chain reaction (qPCR) methods were used to determine N. fowleri, E. coli, and Enterococcus in water collected from Lake Pontchartrain. A total of 158 water samples were analyzed over the 10- month sampling period. Statistically significant positive correlation between water temperature and N. fowleri concentration was observed. N. fowleri target sequence was detected at 35.4% (56/158) of the water samples from ten sites around the Lake ranged from 11.6 GC/100 ml water to 457.8 GC/100 ml water. A single factor (ANOVA) analysis shows the average concentration of N. fowleri in summer (119.8 GC/100 ml) was significantly higher than in winter (58.6 GC/100 ml) (p < 0.01). Statistically significant positive correlations were found between N. fowleri and qPCR E. coli results and N. fowleri and colilert E. coli (culture method), respectively. A weak positive correlation between E. coli and Enterococcus was observed from both qPCR (r = 0.27, p < 0.05) and culture based method (r = 0.52, p < 0.05). Meanwhile, significant positive correlation between qPCR and culture based methods for E. coli (r = 0.30, p < 0.05) and Enterococcus concentration was observed (r = 0.26, p < 0.05), respectively. Future research is needed to determine whether sediment is a source of N. fowleri found in the water column.
Method of identifying clusters representing statistical dependencies in multivariate data
NASA Technical Reports Server (NTRS)
Borucki, W. J.; Card, D. H.; Lyle, G. C.
1975-01-01
Approach is first to cluster and then to compute spatial boundaries for resulting clusters. Next step is to compute, from set of Monte Carlo samples obtained from scrambled data, estimates of probabilities of obtaining at least as many points within boundaries as were actually observed in original data.
Kirgiz, Irina A; Calloway, Cassandra
2017-04-01
Tape lifting and FTA paper scraping methods were directly compared to traditional double swabbing for collecting touch DNA from car steering wheels (n = 70 cars). Touch DNA was collected from the left or right side of each steering wheel (randomized) using two sterile cotton swabs, while the other side was sampled using water-soluble tape or FTA paper cards. DNA was extracted and quantified in duplicate using qPCR. Quantifiable amounts of DNA were detected for 100% of the samples (n = 140) collected independent of the method. However, the DNA collection yield was dependent on the collection method. A statistically significant difference in DNA yield was observed between FTA scraping and double swabbing methods (p = 0.0051), with FTA paper collecting a two-fold higher amount. Statistical analysis showed no significant difference in DNA yields between the double swabbing and tape lifting techniques (p = 0.21). Based on the DNA concentration required for 1 ng input, 47% of the samples collected using FTA paper would be expected to yield a short tandem repeat (STR) profile compared to 30% and 23% using double swabbing or tape, respectively. Further, 55% and 77% of the samples collected using double swabbing or tape, respectively, did not yield a high enough DNA concentration for the 0.5 ng of DNA input recommended for conventional STR kits and would be expected to result in a partial or no profile compared to 35% of the samples collected using FTA paper. STR analysis was conducted for a subset of the higher concentrated samples to confirm that the DNA collected from the steering wheel was from the driver. 32 samples were selected with DNA amounts of at least 1 ng total DNA (100 pg/μl when concentrated if required). A mixed STR profile was observed for 26 samples (88%) and the last driver was the major DNA contributor for 29 samples (94%). For one sample, the last driver was the minor DNA contributor. A full STR profile of the last driver was observed for 21 samples (69%) and a partial profile was observed for nine samples (25%); STR analysis failed for two samples collected using tape (6%). In conclusion, we show that the FTA paper scraping method has the potential to collect higher DNA yields from touch DNA evidence deposited on non-porous surfaces often encountered in criminal cases compared to conventional methods. Copyright © 2017 Elsevier Ltd and Faculty of Forensic and Legal Medicine. All rights reserved.
45 CFR 160.536 - Statistical sampling.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 45 Public Welfare 1 2010-10-01 2010-10-01 false Statistical sampling. 160.536 Section 160.536... REQUIREMENTS GENERAL ADMINISTRATIVE REQUIREMENTS Procedures for Hearings § 160.536 Statistical sampling. (a) In... statistical sampling study as evidence of the number of violations under § 160.406 of this part, or the...
42 CFR 1003.133 - Statistical sampling.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 42 Public Health 5 2011-10-01 2011-10-01 false Statistical sampling. 1003.133 Section 1003.133... AUTHORITIES CIVIL MONEY PENALTIES, ASSESSMENTS AND EXCLUSIONS § 1003.133 Statistical sampling. (a) In meeting... statistical sampling study as evidence of the number and amount of claims and/or requests for payment as...
45 CFR 160.536 - Statistical sampling.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 45 Public Welfare 1 2011-10-01 2011-10-01 false Statistical sampling. 160.536 Section 160.536... REQUIREMENTS GENERAL ADMINISTRATIVE REQUIREMENTS Procedures for Hearings § 160.536 Statistical sampling. (a) In... statistical sampling study as evidence of the number of violations under § 160.406 of this part, or the...
42 CFR 1003.133 - Statistical sampling.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 42 Public Health 5 2010-10-01 2010-10-01 false Statistical sampling. 1003.133 Section 1003.133... AUTHORITIES CIVIL MONEY PENALTIES, ASSESSMENTS AND EXCLUSIONS § 1003.133 Statistical sampling. (a) In meeting... statistical sampling study as evidence of the number and amount of claims and/or requests for payment as...
42 CFR 405.1064 - ALJ decisions involving statistical samples.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 42 Public Health 2 2011-10-01 2011-10-01 false ALJ decisions involving statistical samples. 405... Medicare Coverage Policies § 405.1064 ALJ decisions involving statistical samples. When an appeal from the QIC involves an overpayment issue and the QIC used a statistical sample in reaching its...
42 CFR 405.1064 - ALJ decisions involving statistical samples.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 42 Public Health 2 2010-10-01 2010-10-01 false ALJ decisions involving statistical samples. 405... Medicare Coverage Policies § 405.1064 ALJ decisions involving statistical samples. When an appeal from the QIC involves an overpayment issue and the QIC used a statistical sample in reaching its...
Survey methods for assessing land cover map accuracy
Nusser, S.M.; Klaas, E.E.
2003-01-01
The increasing availability of digital photographic materials has fueled efforts by agencies and organizations to generate land cover maps for states, regions, and the United States as a whole. Regardless of the information sources and classification methods used, land cover maps are subject to numerous sources of error. In order to understand the quality of the information contained in these maps, it is desirable to generate statistically valid estimates of accuracy rates describing misclassification errors. We explored a full sample survey framework for creating accuracy assessment study designs that balance statistical and operational considerations in relation to study objectives for a regional assessment of GAP land cover maps. We focused not only on appropriate sample designs and estimation approaches, but on aspects of the data collection process, such as gaining cooperation of land owners and using pixel clusters as an observation unit. The approach was tested in a pilot study to assess the accuracy of Iowa GAP land cover maps. A stratified two-stage cluster sampling design addressed sample size requirements for land covers and the need for geographic spread while minimizing operational effort. Recruitment methods used for private land owners yielded high response rates, minimizing a source of nonresponse error. Collecting data for a 9-pixel cluster centered on the sampled pixel was simple to implement, and provided better information on rarer vegetation classes as well as substantial gains in precision relative to observing data at a single-pixel.
The effect of different drinks on tooth color after home bleaching
Karadas, Muhammet; Seven, Nilgun
2014-01-01
Objective: This study evaluated the influence of coffee, tea, cola, and red wine staining on the color of teeth after home bleaching. Materials and Methods: A total of 45 samples were obtained from 45 sound maxillary central incisors. The home bleaching procedure was performed using 10% carbamide peroxide gel applied to the sample surface for a period of 6 h each day, for 14 days. After bleaching, baseline color measurements were taken, and the samples were immersed in four staining solutions (coffee, tea, cola, and red wine) or artificial saliva (n = 9). Following 15 min and 6 h of immersion on the first day and next day, respectively, the samples were washed with distilled water for 10 s. After 15 min, 6 h, 1 week, and 1 month immersions, the color values of each sample were remeasured and the color change values (∆E) were calculated. Color change analysis was performed using a spectrophotometer. The results were analyzed using analysis of variance and Tukey's honestly significant difference test (P <0.05). Results: Of all the staining solutions, the lowest ∆E values were observed with coffee staining versus artificial saliva (control group), for all time intervals evaluated after whitening. Although no statistically differences were observed between the coffee and control group at all the time points evaluated, there were statistically significant differences between the red wine, cola, and tea solutions. Conclusion: Following tooth whitening, patients should avoid drinks that cause tooth staining, particularly red wine, tea and cola. PMID:24966778
Further developments in cloud statistics for computer simulations
NASA Technical Reports Server (NTRS)
Chang, D. T.; Willand, J. H.
1972-01-01
This study is a part of NASA's continued program to provide global statistics of cloud parameters for computer simulation. The primary emphasis was on the development of the data bank of the global statistical distributions of cloud types and cloud layers and their applications in the simulation of the vertical distributions of in-cloud parameters such as liquid water content. These statistics were compiled from actual surface observations as recorded in Standard WBAN forms. Data for a total of 19 stations were obtained and reduced. These stations were selected to be representative of the 19 primary cloud climatological regions defined in previous studies of cloud statistics. Using the data compiled in this study, a limited study was conducted of the hemogeneity of cloud regions, the latitudinal dependence of cloud-type distributions, the dependence of these statistics on sample size, and other factors in the statistics which are of significance to the problem of simulation. The application of the statistics in cloud simulation was investigated. In particular, the inclusion of the new statistics in an expanded multi-step Monte Carlo simulation scheme is suggested and briefly outlined.
Hayat, Matthew J.; Powell, Amanda; Johnson, Tessa; Cadwell, Betsy L.
2017-01-01
Statistical literacy and knowledge is needed to read and understand the public health literature. The purpose of this study was to quantify basic and advanced statistical methods used in public health research. We randomly sampled 216 published articles from seven top tier general public health journals. Studies were reviewed by two readers and a standardized data collection form completed for each article. Data were analyzed with descriptive statistics and frequency distributions. Results were summarized for statistical methods used in the literature, including descriptive and inferential statistics, modeling, advanced statistical techniques, and statistical software used. Approximately 81.9% of articles reported an observational study design and 93.1% of articles were substantively focused. Descriptive statistics in table or graphical form were reported in more than 95% of the articles, and statistical inference reported in more than 76% of the studies reviewed. These results reveal the types of statistical methods currently used in the public health literature. Although this study did not obtain information on what should be taught, information on statistical methods being used is useful for curriculum development in graduate health sciences education, as well as making informed decisions about continuing education for public health professionals. PMID:28591190
Hayat, Matthew J; Powell, Amanda; Johnson, Tessa; Cadwell, Betsy L
2017-01-01
Statistical literacy and knowledge is needed to read and understand the public health literature. The purpose of this study was to quantify basic and advanced statistical methods used in public health research. We randomly sampled 216 published articles from seven top tier general public health journals. Studies were reviewed by two readers and a standardized data collection form completed for each article. Data were analyzed with descriptive statistics and frequency distributions. Results were summarized for statistical methods used in the literature, including descriptive and inferential statistics, modeling, advanced statistical techniques, and statistical software used. Approximately 81.9% of articles reported an observational study design and 93.1% of articles were substantively focused. Descriptive statistics in table or graphical form were reported in more than 95% of the articles, and statistical inference reported in more than 76% of the studies reviewed. These results reveal the types of statistical methods currently used in the public health literature. Although this study did not obtain information on what should be taught, information on statistical methods being used is useful for curriculum development in graduate health sciences education, as well as making informed decisions about continuing education for public health professionals.
2006-09-30
strength of the combination is that the tracking system quantifies swimming behaviors of protists in natural seawater samples with large numbers of motile...Sound was to link observations of thin layers to behavioral analysis of protists resident above, within, and below these features. Analysis of our...cells and diatom chains. We are not yet able to make statistical statements about swimming characteristics of the motile protists in our video samples
Speckle in the diffraction patterns of Hendricks-Teller and icosahedral glass models
NASA Technical Reports Server (NTRS)
Garg, Anupam; Levine, Dov
1988-01-01
It is shown that the X-ray diffraction patterns from the Hendricks-Teller model for layered systems and the icosahedral glass models for the icosahedral phases show large fluctuations between nearby scattering wave vectors and from sample to sample, that are quite analogous to laser speckle. The statistics of these fluctuations are studied analytically for the first model and via computer simulations for the second. The observability of these effects is discussed briefly.
Shamey, Renzo; Zubair, Muhammad; Cheema, Hammad
2015-08-01
The aim of this study was twofold, first to determine the effect of field view size and second of illumination conditions on the selection of unique hue samples (UHs: R, Y, G and B) from two rotatable trays, each containing forty highly chromatic Natural Color System (NCS) samples, on one tray corresponding to 1.4° and on the other to 5.7° field of view size. UH selections were made by 25 color-normal observers who repeated assessments three times with a gap of at least 24h between trials. Observers separately assessed UHs under four illumination conditions simulating illuminants D65, A, F2 and F11. An apparent hue shift (statistically significant for UR) was noted for UH selections at 5.7° field of view compared to those at 1.4°. Observers' overall variability was found to be higher for UH stimuli selections at the larger field of view. Intra-observer variability was found to be approximately 18.7% of inter-observer variability in selection of samples for both sample sizes. The highest intra-observer variability was under simulated illuminant D65, followed by A, F11, and F2. Copyright © 2015 Elsevier Ltd. All rights reserved.
Monitoring the impact of Bt maize on butterflies in the field: estimation of required sample sizes.
Lang, Andreas
2004-01-01
The monitoring of genetically modified organisms (GMOs) after deliberate release is important in order to assess and evaluate possible environmental effects. Concerns have been raised that the transgenic crop, Bt maize, may affect butterflies occurring in field margins. Therefore, a monitoring of butterflies was suggested accompanying the commercial cultivation of Bt maize. In this study, baseline data on the butterfly species and their abundance in maize field margins is presented together with implications for butterfly monitoring. The study was conducted in Bavaria, South Germany, between 2000-2002. A total of 33 butterfly species was recorded in field margins. A small number of species dominated the community, and butterflies observed were mostly common species. Observation duration was the most important factor influencing the monitoring results. Field margin size affected the butterfly abundance, and habitat diversity had a tendency to influence species richness. Sample size and statistical power analyses indicated that a sample size in the range of 75 to 150 field margins for treatment (transgenic maize) and control (conventional maize) would detect (power of 80%) effects larger than 15% in species richness and the butterfly abundance pooled across species. However, a much higher number of field margins must be sampled in order to achieve a higher statistical power, to detect smaller effects, and to monitor single butterfly species.
Comparative study of millennials' (age 20-34 years) grip and lateral pinch with the norms.
Fain, Elizabeth; Weatherford, Cara
Cross-sectional research design. Clinical practice continues to use normative data for grip and pinch measurements that were established in 1985. There is no updated norms despite different hand usage patterns in today's society. Measuring and comparing grip and pinch strengths with normative data is a valid method to determine hand function. This research was implemented to compare the grip and pinch measurements obtained from healthy millennials to the established norms and to describe hand usage patterns for millennials. Grip and lateral pinch measurements were obtained from a sample of 237 healthy millennials (ages 20-34 years). Strength scores were statistically lower that older normative data in all millennial grip strengths, with the exception of the women in the age group of 30-34 years. Specifically, this statistically significant trend was observed in all male grip strengths, as well as in women in the age group of 20-24 years (bilateral grip) and 25-29 years (right grip). However, the lateral pinch data reflected was similar to the older norms with variances of 0.5-1 kg. Current data reflect statistically significant differences from the norms for all male grip measurements, as well as for women in the age group of 20-24 years (bilateral grip) and 25-29 years (right grip). No statistical significance was observed in the independent-sample t tests for the lateral pinch in men of all age groups. Statistical significance was noted for lateral pinch for female age groups for the left hand (20-24 years) and for bilateral lateral pinches (30-34 years). IV. Copyright © 2016 Hanley & Belfus. Published by Elsevier Inc. All rights reserved.
Simulation of Wind Profile Perturbations for Launch Vehicle Design
NASA Technical Reports Server (NTRS)
Adelfang, S. I.
2004-01-01
Ideally, a statistically representative sample of measured high-resolution wind profiles with wavelengths as small as tens of meters is required in design studies to establish aerodynamic load indicator dispersions and vehicle control system capability. At most potential launch sites, high- resolution wind profiles may not exist. Representative samples of Rawinsonde wind profiles to altitudes of 30 km are more likely to be available from the extensive network of measurement sites established for routine sampling in support of weather observing and forecasting activity. Such a sample, large enough to be statistically representative of relatively large wavelength perturbations, would be inadequate for launch vehicle design assessments because the Rawinsonde system accurately measures wind perturbations with wavelengths no smaller than 2000 m (1000 m altitude increment). The Kennedy Space Center (KSC) Jimsphere wind profiles (150/month and seasonal 2 and 3.5-hr pairs) are the only adequate samples of high resolution profiles approx. 150 to 300 m effective resolution, but over-sampled at 25 m intervals) that have been used extensively for launch vehicle design assessments. Therefore, a simulation process has been developed for enhancement of measured low-resolution Rawinsonde profiles that would be applicable in preliminary launch vehicle design studies at launch sites other than KSC.
Schlossberg, Scott; Chase, Michael J; Griffin, Curtice R
2016-01-01
Accurate counts of animals are critical for prioritizing conservation efforts. Past research, however, suggests that observers on aerial surveys may fail to detect all individuals of the target species present in the survey area. Such errors could bias population estimates low and confound trend estimation. We used two approaches to assess the accuracy of aerial surveys for African savanna elephants (Loxodonta africana) in northern Botswana. First, we used double-observer sampling, in which two observers make observations on the same herds, to estimate detectability of elephants and determine what variables affect it. Second, we compared total counts, a complete survey of the entire study area, against sample counts, in which only a portion of the study area is sampled. Total counts are often considered a complete census, so comparing total counts against sample counts can help to determine if sample counts are underestimating elephant numbers. We estimated that observers detected only 76% ± SE of 2% of elephant herds and 87 ± 1% of individual elephants present in survey strips. Detectability increased strongly with elephant herd size. Out of the four observers used in total, one observer had a lower detection probability than the other three, and detectability was higher in the rear row of seats than the front. The habitat immediately adjacent to animals also affected detectability, with detection more likely in more open habitats. Total counts were not statistically distinguishable from sample counts. Because, however, the double-observer samples revealed that observers missed 13% of elephants, we conclude that total counts may be undercounting elephants as well. These results suggest that elephant population estimates from both sample and total counts are biased low. Because factors such as observer and habitat affected detectability of elephants, comparisons of elephant populations across time or space may be confounded. We encourage survey teams to incorporate detectability analysis in all aerial surveys for mammals.
Schlossberg, Scott; Chase, Michael J.; Griffin, Curtice R.
2016-01-01
Accurate counts of animals are critical for prioritizing conservation efforts. Past research, however, suggests that observers on aerial surveys may fail to detect all individuals of the target species present in the survey area. Such errors could bias population estimates low and confound trend estimation. We used two approaches to assess the accuracy of aerial surveys for African savanna elephants (Loxodonta africana) in northern Botswana. First, we used double-observer sampling, in which two observers make observations on the same herds, to estimate detectability of elephants and determine what variables affect it. Second, we compared total counts, a complete survey of the entire study area, against sample counts, in which only a portion of the study area is sampled. Total counts are often considered a complete census, so comparing total counts against sample counts can help to determine if sample counts are underestimating elephant numbers. We estimated that observers detected only 76% ± SE of 2% of elephant herds and 87 ± 1% of individual elephants present in survey strips. Detectability increased strongly with elephant herd size. Out of the four observers used in total, one observer had a lower detection probability than the other three, and detectability was higher in the rear row of seats than the front. The habitat immediately adjacent to animals also affected detectability, with detection more likely in more open habitats. Total counts were not statistically distinguishable from sample counts. Because, however, the double-observer samples revealed that observers missed 13% of elephants, we conclude that total counts may be undercounting elephants as well. These results suggest that elephant population estimates from both sample and total counts are biased low. Because factors such as observer and habitat affected detectability of elephants, comparisons of elephant populations across time or space may be confounded. We encourage survey teams to incorporate detectability analysis in all aerial surveys for mammals. PMID:27755570
Relativistic effects on galaxy redshift samples due to target selection
NASA Astrophysics Data System (ADS)
Alam, Shadab; Croft, Rupert A. C.; Ho, Shirley; Zhu, Hongyu; Giusarma, Elena
2017-10-01
In a galaxy redshift survey, the objects to be targeted for spectra are selected from a photometrically observed sample. The observed magnitudes and colours of galaxies in this parent sample will be affected by their peculiar velocities, through relativistic Doppler and relativistic beaming effects. In this paper, we compute the resulting expected changes in galaxy photometry. The magnitudes of the relativistic effects are a function of redshift, stellar mass, galaxy velocity and velocity direction. We focus on the CMASS sample from the Sloan Digital Sky Survey (SDSS) and Baryon Oscillation Spectroscopic Survey (BOSS), which is selected on the basis of colour and magnitude. We find that 0.10 per cent of the sample (∼585 galaxies) has been scattered into the targeted region of colour-magnitude space by relativistic effects, and conversely 0.09 per cent of the sample (∼532 galaxies) has been scattered out. Observational consequences of these effects include an asymmetry in clustering statistics, which we explore in a companion paper. Here, we compute a set of weights that can be used to remove the effect of modulations introduced into the density field inferred from a galaxy sample. We conclude by investigating the possible effects of these relativistic modulation on large-scale clustering of the galaxy sample.
Kulesz, Paulina A.; Tian, Siva; Juranek, Jenifer; Fletcher, Jack M.; Francis, David J.
2015-01-01
Objective Weak structure-function relations for brain and behavior may stem from problems in estimating these relations in small clinical samples with frequently occurring outliers. In the current project, we focused on the utility of using alternative statistics to estimate these relations. Method Fifty-four children with spina bifida meningomyelocele performed attention tasks and received MRI of the brain. Using a bootstrap sampling process, the Pearson product moment correlation was compared with four robust correlations: the percentage bend correlation, the Winsorized correlation, the skipped correlation using the Donoho-Gasko median, and the skipped correlation using the minimum volume ellipsoid estimator Results All methods yielded similar estimates of the relations between measures of brain volume and attention performance. The similarity of estimates across correlation methods suggested that the weak structure-function relations previously found in many studies are not readily attributable to the presence of outlying observations and other factors that violate the assumptions behind the Pearson correlation. Conclusions Given the difficulty of assembling large samples for brain-behavior studies, estimating correlations using multiple, robust methods may enhance the statistical conclusion validity of studies yielding small, but often clinically significant, correlations. PMID:25495830
Kulesz, Paulina A; Tian, Siva; Juranek, Jenifer; Fletcher, Jack M; Francis, David J
2015-03-01
Weak structure-function relations for brain and behavior may stem from problems in estimating these relations in small clinical samples with frequently occurring outliers. In the current project, we focused on the utility of using alternative statistics to estimate these relations. Fifty-four children with spina bifida meningomyelocele performed attention tasks and received MRI of the brain. Using a bootstrap sampling process, the Pearson product-moment correlation was compared with 4 robust correlations: the percentage bend correlation, the Winsorized correlation, the skipped correlation using the Donoho-Gasko median, and the skipped correlation using the minimum volume ellipsoid estimator. All methods yielded similar estimates of the relations between measures of brain volume and attention performance. The similarity of estimates across correlation methods suggested that the weak structure-function relations previously found in many studies are not readily attributable to the presence of outlying observations and other factors that violate the assumptions behind the Pearson correlation. Given the difficulty of assembling large samples for brain-behavior studies, estimating correlations using multiple, robust methods may enhance the statistical conclusion validity of studies yielding small, but often clinically significant, correlations. PsycINFO Database Record (c) 2015 APA, all rights reserved.
Experimental and environmental factors affect spurious detection of ecological thresholds
Daily, Jonathan P.; Hitt, Nathaniel P.; Smith, David; Snyder, Craig D.
2012-01-01
Threshold detection methods are increasingly popular for assessing nonlinear responses to environmental change, but their statistical performance remains poorly understood. We simulated linear change in stream benthic macroinvertebrate communities and evaluated the performance of commonly used threshold detection methods based on model fitting (piecewise quantile regression [PQR]), data partitioning (nonparametric change point analysis [NCPA]), and a hybrid approach (significant zero crossings [SiZer]). We demonstrated that false detection of ecological thresholds (type I errors) and inferences on threshold locations are influenced by sample size, rate of linear change, and frequency of observations across the environmental gradient (i.e., sample-environment distribution, SED). However, the relative importance of these factors varied among statistical methods and between inference types. False detection rates were influenced primarily by user-selected parameters for PQR (τ) and SiZer (bandwidth) and secondarily by sample size (for PQR) and SED (for SiZer). In contrast, the location of reported thresholds was influenced primarily by SED. Bootstrapped confidence intervals for NCPA threshold locations revealed strong correspondence to SED. We conclude that the choice of statistical methods for threshold detection should be matched to experimental and environmental constraints to minimize false detection rates and avoid spurious inferences regarding threshold location.
Review of research designs and statistical methods employed in dental postgraduate dissertations.
Shirahatti, Ravi V; Hegde-Shetiya, Sahana
2015-01-01
There is a need to evaluate the quality of postgraduate dissertations of dentistry submitted to university in the light of the international standards of reporting. We conducted the review with an objective to document the use of sampling methods, measurement standardization, blinding, methods to eliminate bias, appropriate use of statistical tests, appropriate use of data presentation in postgraduate dental research and suggest and recommend modifications. The public access database of the dissertations from Rajiv Gandhi University of Health Sciences was reviewed. Three hundred and thirty-three eligible dissertations underwent preliminary evaluation followed by detailed evaluation of 10% of randomly selected dissertations. The dissertations were assessed based on international reporting guidelines such as strengthening the reporting of observational studies in epidemiology (STROBE), consolidated standards of reporting trials (CONSORT), and other scholarly resources. The data were compiled using MS Excel and SPSS 10.0. Numbers and percentages were used for describing the data. The "in vitro" studies were the most common type of research (39%), followed by observational (32%) and experimental studies (29%). The disciplines conservative dentistry (92%) and prosthodontics (75%) reported high numbers of in vitro research. Disciplines oral surgery (80%) and periodontics (67%) had conducted experimental studies as a major share of their research. Lacunae in the studies included observational studies not following random sampling (70%), experimental studies not following random allocation (75%), not mentioning about blinding, confounding variables and calibrations in measurements, misrepresenting the data by inappropriate data presentation, errors in reporting probability values and not reporting confidence intervals. Few studies showed grossly inappropriate choice of statistical tests and many studies needed additional tests. Overall observations indicated the need to comply with standard guidelines of reporting research.
Comparative Financial Statistics for Public Two-Year Colleges: FY 1993 National Sample.
ERIC Educational Resources Information Center
Dickmeyer, Nathan; Meeker, Bradley
This report provides comparative information derived from a national sample of 516 public two-year colleges, highlighting financial statistics for fiscal year, 1992-93. This report provides space for colleges to compare their institutional statistics with national sample medians, quartile data for the national sample, and statistics presented in a…
7 CFR 52.38a - Definitions of terms applicable to statistical sampling.
Code of Federal Regulations, 2011 CFR
2011-01-01
... 7 Agriculture 2 2011-01-01 2011-01-01 false Definitions of terms applicable to statistical... Sampling § 52.38a Definitions of terms applicable to statistical sampling. (a) Terms applicable to both on... acceptable as a process average. At the AQL's contained in the statistical sampling plans of this subpart...
7 CFR 52.38a - Definitions of terms applicable to statistical sampling.
Code of Federal Regulations, 2010 CFR
2010-01-01
... 7 Agriculture 2 2010-01-01 2010-01-01 false Definitions of terms applicable to statistical... Sampling § 52.38a Definitions of terms applicable to statistical sampling. (a) Terms applicable to both on... acceptable as a process average. At the AQL's contained in the statistical sampling plans of this subpart...
Wang, Dan; Singhasemanon, Nan; Goh, Kean S
2016-11-15
Pesticides are routinely monitored in surface waters and resultant data are analyzed to assess whether their uses will damage aquatic eco-systems. However, the utility of the monitoring data is limited because of the insufficiency in the temporal and spatial sampling coverage and the inability to detect and quantify trace concentrations. This study developed a novel assessment procedure that addresses those limitations by combining 1) statistical methods capable of extracting information from concentrations below changing detection limits, 2) statistical resampling techniques that account for uncertainties rooted in the non-detects and insufficient/irregular sampling coverage, and 3) multiple lines of evidence that improve confidence in the final conclusion. This procedure was demonstrated by an assessment on chlorpyrifos monitoring data in surface waters of California's Central Valley (2005-2013). We detected a significant downward trend in the concentrations, which cannot be observed by commonly-used statistical approaches. We assessed that the aquatic risk was low using a probabilistic method that works with non-detects and has the ability to differentiate indicator groups with varying sensitivity. In addition, we showed that the frequency of exceedance over ambient aquatic life water quality criteria was affected by pesticide use, precipitation and irrigation demand in certain periods anteceding the water sampling events. Copyright © 2016 Elsevier B.V. All rights reserved.
Predictive data modeling of human type II diabetes related statistics
NASA Astrophysics Data System (ADS)
Jaenisch, Kristina L.; Jaenisch, Holger M.; Handley, James W.; Albritton, Nathaniel G.
2009-04-01
During the course of routine Type II treatment of one of the authors, it was decided to derive predictive analytical Data Models of the daily sampled vital statistics: namely weight, blood pressure, and blood sugar, to determine if the covariance among the observed variables could yield a descriptive equation based model, or better still, a predictive analytical model that could forecast the expected future trend of the variables and possibly eliminate the number of finger stickings required to montior blood sugar levels. The personal history and analysis with resulting models are presented.
Considerations for the design, analysis and presentation of in vivo studies.
Ranstam, J; Cook, J A
2017-03-01
To describe, explain and give practical suggestions regarding important principles and key methodological challenges in the study design, statistical analysis, and reporting of results from in vivo studies. Pre-specifying endpoints and analysis, recognizing the common underlying assumption of statistically independent observations, performing sample size calculations, and addressing multiplicity issues are important parts of an in vivo study. A clear reporting of results and informative graphical presentations of data are other important parts. Copyright © 2016 Osteoarthritis Research Society International. Published by Elsevier Ltd. All rights reserved.
X-ray studies of quasars with the Einstein Observatory. IV - X-ray dependence on radio emission
NASA Technical Reports Server (NTRS)
Worrall, D. M.; Tananbaum, H.; Giommi, P.; Zamorani, G.
1987-01-01
The X-ray properties of a sample of 114 radio-loud quasars observed with the Einstein Observatory are examined, and the results are compared with those obtained from a large sample of radio-quiet quasars. The results of statistical analysis of the dependence of X-ray luminosity on combined functions of optical and radio luminosity show that the dependence on both luminosities is important. However, statistically significant differences are found between subsamples of flat radio spectra quasars and steep radio spectra quasars with regard to dependence of X-ray luminosity on only radio luminosity. The data are consistent with radio-loud quasars having a physical component, not directly related to the optical luminosity, which produces the core radio luminosity plus 'extra' X-ray emission.
NASA Technical Reports Server (NTRS)
Torres-Pomales, Wilfredo
2014-01-01
This report describes a modeling and simulation approach for disturbance patterns representative of the environment experienced by a digital system in an electromagnetic reverberation chamber. The disturbance is modeled by a multi-variate statistical distribution based on empirical observations. Extended versions of the Rejection Samping and Inverse Transform Sampling techniques are developed to generate multi-variate random samples of the disturbance. The results show that Inverse Transform Sampling returns samples with higher fidelity relative to the empirical distribution. This work is part of an ongoing effort to develop a resilience assessment methodology for complex safety-critical distributed systems.
Using GEO Optical Observations to Infer Orbit Populations
NASA Technical Reports Server (NTRS)
Matney, Mark; Africano, John
2002-01-01
NASA's Orbital Debris measurements program has a goal to characterize the small debris environment in the geosynchronous Earth-orbit (GEO) region using optical telescopes ("small" refers to objects too small to catalog and track with current systems). Traditionally, observations of GEO and near-GEO objects involve following the object with the telescope long enough to obtain an orbit. When observing very dim objects with small field-of-view telescopes, though, the observations are generally too short to obtain accurate orbital elements. However, it is possible to use such observations to statistically characterize the small object environment. A telescope pointed at a particular spot could potentially see objects in a number of different orbits. Inevitably, when looking at one region for certain types of orbits, there are objects in other types of orbits that cannot be seen. Observation campaigns are designed with these limitations in mind and are set up to span a number of regions of the sky, making it possible to sample all potential orbits under consideration. Each orbit is not seen with the same probability, however, so there are observation biases intrinsic to any observation campaign. Fortunately, it is possible to remove such biases and reconstruct a meaningful estimate of the statistical orbit populations of small objects in GEO. This information, in turn, can be used to investigate the nature of debris sources and to characterize the risk to GEO spacecraft. This paper describes these statistical tools and presents estimates of small object GEO populations.
Kandler, Anne; Shennan, Stephen
2015-12-06
Cultural change can be quantified by temporal changes in frequency of different cultural artefacts and it is a central question to identify what underlying cultural transmission processes could have caused the observed frequency changes. Observed changes, however, often describe the dynamics in samples of the population of artefacts, whereas transmission processes act on the whole population. Here we develop a modelling framework aimed at addressing this inference problem. To do so, we firstly generate population structures from which the observed sample could have been drawn randomly and then determine theoretical samples at a later time t2 produced under the assumption that changes in frequencies are caused by a specific transmission process. Thereby we also account for the potential effect of time-averaging processes in the generation of the observed sample. Subsequent statistical comparisons (e.g. using Bayesian inference) of the theoretical and observed samples at t2 can establish which processes could have produced the observed frequency data. In this way, we infer underlying transmission processes directly from available data without any equilibrium assumption. We apply this framework to a dataset describing pottery from settlements of some of the first farmers in Europe (the LBK culture) and conclude that the observed frequency dynamic of different types of decorated pottery is consistent with age-dependent selection, a preference for 'young' pottery types which is potentially indicative of fashion trends. © 2015 The Author(s).
Testing for qualitative heterogeneity: An application to composite endpoints in survival analysis.
Oulhaj, Abderrahim; El Ghouch, Anouar; Holman, Rury R
2017-01-01
Composite endpoints are frequently used in clinical outcome trials to provide more endpoints, thereby increasing statistical power. A key requirement for a composite endpoint to be meaningful is the absence of the so-called qualitative heterogeneity to ensure a valid overall interpretation of any treatment effect identified. Qualitative heterogeneity occurs when individual components of a composite endpoint exhibit differences in the direction of a treatment effect. In this paper, we develop a general statistical method to test for qualitative heterogeneity, that is to test whether a given set of parameters share the same sign. This method is based on the intersection-union principle and, provided that the sample size is large, is valid whatever the model used for parameters estimation. We propose two versions of our testing procedure, one based on a random sampling from a Gaussian distribution and another version based on bootstrapping. Our work covers both the case of completely observed data and the case where some observations are censored which is an important issue in many clinical trials. We evaluated the size and power of our proposed tests by carrying out some extensive Monte Carlo simulations in the case of multivariate time to event data. The simulations were designed under a variety of conditions on dimensionality, censoring rate, sample size and correlation structure. Our testing procedure showed very good performances in terms of statistical power and type I error. The proposed test was applied to a data set from a single-center, randomized, double-blind controlled trial in the area of Alzheimer's disease.
Limitations of Poisson statistics in describing radioactive decay.
Sitek, Arkadiusz; Celler, Anna M
2015-12-01
The assumption that nuclear decays are governed by Poisson statistics is an approximation. This approximation becomes unjustified when data acquisition times longer than or even comparable with the half-lives of the radioisotope in the sample are considered. In this work, the limits of the Poisson-statistics approximation are investigated. The formalism for the statistics of radioactive decay based on binomial distribution is derived. The theoretical factor describing the deviation of variance of the number of decays predicated by the Poisson distribution from the true variance is defined and investigated for several commonly used radiotracers such as (18)F, (15)O, (82)Rb, (13)N, (99m)Tc, (123)I, and (201)Tl. The variance of the number of decays estimated using the Poisson distribution is significantly different than the true variance for a 5-minute observation time of (11)C, (15)O, (13)N, and (82)Rb. Durations of nuclear medicine studies often are relatively long; they may be even a few times longer than the half-lives of some short-lived radiotracers. Our study shows that in such situations the Poisson statistics is unsuitable and should not be applied to describe the statistics of the number of decays in radioactive samples. However, the above statement does not directly apply to counting statistics at the level of event detection. Low sensitivities of detectors which are used in imaging studies make the Poisson approximation near perfect. Copyright © 2015 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.
A Bayesian nonparametric method for prediction in EST analysis
Lijoi, Antonio; Mena, Ramsés H; Prünster, Igor
2007-01-01
Background Expressed sequence tags (ESTs) analyses are a fundamental tool for gene identification in organisms. Given a preliminary EST sample from a certain library, several statistical prediction problems arise. In particular, it is of interest to estimate how many new genes can be detected in a future EST sample of given size and also to determine the gene discovery rate: these estimates represent the basis for deciding whether to proceed sequencing the library and, in case of a positive decision, a guideline for selecting the size of the new sample. Such information is also useful for establishing sequencing efficiency in experimental design and for measuring the degree of redundancy of an EST library. Results In this work we propose a Bayesian nonparametric approach for tackling statistical problems related to EST surveys. In particular, we provide estimates for: a) the coverage, defined as the proportion of unique genes in the library represented in the given sample of reads; b) the number of new unique genes to be observed in a future sample; c) the discovery rate of new genes as a function of the future sample size. The Bayesian nonparametric model we adopt conveys, in a statistically rigorous way, the available information into prediction. Our proposal has appealing properties over frequentist nonparametric methods, which become unstable when prediction is required for large future samples. EST libraries, previously studied with frequentist methods, are analyzed in detail. Conclusion The Bayesian nonparametric approach we undertake yields valuable tools for gene capture and prediction in EST libraries. The estimators we obtain do not feature the kind of drawbacks associated with frequentist estimators and are reliable for any size of the additional sample. PMID:17868445
A statistical model for water quality predictions from a river discharge using coastal observations
NASA Astrophysics Data System (ADS)
Kim, S.; Terrill, E. J.
2007-12-01
Understanding and predicting coastal ocean water quality has benefits for reducing human health risks, protecting the environment, and improving local economies which depend on clean beaches. Continuous observations of coastal physical oceanography increase the understanding of the processes which control the fate and transport of a riverine plume which potentially contains high levels of contaminants from the upstream watershed. A data-driven model of the fate and transport of river plume water from the Tijuana River has been developed using surface current observations provided by a network of HF radar operated as part of a local coastal observatory that has been in place since 2002. The model outputs are compared with water quality sampling of shoreline indicator bacteria, and the skill of an alarm for low water quality is evaluated using the receiver operating characteristic (ROC) curve. In addition, statistical analysis of beach closures in comparison with environmental variables is also discussed.
Using the Properties of Broad Absorption Line Quasars to Illuminate Quasar Structure
NASA Astrophysics Data System (ADS)
Yong, Suk Yee; King, Anthea L.; Webster, Rachel L.; Bate, Nicholas F.; O'Dowd, Matthew J.; Labrie, Kathleen
2018-06-01
A key to understanding quasar unification paradigms is the emission properties of broad absorption line quasars (BALQs). The fact that only a small fraction of quasar spectra exhibit deep absorption troughs blueward of the broad permitted emission lines provides a crucial clue to the structure of quasar emitting regions. To learn whether it is possible to discriminate between the BALQ and non-BALQ populations given the observed spectral properties of a quasar, we employ two approaches: one based on statistical methods and the other supervised machine learning classification, applied to quasar samples from the Sloan Digital Sky Survey. The features explored include continuum and emission line properties, in particular the absolute magnitude, redshift, spectral index, line width, asymmetry, strength, and relative velocity offsets of high-ionisation C IV λ1549 and low-ionisation Mg II λ2798 lines. We consider a complete population of quasars, and assume that the statistical distributions of properties represent all angles where the quasar is viewed without obscuration. The distributions of the BALQ and non-BALQ sample properties show few significant differences. None of the observed continuum and emission line features are capable of differentiating between the two samples. Most published narrow disk-wind models are inconsistent with these observations, and an alternative disk-wind model is proposed. The key feature of the proposed model is a disk-wind filling a wide opening angle with multiple radial streams of dense clumps.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shettel, D.L. Jr.; Langfeldt, S.L.; Youngquist, C.A.
This report presents a Hydrogeochemical and Stream Sediment Reconnaissance of the Christian NTMS Quadrangle, Alaska. In addition to this abbreviated data release, more complete data are available to the public in machine-readable form. These machine-readable data, as well as quarterly or semiannual program progress reports containing further information on the HSSR program in general, or on the Los Alamos National Laboratory portion of the program in particular, are available from DOE's Technical Library at its Grand Junction Area Office. Presented in this data release are location data, field analyses, and laboratory analyses of several different sample media. For the sakemore » of brevity, many field site observations have not been included in this volume; these data are, however, available on the magnetic tape. Appendices A through D describe the sample media and summarize the analytical results for each medium. The data have been subdivided by one of the Los Alamos National Laboratory sorting programs of Zinkl and others (1981a) into groups of stream-sediment, lake-sediment, stream-water, lake-water, and ground-water samples. For each group which contains a sufficient number of observations, statistical tables, tables of raw data, and 1:1,000,000 scale maps of pertinent elements have been included in this report. Also included are maps showing results of multivariate statistical analyses.« less
Accounting for undetected compounds in statistical analyses of mass spectrometry 'omic studies.
Taylor, Sandra L; Leiserowitz, Gary S; Kim, Kyoungmi
2013-12-01
Mass spectrometry is an important high-throughput technique for profiling small molecular compounds in biological samples and is widely used to identify potential diagnostic and prognostic compounds associated with disease. Commonly, this data generated by mass spectrometry has many missing values resulting when a compound is absent from a sample or is present but at a concentration below the detection limit. Several strategies are available for statistically analyzing data with missing values. The accelerated failure time (AFT) model assumes all missing values result from censoring below a detection limit. Under a mixture model, missing values can result from a combination of censoring and the absence of a compound. We compare power and estimation of a mixture model to an AFT model. Based on simulated data, we found the AFT model to have greater power to detect differences in means and point mass proportions between groups. However, the AFT model yielded biased estimates with the bias increasing as the proportion of observations in the point mass increased while estimates were unbiased with the mixture model except if all missing observations came from censoring. These findings suggest using the AFT model for hypothesis testing and mixture model for estimation. We demonstrated this approach through application to glycomics data of serum samples from women with ovarian cancer and matched controls.
Sim, Julius; Lewis, Martyn
2012-03-01
To investigate methods to determine the size of a pilot study to inform a power calculation for a randomized controlled trial (RCT) using an interval/ratio outcome measure. Calculations based on confidence intervals (CIs) for the sample standard deviation (SD). Based on CIs for the sample SD, methods are demonstrated whereby (1) the observed SD can be adjusted to secure the desired level of statistical power in the main study with a specified level of confidence; (2) the sample for the main study, if calculated using the observed SD, can be adjusted, again to obtain the desired level of statistical power in the main study; (3) the power of the main study can be calculated for the situation in which the SD in the pilot study proves to be an underestimate of the true SD; and (4) an "efficient" pilot size can be determined to minimize the combined size of the pilot and main RCT. Trialists should calculate the appropriate size of a pilot study, just as they should the size of the main RCT, taking into account the twin needs to demonstrate efficiency in terms of recruitment and to produce precise estimates of treatment effect. Copyright © 2012 Elsevier Inc. All rights reserved.
Caviedes-Bucheli, J; Canales-Sánchez, P; Castrillón-Sarria, N; Jovel-Garcia, J; Alvarez-Vásquez, J; Rivero, C; Azuero-Holguín, M M; Diaz, E; Munoz, H R
2009-08-01
To quantify the expression of insulin-like growth factor-1 (IGF-1) and proliferating cell nuclear antigen (PCNA) in human pulp cells of teeth with complete or incomplete root development, to support the specific role of IGF-1 in cell proliferation during tooth development and pulp reparative processes. Twenty six pulp samples were obtained from freshly extracted human third molars, equally divided in two groups according to root development stage (complete or incomplete root development). All samples were processed and immunostained to determine the expression of IGF-1 and PCNA in pulp cells. Sections were observed with a light microscope at 80x and morphometric analyses were performed to calculate the area of PCNA and IGF-1 immunostaining using digital image software. Mann-Whitney's test was used to determine statistically significant differences between groups (P < 0.05) for each peptide and the co-expression of both. Expression of IGF-1 and PCNA was observed in all human pulp samples with a statistically significant higher expression in cells of pulps having complete root development (P = 0.0009). Insulin-like growth factor-1 and PCNA are expressed in human pulp cells, with a significant greater expression in pulp cells of teeth having complete root development.
Validity of strong lensing statistics for constraints on the galaxy evolution model
NASA Astrophysics Data System (ADS)
Matsumoto, Akiko; Futamase, Toshifumi
2008-02-01
We examine the usefulness of the strong lensing statistics to constrain the evolution of the number density of lensing galaxies by adopting the values of the cosmological parameters determined by recent Wilkinson Microwave Anisotropy Probe observation. For this purpose, we employ the lens-redshift test proposed by Kochanek and constrain the parameters in two evolution models, simple power-law model characterized by the power-law indexes νn and νv, and the evolution model by Mitchell et al. based on cold dark matter structure formation scenario. We use the well-defined lens sample from the Sloan Digital Sky Survey (SDSS) and this is similarly sized samples used in the previous studies. Furthermore, we adopt the velocity dispersion function of early-type galaxies based on SDSS DR1 and DR5. It turns out that the indexes of power-law model are consistent with the previous studies, thus our results indicate the mild evolution in the number and velocity dispersion of early-type galaxies out to z = 1. However, we found that the values for p and q used by Mitchell et al. are inconsistent with the presently available observational data. More complete sample is necessary to withdraw more realistic determination on these parameters.
Savoca, Mark E.; Sadorf, Eric M.; Linhart, S. Mike; Akers, Kim K.B.
2000-01-01
Factors other than land use may contribute to observed differences in water quality between and within agricultural and urban areas. Nitrate, atrazine, deethylatrazine, and deisopropylatrazine concentrations were significantly higher in shallow wells with sample intervals nearer the water table and in wells with thinner cumulative clay thickness above the sample intervals. These relations suggest that longer flow paths allow for greater residence time and increase opportunities for sorption, degradation, and dispersion, which may contribute to decreases in nutrient and pesticide concentrations with depth. Nitrogen speciation was influenced by redox conditions. Nitrate concentrations were significantly higher in ground water with dissolved-oxygen concentrations in excess of 0.5 milligram per liter. Ammonia concentrations were higher in ground water with dissolved-oxygen concentrations of 0.5 milligram per liter or less; however, this relation was not statistically significant. The amount of available organic matter may limit denitrification rates. Elevated nitrate concentrations (greater than 2.0 mg/L) were significantly related to lower dissolved organic carbon concentrations in water samples from both agricultural and urban areas. A similar relation between nitrate concentrations (in water) and organic carbon concentrations (in aquifer material) also was observed but was not statistically significant.
Clinical Efficacy of Psychoeducational Interventions with Family Caregivers
ERIC Educational Resources Information Center
Limiñana-Gras, Rosa M.; Colodro-Conde, Lucía; Cuéllar-Flores, Isabel; Sánchez-López, M. Pilar
2016-01-01
The goal of this study is to investigate the efficacy of psychoeducational interventions geared to reducing psychological distress for caregivers in a sample of 90 family caregivers of elderly dependent (78 women and 12 men). We conducted an analysis of the statistical and clinical significance of the changes observed in psychological health…
Shoulder strength value differences between genders and age groups.
Balcells-Diaz, Eudald; Daunis-I-Estadella, Pepus
2018-03-01
The strength of a normal shoulder differs according to gender and decreases with age. Therefore, the Constant score, which is a shoulder function measurement tool that allocates 25% of the final score to strength, differs from the absolute values but likely reflects a normal shoulder. To compare group results, a normalized Constant score is needed, and the first step to achieving normalization involves statistically establishing the gender differences and age-related decline. In this investigation, we sought to verify the gender difference and age-related decline in strength. We obtained a randomized representative sample of the general population in a small to medium-sized Spanish city. We then invited this population to participate in our study, and we measured their shoulder strength. We performed a statistical analysis with a power of 80% and a P value < .05. We observed a statistically significant difference between the genders and a statistically significant decline with age. To the best of our knowledge, this is the first investigation to study a representative sample of the general population from which conclusions can be drawn regarding Constant score normalization. Copyright © 2017 Journal of Shoulder and Elbow Surgery Board of Trustees. Published by Elsevier Inc. All rights reserved.
Statistical Analyses of Scatterplots to Identify Important Factors in Large-Scale Simulations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kleijnen, J.P.C.; Helton, J.C.
1999-04-01
The robustness of procedures for identifying patterns in scatterplots generated in Monte Carlo sensitivity analyses is investigated. These procedures are based on attempts to detect increasingly complex patterns in the scatterplots under consideration and involve the identification of (1) linear relationships with correlation coefficients, (2) monotonic relationships with rank correlation coefficients, (3) trends in central tendency as defined by means, medians and the Kruskal-Wallis statistic, (4) trends in variability as defined by variances and interquartile ranges, and (5) deviations from randomness as defined by the chi-square statistic. The following two topics related to the robustness of these procedures are consideredmore » for a sequence of example analyses with a large model for two-phase fluid flow: the presence of Type I and Type II errors, and the stability of results obtained with independent Latin hypercube samples. Observations from analysis include: (1) Type I errors are unavoidable, (2) Type II errors can occur when inappropriate analysis procedures are used, (3) physical explanations should always be sought for why statistical procedures identify variables as being important, and (4) the identification of important variables tends to be stable for independent Latin hypercube samples.« less
Rasch fit statistics and sample size considerations for polytomous data.
Smith, Adam B; Rush, Robert; Fallowfield, Lesley J; Velikova, Galina; Sharpe, Michael
2008-05-29
Previous research on educational data has demonstrated that Rasch fit statistics (mean squares and t-statistics) are highly susceptible to sample size variation for dichotomously scored rating data, although little is known about this relationship for polytomous data. These statistics help inform researchers about how well items fit to a unidimensional latent trait, and are an important adjunct to modern psychometrics. Given the increasing use of Rasch models in health research the purpose of this study was therefore to explore the relationship between fit statistics and sample size for polytomous data. Data were collated from a heterogeneous sample of cancer patients (n = 4072) who had completed both the Patient Health Questionnaire - 9 and the Hospital Anxiety and Depression Scale. Ten samples were drawn with replacement for each of eight sample sizes (n = 25 to n = 3200). The Rating and Partial Credit Models were applied and the mean square and t-fit statistics (infit/outfit) derived for each model. The results demonstrated that t-statistics were highly sensitive to sample size, whereas mean square statistics remained relatively stable for polytomous data. It was concluded that mean square statistics were relatively independent of sample size for polytomous data and that misfit to the model could be identified using published recommended ranges.
Rasch fit statistics and sample size considerations for polytomous data
Smith, Adam B; Rush, Robert; Fallowfield, Lesley J; Velikova, Galina; Sharpe, Michael
2008-01-01
Background Previous research on educational data has demonstrated that Rasch fit statistics (mean squares and t-statistics) are highly susceptible to sample size variation for dichotomously scored rating data, although little is known about this relationship for polytomous data. These statistics help inform researchers about how well items fit to a unidimensional latent trait, and are an important adjunct to modern psychometrics. Given the increasing use of Rasch models in health research the purpose of this study was therefore to explore the relationship between fit statistics and sample size for polytomous data. Methods Data were collated from a heterogeneous sample of cancer patients (n = 4072) who had completed both the Patient Health Questionnaire – 9 and the Hospital Anxiety and Depression Scale. Ten samples were drawn with replacement for each of eight sample sizes (n = 25 to n = 3200). The Rating and Partial Credit Models were applied and the mean square and t-fit statistics (infit/outfit) derived for each model. Results The results demonstrated that t-statistics were highly sensitive to sample size, whereas mean square statistics remained relatively stable for polytomous data. Conclusion It was concluded that mean square statistics were relatively independent of sample size for polytomous data and that misfit to the model could be identified using published recommended ranges. PMID:18510722
Charm dimuon production in neutrino-nucleon interactions in the NOMAD experiment
NASA Astrophysics Data System (ADS)
Petti, Roberto; Samoylov, Oleg
2012-09-01
We present our new measurement of charm dimuon production in neutrino-iron interactions based upon the full statistics collected by the NOMAD experiment. After background subtraction we observe 15,340 charm dimuon events, providing the largest sample currently available. The analysis exploits the large inclusive charged current sample (about 9 million events after all analysis cuts) to constrain the total systematic uncertainty to about 2%. The extraction of strange sea and charm production parameters is also discussed.
Charm dimuon production in neutrino-nucleon interactions in the NOMAD experiment
NASA Astrophysics Data System (ADS)
Petti, R.; Samoylov, O. B.
2011-12-01
We present our new measurement of charm dimuon production in neutrino-iron interactions based upon the full statistics collected by the NOMAD experiment. After background subtraction we observe 15,340 charm dimuon events, providing the largest sample currently available. The analysis exploits the large inclusive charged current sample (about 9 million events after all analysis cuts) to constrain the total systematic uncertainty to ˜2%. The extraction of strange sea and charm production parameters is also discussed.
NASA Astrophysics Data System (ADS)
Vogel, Thomas; Perez, Danny; Junghans, Christoph
2014-03-01
We show direct formal relationships between the Wang-Landau iteration [PRL 86, 2050 (2001)], metadynamics [PNAS 99, 12562 (2002)] and statistical temperature molecular dynamics [PRL 97, 050601 (2006)], the major Monte Carlo and molecular dynamics work horses for sampling from a generalized, multicanonical ensemble. We aim at helping to consolidate the developments in the different areas by indicating how methodological advancements can be transferred in a straightforward way, avoiding the parallel, largely independent, developments tracks observed in the past.
Acute health effects of PM10 pollution on symptomatic and asymptomatic children
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pope, C.A. 3d.; Dockery, D.W.
1992-05-01
This study assessed the association between daily changes in respiratory health and respirable particulate pollution (PM10) in Utah Valley during the winter of 1990-1991. During the study period, 24-h PM10 concentrations ranged from 7 to 251 micrograms/m3. Participants included symptomatic and asymptomatic samples of fifth- and sixth-grade students. Relatively small but statistically significant (p less than 0.01) negative associations between peak expiratory flow (PEF) and PM10 were observed for both the symptomatic and asymptomatic samples. The association was strongest for the symptomatic children. Large associations between the incidence of respiratory symptoms, especially cough, and PM10 pollution were also observed formore » both samples. Again the association was strongest for the symptomatic sample. Immediate and delayed PM10 effects were observed. Respiratory symptoms and PEF changes were more closely associated with 5-day moving-average PM10 levels than with concurrent-day levels. These associations were also observed at PM10 levels below the 24-h standard of 150 micrograms/m3. This study indicates that both symptomatic and asymptomatic children may suffer acute health effects of respirable particulate pollution, with symptomatic children suffering the most.« less
NASA Astrophysics Data System (ADS)
Liuzzo, E.; Giovannini, G.; Giroletti, M.; Taylor, G. B.
2009-10-01
Aims: To study statistical properties of different classes of sources, it is necessary to observe a sample that is free of selection effects. To do this, we initiated a project to observe a complete sample of radio galaxies selected from the B2 Catalogue of Radio Sources and the Third Cambridge Revised Catalogue (3CR), with no selection constraint on the nuclear properties. We named this sample “the Bologna Complete Sample” (BCS). Methods: We present new VLBI observations at 5 and 1.6 GHz for 33 sources drawn from a sample not biased toward orientation. By combining these data with those in the literature, information on the parsec-scale morphology is available for a total of 76 of 94 radio sources with a range in radio power and kiloparsec-scale morphologies. Results: The fraction of two-sided sources at milliarcsecond resolution is high (30%), compared to the fraction found in VLBI surveys selected at centimeter wavelengths, as expected from the predictions of unified models. The parsec-scale jets are generally found to be straight and to line up with the kiloparsec-scale jets. A few peculiar sources are discussed in detail. Tables 1-4 are only available in electronic form at http://www.aanda.org
Hernández-Morera, Pablo; Castaño-González, Irene; Travieso-González, Carlos M.; Mompeó-Corredera, Blanca; Ortega-Santana, Francisco
2016-01-01
Purpose To develop a digital image processing method to quantify structural components (smooth muscle fibers and extracellular matrix) in the vessel wall stained with Masson’s trichrome, and a statistical method suitable for small sample sizes to analyze the results previously obtained. Methods The quantification method comprises two stages. The pre-processing stage improves tissue image appearance and the vessel wall area is delimited. In the feature extraction stage, the vessel wall components are segmented by grouping pixels with a similar color. The area of each component is calculated by normalizing the number of pixels of each group by the vessel wall area. Statistical analyses are implemented by permutation tests, based on resampling without replacement from the set of the observed data to obtain a sampling distribution of an estimator. The implementation can be parallelized on a multicore machine to reduce execution time. Results The methods have been tested on 48 vessel wall samples of the internal saphenous vein stained with Masson’s trichrome. The results show that the segmented areas are consistent with the perception of a team of doctors and demonstrate good correlation between the expert judgments and the measured parameters for evaluating vessel wall changes. Conclusion The proposed methodology offers a powerful tool to quantify some components of the vessel wall. It is more objective, sensitive and accurate than the biochemical and qualitative methods traditionally used. The permutation tests are suitable statistical techniques to analyze the numerical measurements obtained when the underlying assumptions of the other statistical techniques are not met. PMID:26761643
Understanding the Sampling Distribution and the Central Limit Theorem.
ERIC Educational Resources Information Center
Lewis, Charla P.
The sampling distribution is a common source of misuse and misunderstanding in the study of statistics. The sampling distribution, underlying distribution, and the Central Limit Theorem are all interconnected in defining and explaining the proper use of the sampling distribution of various statistics. The sampling distribution of a statistic is…
Alpha1 LASSO data bundles Lamont, OK
Gustafson, William Jr; Vogelmann, Andrew; Endo, Satoshi; Toto, Tami; Xiao, Heng; Li, Zhijin; Cheng, Xiaoping; Krishna, Bhargavi (ORCID:000000018828528X)
2016-08-03
A data bundle is a unified package consisting of LASSO LES input and output, observations, evaluation diagnostics, and model skill scores. LES input includes model configuration information and forcing data. LES output includes profile statistics and full domain fields of cloud and environmental variables. Model evaluation data consists of LES output and ARM observations co-registered on the same grid and sampling frequency. Model performance is quantified by skill scores and diagnostics in terms of cloud and environmental variables.
Past, Present and Emerging Toxicity Issues for Jet Fuel
2011-01-01
Statistically significant dominant lethal effects were not observed for either mice or rats (Air Force, 1978). However, because of the small sample...Adams, M.M., 2004. Immunological and hematological effects observed in B6C3F1 mice exposed to JP-8 jet fuel for 14 days. J. Toxicol. Environ. Health A...acute toxicity with the adverse effects being slight dermal irritation and weak dermal sensitization in animals. JP-4 also has low acute toxicity with
Relationship between uterine biopsy score, endometrial infection and inflammation in the mare.
Buczkowska, Justyna; Kozdrowski, Roland; Nowak, Marcin; Sikora, Monika
2016-06-16
Endometrial biopsy score is an accepted marker of uterine health and predicted fertility, and it has been suggested that endometrial alternations are correlated with susceptibility to persistent infectious endometritis. The objective of this study was to investigate associations of endometrial biopsy score with: 1) presence of polymorphonuclear cells (PMNs) in the epithelium and stratum compactum in histopathology; 2) presence of PMNs in cytology and 3) presence of infection in microbiology. The material for examination was collected from 69 mares suspected for subclinical endometritis (bred three or more times unsuccessfully in the same breeding season) and from 15 maiden mares. Samples were collected by endometrial biopsy and cytobrush technique. Endometrial alterations (biopsy score IIA, IIB, III) were found in 64 of 82 mares (78%). There was an increase in PMN occurrence for grades IIA, IIB and III. When comparing grades and PMNs infiltration, we observed statistically significant differences between grades I and IIA (p = 0.222) and grades I and IIB (p = 0.042) in samples collected by endometrial biopsy. Statistically significant differences were found in microbiological examination (biopsy p = 0.036; cytobrush p = 0.189), cytological examination (biopsy p = 0.040; cytobrush p = 0.079) and PMN infiltration (p = 0.042) between mares with biopsy scores I and IIB. Furthermore, the highest percentage of infected mares was in grade IIA and IIB, and we found statistically significant differences between grades I and IIA (p = 0.043), and grades I and IIB (p = 0.036) in biopsy samples. We observed a tendency to higher prevalence of endometrial infection in mares with biopsy score IIA, IIB and III than with biopsy score I in samples collected using cytobrush technique. However, there were no statistical significant differences. Degenerative endometrial changes can predispose to uterine infection and inflammation. Our study shows that mares with endometrial score I are less predisposed to infection than mares with category IIA, IIB and III. Endometrial biopsy is a reliable diagnostic tool.
ERIC Educational Resources Information Center
Garfield, Joan; Le, Laura; Zieffler, Andrew; Ben-Zvi, Dani
2015-01-01
This paper describes the importance of developing students' reasoning about samples and sampling variability as a foundation for statistical thinking. Research on expert-novice thinking as well as statistical thinking is reviewed and compared. A case is made that statistical thinking is a type of expert thinking, and as such, research…
Applying the J-optimal channelized quadratic observer to SPECT myocardial perfusion defect detection
NASA Astrophysics Data System (ADS)
Kupinski, Meredith K.; Clarkson, Eric; Ghaly, Michael; Frey, Eric C.
2016-03-01
To evaluate performance on a perfusion defect detection task from 540 image pairs of myocardial perfusion SPECT image data we apply the J-optimal channelized quadratic observer (J-CQO). We compare AUC values of the linear Hotelling observer and J-CQO when the defect location is fixed and when it occurs in one of two locations. As expected, when the location is fixed a single channels maximizes AUC; location variability requires multiple channels to maximize the AUC. The AUC is estimated from both the projection data and reconstructed images. J-CQO is quadratic since it uses the first- and second- order statistics of the image data from both classes. The linear data reduction by the channels is described by an L x M channel matrix and in prior work we introduced an iterative gradient-based method for calculating the channel matrix. The dimensionality reduction from M measurements to L channels yields better estimates of these sample statistics from smaller sample sizes, and since the channelized covariance matrix is L x L instead of M x M, the matrix inverse is easier to compute. The novelty of our approach is the use of Jeffrey's divergence (J) as the figure of merit (FOM) for optimizing the channel matrix. We previously showed that the J-optimal channels are also the optimum channels for the AUC and the Bhattacharyya distance when the channel outputs are Gaussian distributed with equal means. This work evaluates the use of J as a surrogate FOM (SFOM) for AUC when these statistical conditions are not satisfied.
Marcinek-Jacel, Marta; Albińska, Jadwiga; Pawlaczyk, Aleksandra; Szynkowska, Małgorzata Iwona
2017-10-01
The aim of this work was to access the influence of different factors such as sex, age, fish consumption, hair dyeing or smoking habit on the content of mercury in human hair samples. The research was carried on 444 samples (102 males and 342 females) collected from the population of people living in the region of Lodz (central Poland). The content of mercury in human hair samples was determined using the Mercury Analyzer MA 3000 (Nippon Instruments, Japan). The obtained results were elaborated using Statistica ver. 10.0 software. The mean value of mercury in investigated human hair samples was found to be 0.174±0.137mg/kg. We observed the statistically significant correlations (p<0.05) between the content of Hg in hair of the studied population and factors such as gender, age, and fish consumption. However, no statistically significant differences were found in relation to cosmetic treatment such as hair dyeing or smoking. Copyright © 2017 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Dennison, J. E.; Lipschutz, M. E.
1987-03-01
The authors report RNAA data for 14 siderophile, lithophile and chalcophile volatile/mobile trace elements in interior portions of 45 different H4-6 chondrites (49 samples) from Victoria Land, Antarctica and 5 H5 chondrites from the Yamato Mts., Antarctica. Relative to H5 chondrites of weathering types A and B, all elements are depleted (10 at statistically significant levels) in extensively weathered (types B/C and C) samples. Chondrites of weathering types A and B seem compositionally uncompromised and as useful as contemporary falls for trace-element studies. When data distributions for these 14 trace elements in non-Antarctic H chondrite falls and unpaired samples from Victoria Land and from the Yamato Mts. (Queen Maud Land) are compared statistically, numerous significant differences are apparent. These and other differences give ample cause to doubt that the various sample populations derive from the same parent population. The observed differences do no reflect weathering, chance or other trivial causes: a preterrestrial source must be responsible.
Evola, Francesco Roberto; Costarella, Luciano; Evola, Giuseppe; Barchitta, Martina; Agodi, Antonella; Sessa, Giuseppe
2017-01-01
AIM To evaluate the clinical and X-ray results of acetabular components and tantalum augments in prosthetic hip revisions. METHODS Fifty-eight hip prostheses with primary failure of the acetabular component were reviewed with tantalum implants. The clinical records and X-rays of these cases were retrospectively reviewed. Bone defect evaluations were based on preoperative CT scans and classified according to Paprosky criteria of Radiolucent lines and periprosthetic gaps; implant mobilization and osteolysis were evaluated by X-ray. An ad hoc database was created and statistical analyses were performed with SPSS software (IBM SPSS Statistics for Windows, version 23.0). Statistical analyses were carried out using the Student’s t test for independent and paired samples. A P value of < 0.05 was considered statistically significant and cumulative survival was calculated by the Kaplan-Meier method. RESULTS The mean follow-up was 87.6 ± 25.6 mo (range 3-120 mo). 25 cases (43.1%) were classified as minor defects, and 33 cases (56.9%) as major defects. The preoperative HHS rating improved significantly from a mean of 40.7 ± 6.1 (range: 29-53) before revision, to a mean of 85.8 ± 6.1 (range: 70-94) at the end of the follow-up (Student’s t test for paired samples: P < 0.001). Considering HHS only at the end of follow-up, no statistically significant difference was observed between patients with a major or minor defect (Student’s t test for independent samples: P > 0.05). Radiolucent lines were found in 4 implants (6.9%). Postoperative acetabular gaps were observed in 5 hips (8.6%). No signs of implant mobilization or areas of periprosthetic osteolysis were found in the x-rays at the final follow-up. Only 3 implants failed: 1 case of infection and 2 cases of instability. Defined as the end-point, cumulative survival at 10 years was 95% (for all reasons) and 100% for aseptic loosening of the acetabular component. CONCLUSION The medium-term use of prosthetic tantalum components in prosthetic hip revisions is safe and effective in a wide variety of acetabular bone defects. PMID:28808626
Evola, Francesco Roberto; Costarella, Luciano; Evola, Giuseppe; Barchitta, Martina; Agodi, Antonella; Sessa, Giuseppe
2017-07-18
To evaluate the clinical and X-ray results of acetabular components and tantalum augments in prosthetic hip revisions. Fifty-eight hip prostheses with primary failure of the acetabular component were reviewed with tantalum implants. The clinical records and X-rays of these cases were retrospectively reviewed. Bone defect evaluations were based on preoperative CT scans and classified according to Paprosky criteria of Radiolucent lines and periprosthetic gaps; implant mobilization and osteolysis were evaluated by X-ray. An ad hoc database was created and statistical analyses were performed with SPSS software (IBM SPSS Statistics for Windows, version 23.0). Statistical analyses were carried out using the Student's t test for independent and paired samples. A P value of < 0.05 was considered statistically significant and cumulative survival was calculated by the Kaplan-Meier method. The mean follow-up was 87.6 ± 25.6 mo (range 3-120 mo). 25 cases (43.1%) were classified as minor defects, and 33 cases (56.9%) as major defects. The preoperative HHS rating improved significantly from a mean of 40.7 ± 6.1 (range: 29-53) before revision, to a mean of 85.8 ± 6.1 (range: 70-94) at the end of the follow-up (Student's t test for paired samples: P < 0.001). Considering HHS only at the end of follow-up, no statistically significant difference was observed between patients with a major or minor defect (Student's t test for independent samples: P > 0.05). Radiolucent lines were found in 4 implants (6.9%). Postoperative acetabular gaps were observed in 5 hips (8.6%). No signs of implant mobilization or areas of periprosthetic osteolysis were found in the x-rays at the final follow-up. Only 3 implants failed: 1 case of infection and 2 cases of instability. Defined as the end-point, cumulative survival at 10 years was 95% (for all reasons) and 100% for aseptic loosening of the acetabular component. The medium-term use of prosthetic tantalum components in prosthetic hip revisions is safe and effective in a wide variety of acetabular bone defects.
Poisson Statistics of Combinatorial Library Sampling Predict False Discovery Rates of Screening
2017-01-01
Microfluidic droplet-based screening of DNA-encoded one-bead-one-compound combinatorial libraries is a miniaturized, potentially widely distributable approach to small molecule discovery. In these screens, a microfluidic circuit distributes library beads into droplets of activity assay reagent, photochemically cleaves the compound from the bead, then incubates and sorts the droplets based on assay result for subsequent DNA sequencing-based hit compound structure elucidation. Pilot experimental studies revealed that Poisson statistics describe nearly all aspects of such screens, prompting the development of simulations to understand system behavior. Monte Carlo screening simulation data showed that increasing mean library sampling (ε), mean droplet occupancy, or library hit rate all increase the false discovery rate (FDR). Compounds identified as hits on k > 1 beads (the replicate k class) were much more likely to be authentic hits than singletons (k = 1), in agreement with previous findings. Here, we explain this observation by deriving an equation for authenticity, which reduces to the product of a library sampling bias term (exponential in k) and a sampling saturation term (exponential in ε) setting a threshold that the k-dependent bias must overcome. The equation thus quantitatively describes why each hit structure’s FDR is based on its k class, and further predicts the feasibility of intentionally populating droplets with multiple library beads, assaying the micromixtures for function, and identifying the active members by statistical deconvolution. PMID:28682059
Climate and Edaphic Controls on Humid Tropical Forest Tree Height
NASA Astrophysics Data System (ADS)
Yang, Y.; Saatchi, S. S.; Xu, L.
2014-12-01
Uncertainty in the magnitude and spatial variations of forest carbon density in tropical regions is due to under sampling of forest structure from inventory plots and the lack of regional allometry to estimate the carbon density from structure. Here we quantify the variation of tropical forest structure by using more than 2.5 million measurements of canopy height from systematic sampling of Geoscience Laser Altimeter System (GLAS) satellite observations between 2004 to 2008 and examine the climate and edaphic variables influencing the variations. We used top canopy height of GLAS footprints (~ 0.25 ha) to grid the statistical mean and 90 percentile of samples at 0.5 degrees to capture the regional variability of large trees in tropics. GLAS heights were also aggregated based on a stratification of tropical regions using soil, elevation, and forest types. Both approaches provided consistent patterns of statistically dominant large trees and the least heterogeneity, both as strong drivers of distribution of high biomass forests. Statistical models accounting for spatial autocorrelation suggest that climate, soil and spatial features together can explain more than 60% of the variations in observed tree height information, while climate-only variables explains about one third of the first-order changes in tree height. Soil basics, including physical compositions such as clay and sand contents, chemical properties such as PH values and cation-exchange capacity, as well as biological variables such as organic matters, all present independent but statistically significant relationships to tree height variations. The results confirm other landscape and regional studies that soil fertility, geology and climate may jointly control a majority of the regional variations of forest structure in pan-tropics and influencing both biomass stocks and dynamics. Consequently, other factors such as biotic and disturbance regimes, not included in this study, may have less influence on regional variations but strongly mediate landscape and small-scale forest structure and dynamics.
Statistical power analysis in wildlife research
Steidl, R.J.; Hayes, J.P.
1997-01-01
Statistical power analysis can be used to increase the efficiency of research efforts and to clarify research results. Power analysis is most valuable in the design or planning phases of research efforts. Such prospective (a priori) power analyses can be used to guide research design and to estimate the number of samples necessary to achieve a high probability of detecting biologically significant effects. Retrospective (a posteriori) power analysis has been advocated as a method to increase information about hypothesis tests that were not rejected. However, estimating power for tests of null hypotheses that were not rejected with the effect size observed in the study is incorrect; these power estimates will always be a??0.50 when bias adjusted and have no relation to true power. Therefore, retrospective power estimates based on the observed effect size for hypothesis tests that were not rejected are misleading; retrospective power estimates are only meaningful when based on effect sizes other than the observed effect size, such as those effect sizes hypothesized to be biologically significant. Retrospective power analysis can be used effectively to estimate the number of samples or effect size that would have been necessary for a completed study to have rejected a specific null hypothesis. Simply presenting confidence intervals can provide additional information about null hypotheses that were not rejected, including information about the size of the true effect and whether or not there is adequate evidence to 'accept' a null hypothesis as true. We suggest that (1) statistical power analyses be routinely incorporated into research planning efforts to increase their efficiency, (2) confidence intervals be used in lieu of retrospective power analyses for null hypotheses that were not rejected to assess the likely size of the true effect, (3) minimum biologically significant effect sizes be used for all power analyses, and (4) if retrospective power estimates are to be reported, then the I?-level, effect sizes, and sample sizes used in calculations must also be reported.
Exhaled isoprene for monitoring recovery from acute hypoxic stress.
Harshman, Sean W; Geier, Brian A; Qualley, Anthony V; Drummond, Leslie A; Flory, Laura E; Fan, Maomian; Pitsch, Rhonda L; Grigsby, Claude C; Phillips, Jeffrey B; Martin, Jennifer A
2017-11-29
Hypoxia-like incidents in-flight have increased over the past decade causing severe safety concerns across the aviation community. As a result, the need to monitor flight crews in real-time for the onset of hypoxic conditions is paramount for continued aeronautical safety. Here, hypoxic events were simulated in the laboratory via a reduced oxygen-breathing device to determine the effect of recovery gas oxygen concentration (21% and 100%) on exhaled breath volatile organic compound composition. Data from samples collected both serially (throughout the exposure), prior to, and following exposures yielded 326 statistically significant features, 203 of which were unique. Of those, 72 features were tentatively identified while 51 were verified with authentic standards. A comparison of samples collected serially between recovery and hypoxia time points shows a statistically significant reduction in exhaled breath isoprene (2-methyl-1,3-butadiene, log 2 FC -0.399, p = 0.005, FDR = 0.034, q = 0.033), however no significant difference in isoprene abundance was observed when comparing recovery gases (21% or 100% O 2 , p = 0.152). Furthermore, examination of pre-/post-exposure 1 l bag breath samples illustrate an overall increase in exhaled isoprene abundance post-exposure (log 2 FC 0.393, p = 0.005, FDR = 0.094, q = 0.033) but again no significant difference between recovery gas (21% and 100%, p = 0.798) was observed. A statistically significant difference in trend was observed between isoprene abundance and recovery gases O 2 concentration when plotted against minimum oxygen saturation (p = 0.0419 100% O 2 , p = 0.7034 21% O 2 ). Collectively, these results suggest exhaled isoprene is dynamic in the laboratory ROBD setup and additional experimentation will be required to fully understand the dynamics of isoprene in response to acute hypoxic stress.
[Epidemiological study of dental and facial asymmetries in a sample of preschool subjects].
Vitale, Marina Consuelo; Barbieri, Federica; Ricotta, Riccardo; Arpesella, Marisa; Emanuelli, Maria Teresa
2015-01-01
to identify the typologies of facial and dental asymmetries in a sample of children aged between 3 and 6 years and to correlate these asymmetries with possible morphological and functional situations. cross-sectional observational study. sample of 95 subjects aged between 3 and 6 years. Clinical data were collected in 10 sessions conducted during school hours in April 2013 by a doctor of Dentistry at two preschools in the city of Sanremo (Liguria Region, Northern Italy) and a kindergarten in the city of Pavia (Lombardy Region, Northern Italy). To collect the data, a weighted clinical questionnaire was used. presence and type of bad habit, type of breathing, presence and type of facial asymmetry, dental formula, presence of diastema, presence and type of occlusal asymmetries, presence and type of dental malocclusions. analysed sample consisted of 53.7% (51/95) of males and 46.3 % (44/95) females; the mean age was 4.3 ± 0.9 years. Most frequent facial asymmetry is orbits asymmetry (35%, 33/95); dental malocclusions are detected in 70%(67/95) of cases. High percentage of subjects (69.5%, 66/95) presents displacement between superior dental midline (SDM) and inferior dental midline (IDM). Several statistically significant associations are observed: in particular, asymmetry of molar ratios is linked to asymmetry of the cheekbones and displacement of the SDM; facial midline has statistical association with asymmetry of the cheekbones (p <0.001). the results of this study agree with scientific literature, in particular as regards the prevalence of compromising habits observed and the close correlation between: the presence of dental malocclusions and the presence of compromising habits, the presence of dental malocclusions and the presence of oral breathing.
NASA Astrophysics Data System (ADS)
Vogelmann, A. M.; Gustafson, W. I., Jr.; Toto, T.; Endo, S.; Cheng, X.; Li, Z.; Xiao, H.
2015-12-01
The Department of Energy's Atmospheric Radiation Measurement (ARM) Climate Research Facilities' Large-Eddy Simulation (LES) ARM Symbiotic Simulation and Observation (LASSO) Workflow is currently being designed to provide output from routine LES to complement its extensive observations. The modeling portion of the LASSO workflow is presented by Gustafson et al., which will initially focus on shallow convection over the ARM megasite in Oklahoma, USA. This presentation describes how the LES output will be combined with observations to construct multi-dimensional and dynamically consistent "data cubes", aimed at providing the best description of the atmospheric state for use in analyses by the community. The megasite observations are used to constrain large-eddy simulations that provide a complete spatial and temporal coverage of observables and, further, the simulations also provide information on processes that cannot be observed. Statistical comparisons of model output with their observables are used to assess the quality of a given simulated realization and its associated uncertainties. A data cube is a model-observation package that provides: (1) metrics of model-observation statistical summaries to assess the simulations and the ensemble spread; (2) statistical summaries of additional model property output that cannot be or are very difficult to observe; and (3) snapshots of the 4-D simulated fields from the integration period. Searchable metrics are provided that characterize the general atmospheric state to assist users in finding cases of interest, such as categorization of daily weather conditions and their specific attributes. The data cubes will be accompanied by tools designed for easy access to cube contents from within the ARM archive and externally, the ability to compare multiple data streams within an event as well as across events, and the ability to use common grids and time sampling, where appropriate.
NASA Astrophysics Data System (ADS)
Duari, Debiprosad; Narlikar, Jayant V.
This paper examines, in the light of the available data, the hypothesis that the heavy element absorption line systems in the spectra of QSOs originate through en-route absorption by intervening galaxies, halos etc. Several statistical tests are applied in two different ways to compare the predictions of the intervening galaxies hypothesis (IGH) with actual observations. The database is taken from a recent 1991 compilation of absorption line systems by Junkkarinen, Hewitt and Burbidge. Although, prima facie, a considerable gap is found between the predictions of the intervening galaxies hypothesis and the actual observations despite inclusion of any effects of clustering and some likely selection effects, the gap narrows after invoking evolution in the number density of absorbers and allowing for the incompleteness and inhomogeneity of samples examined. On the latter count the gap might be bridgeable by stretching the parameters of the theory. It is concluded that although the intervening galaxies hypothesis is a possible natural explanation to account for the absorption line systems and may in fact do so in several cases, it seems too simplistic to be able to account for all the available data. It is further stressed that the statistical techniques described here will be useful for future studies of complete and homogenous samples with a view to deciding the extent of applicability of the IGH.
Trends in bromide wet deposition concentrations in the contiguous United States, 2001-2016.
Wetherbee, Gregory A; Lehmann, Christopher M B; Kerschner, Brian M; Ludtke, Amy S; Green, Lee A; Rhodes, Mark F
2018-02-01
Bromide (Br - ) and other solute concentration data from wet deposition samples collected and analyzed by the National Atmospheric Deposition Program (NADP) from 2001 to 2016, were statistically analyzed for trends both geographically and temporally by precipitation type. Analysis was limited to NADP sites in the contiguous 48 United States. The Br - concentrations for this time period had a high number of values censored at the detection limits with greater than 86 percent of sample concentrations below analytical detection. Bromide was more frequently detected at NADP sites in coastal regions. Analysis using specialized statistical techniques for censored data revealed that Br - concentrations varied by precipitation type with higher concentrations usually observed in liquid versus precipitation containing snow. Negative temporal trends in Br - wet deposition concentrations were observed at a majority of NADP sites; approximately 25 percent of these trend values were statistically significant at less than 0.05 to 0.10 significance levels. Potential causes for the negative trends were explored, including annual and seasonal changes in precipitation depth, reduced emissions of methyl bromide (CH 3 Br) from coastal wetlands, and declining industrial use of bromine compounds. The results indicate that Br - in non-coastal wet-deposition comes mainly from long-range transport, not local sources. Correlations between Br - , chloride, and nitrate concentrations also were evaluated. Published by Elsevier Ltd.
Eisinger, Stephen W; Schwartz, Matthew; Dam, Lisa; Riedel, Stefan
2013-09-01
The stability of urine specimens submitted for culture remains a challenge for many laboratories because of delays in specimen transport. We evaluated the usefulness of BD Vacutainer Plus Urine C&S Preservative Tube in ensuring specimen stability. Clinical urine specimens collected in sterile collection cups (n = 110) were plated onto sheep blood and MacConkey agar following standard laboratory procedures guidelines. Thereafter, specimens were divided into 3 storage conditions: nonpreservative, refrigerated; nonpreservative, room temperature (RT); BD Vacutainer Plus Urine C&S Preservative Tube, RT. For each sample type, additional cultures were set up at 2, 4, 24, and 48 hours. Initially, 18 specimens had no growth, 32 showed mixed skin flora, and 60 yielded at least 1 uropathogen. Increased colony counts of uropathogens were observed for nonpreserved urine samples stored at RT; these changes were statistically significant. Minor differences between refrigerated urine samples and BD Vacutainer Plus Urine C&S Preservative Tube samples were seen but were not statistically significant. The use of preservative-containing collection tubes is desirable to ensure specimen stability when prompt processing or refrigeration is not feasible.
Microjets in the penumbra of a sunspot
NASA Astrophysics Data System (ADS)
Drews, Ainar; Rouppe van der Voort, Luc
2017-06-01
Context. Penumbral microjets (PMJs) are short-lived jets found in the penumbra of sunspots, first observed in wide-band Ca II H line observations as localized brightenings, and are thought to be caused by magnetic reconnection. Earlier work on PMJs has focused on smaller samples of by-eye selected events and case studies. Aims: It is our goal to present an automated study of a large sample of PMJs to place the basic statistics of PMJs on a sure footing and to study the PMJ Ca II 8542 Å spectral profile in detail. Methods: High spatial resolution and spectrally well-sampled observations in the Ca II 8542 Å line obtained from the Swedish 1-m Solar Telescope (SST) were reduced by a principle component analysis and subsequently used in the automated detection of PMJs using the simple machine learning algorithm k-nearest neighbour. PMJ detections were verified with co-temporal Ca II H line observations. Results: We find a total of 453 tracked PMJ events, 4253 PMJs detections tallied over all timeframes, and a detection rate of 21 events per timestep. From these, an average length, width and lifetime of 640 km, 210 km and 90 s are obtained. The average PMJ Ca II 8542 Å line profile is characterized by enhanced inner wings, often in the form of one or two distinct peaks, and a brighter line core as compared to the quiet-Sun average. Average blue and red peak positions are determined at - 10.4 km s-1 and + 10.2 km s-1 offsets from the Ca II 8542 Å line core. We find several clusters of PMJ hot-spots within the sunspot penumbra, in which PMJ events occur in the same general area repeatedly over time. Conclusions: Our results indicate smaller average PMJs sizes and longer lifetimes compared to previously published values, but with statistics still in the same orders of magnitude. The investigation and analysis of the PMJ line profiles strengthens the proposed heating of PMJs to transition region temperatures. The presented statistics on PMJs form a solid basis for future investigations and numerical modelling of PMJs.
Multivariate Statistical Inference of Lightning Occurrence, and Using Lightning Observations
NASA Technical Reports Server (NTRS)
Boccippio, Dennis
2004-01-01
Two classes of multivariate statistical inference using TRMM Lightning Imaging Sensor, Precipitation Radar, and Microwave Imager observation are studied, using nonlinear classification neural networks as inferential tools. The very large and globally representative data sample provided by TRMM allows both training and validation (without overfitting) of neural networks with many degrees of freedom. In the first study, the flashing / or flashing condition of storm complexes is diagnosed using radar, passive microwave and/or environmental observations as neural network inputs. The diagnostic skill of these simple lightning/no-lightning classifiers can be quite high, over land (above 80% Probability of Detection; below 20% False Alarm Rate). In the second, passive microwave and lightning observations are used to diagnose radar reflectivity vertical structure. A priori diagnosis of hydrometeor vertical structure is highly important for improved rainfall retrieval from either orbital radars (e.g., the future Global Precipitation Mission "mothership") or radiometers (e.g., operational SSM/I and future Global Precipitation Mission passive microwave constellation platforms), we explore the incremental benefit to such diagnosis provided by lightning observations.
NASA Technical Reports Server (NTRS)
Bunting, Charles F.; Yu, Shih-Pin
2006-01-01
This paper emphasizes the application of numerical methods to explore the ideas related to shielding effectiveness from a statistical view. An empty rectangular box is examined using a hybrid modal/moment method. The basic computational method is presented followed by the results for single- and multiple observation points within the over-moded empty structure. The statistics of the field are obtained by using frequency stirring, borrowed from the ideas connected with reverberation chamber techniques, and extends the ideas of shielding effectiveness well into the multiple resonance regions. The study presented in this paper will address the average shielding effectiveness over a broad spatial sample within the enclosure as the frequency is varied.
Physical and dynamical studies of meteors. Meteor-fragmentation and stream-distribution studies
NASA Technical Reports Server (NTRS)
Sekanina, Z.; Southworth, R. B.
1975-01-01
Population parameters of 275 streams including 20 additional streams in the synoptic-year sample were found by a computer technique. Some 16 percent of the sample is in these streams. Four meteor streams that have close orbital resemblance to Adonis cannot be positively identified as meteors ejected by Adonis within the last 12000 years. Ceplecha's discrete levels of meteor height are not evident in radar meteors. The spread of meteoroid fragments along their common trajectory was computed for most of the observed radar meteors. There is an unexpected relationship between spread and velocity that perhaps conceals relationships between fragmentation and orbits; a theoretical treatment will be necessary to resolve these relationships. Revised unbiased statistics of synoptic-year orbits are presented, together with parallel statistics for the 1961 to 1965 radar meteor orbits.
A methodological analysis of chaplaincy research: 2000-2009.
Galek, Kathleen; Flannelly, Kevin J; Jankowski, Katherine R B; Handzo, George F
2011-01-01
The present article presents a comprehensive review and analysis of quantitative research conducted in the United States on chaplaincy and closely related topics published between 2000 and 2009. A combined search strategy identified 49 quantitative studies in 13 journals. The analysis focuses on the methodological sophistication of the studies, compared to earlier research on chaplaincy and pastoral care. Cross-sectional surveys of convenience samples still dominate the field, but sample sizes have increased somewhat over the past three decades. Reporting of the validity and reliability of measures continues to be low, although reporting of response rates has improved. Improvements in the use of inferential statistics and statistical controls were also observed, compared to previous research. The authors conclude that more experimental research is needed on chaplaincy, along with an increased use of hypothesis testing, regardless of the research designs that are used.
Mirjalali, H; Abbasi, M R; Naderi, N; Hasani, Z; Mirsamadi, E S; Stensvold, C R; Balaii, H; Asadzadeh Aghdaei, H; Zali, M R
2017-12-01
Blastocystis is a single-celled intestinal parasite commonly found in humans and a broad range of animals all over the world. In humans, its role in health and disease remains unsettled. The aim of our study was to investigate the distribution of Blastocystis and Blastocystis subtypes (ST) in patients with inflammatory bowel disease (IBD) and control subjects. A total of 71 stool samples were collected from IBD patients, 69 and 2 of whom had ulcerative colitis (UC) and Crohn's Disease (CD), respectively. Moreover, 166 stool samples from healthy subjects were included as control samples. All stool samples were cultivated, and 550-bp fragments of the small subunit ribosomal RNA gene was amplified from Blastocystis-positive cultures. All PCR-positive samples were sequenced. Blastocystis was observed in 9 (12.67%) and 35 (21.1%) IBD patients and healthy controls, respectively. There was no statistically significant correlation between IBD and presence of Blastocystis (P = 0.147). There was a statistically significant correlation between age and Blastocystis colonization in the IBD group (P < 0.05), but not among healthy controls. No significant correlation between gender and colonization was observed. ST1 and ST3 were obtained from 1 (12.5%) and 7 (87.5%) IBD patients, respectively, while in the healthy control group, subtypes 1, 2, and 3 were found in 14 (40%), 12 (34.28%), and 9 (25.72%), respectively. Phylogenetic analysis showed no variation in the distribution of subtypes nor intra-subtype genetic diversity between samples acquired from IBD patients and healthy controls. This study showed a trend towards a lower prevalence of Blastocystis in IBD patients than in control subjects. ST3 sequences isolated from IBD patients and control individuals did not appear to differ genetically.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Piepel, Gregory F.; Matzke, Brett D.; Sego, Landon H.
2013-04-27
This report discusses the methodology, formulas, and inputs needed to make characterization and clearance decisions for Bacillus anthracis-contaminated and uncontaminated (or decontaminated) areas using a statistical sampling approach. Specifically, the report includes the methods and formulas for calculating the • number of samples required to achieve a specified confidence in characterization and clearance decisions • confidence in making characterization and clearance decisions for a specified number of samples for two common statistically based environmental sampling approaches. In particular, the report addresses an issue raised by the Government Accountability Office by providing methods and formulas to calculate the confidence that amore » decision area is uncontaminated (or successfully decontaminated) if all samples collected according to a statistical sampling approach have negative results. Key to addressing this topic is the probability that an individual sample result is a false negative, which is commonly referred to as the false negative rate (FNR). The two statistical sampling approaches currently discussed in this report are 1) hotspot sampling to detect small isolated contaminated locations during the characterization phase, and 2) combined judgment and random (CJR) sampling during the clearance phase. Typically if contamination is widely distributed in a decision area, it will be detectable via judgment sampling during the characterization phrase. Hotspot sampling is appropriate for characterization situations where contamination is not widely distributed and may not be detected by judgment sampling. CJR sampling is appropriate during the clearance phase when it is desired to augment judgment samples with statistical (random) samples. The hotspot and CJR statistical sampling approaches are discussed in the report for four situations: 1. qualitative data (detect and non-detect) when the FNR = 0 or when using statistical sampling methods that account for FNR > 0 2. qualitative data when the FNR > 0 but statistical sampling methods are used that assume the FNR = 0 3. quantitative data (e.g., contaminant concentrations expressed as CFU/cm2) when the FNR = 0 or when using statistical sampling methods that account for FNR > 0 4. quantitative data when the FNR > 0 but statistical sampling methods are used that assume the FNR = 0. For Situation 2, the hotspot sampling approach provides for stating with Z% confidence that a hotspot of specified shape and size with detectable contamination will be found. Also for Situation 2, the CJR approach provides for stating with X% confidence that at least Y% of the decision area does not contain detectable contamination. Forms of these statements for the other three situations are discussed in Section 2.2. Statistical methods that account for FNR > 0 currently only exist for the hotspot sampling approach with qualitative data (or quantitative data converted to qualitative data). This report documents the current status of methods and formulas for the hotspot and CJR sampling approaches. Limitations of these methods are identified. Extensions of the methods that are applicable when FNR = 0 to account for FNR > 0, or to address other limitations, will be documented in future revisions of this report if future funding supports the development of such extensions. For quantitative data, this report also presents statistical methods and formulas for 1. quantifying the uncertainty in measured sample results 2. estimating the true surface concentration corresponding to a surface sample 3. quantifying the uncertainty of the estimate of the true surface concentration. All of the methods and formulas discussed in the report were applied to example situations to illustrate application of the methods and interpretation of the results.« less
Errors in radial velocity variance from Doppler wind lidar
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang, H.; Barthelmie, R. J.; Doubrawa, P.
A high-fidelity lidar turbulence measurement technique relies on accurate estimates of radial velocity variance that are subject to both systematic and random errors determined by the autocorrelation function of radial velocity, the sampling rate, and the sampling duration. Our paper quantifies the effect of the volumetric averaging in lidar radial velocity measurements on the autocorrelation function and the dependence of the systematic and random errors on the sampling duration, using both statistically simulated and observed data. For current-generation scanning lidars and sampling durations of about 30 min and longer, during which the stationarity assumption is valid for atmospheric flows, themore » systematic error is negligible but the random error exceeds about 10%.« less
Errors in radial velocity variance from Doppler wind lidar
Wang, H.; Barthelmie, R. J.; Doubrawa, P.; ...
2016-08-29
A high-fidelity lidar turbulence measurement technique relies on accurate estimates of radial velocity variance that are subject to both systematic and random errors determined by the autocorrelation function of radial velocity, the sampling rate, and the sampling duration. Our paper quantifies the effect of the volumetric averaging in lidar radial velocity measurements on the autocorrelation function and the dependence of the systematic and random errors on the sampling duration, using both statistically simulated and observed data. For current-generation scanning lidars and sampling durations of about 30 min and longer, during which the stationarity assumption is valid for atmospheric flows, themore » systematic error is negligible but the random error exceeds about 10%.« less
Observation of e+e-→ηJ/ψ at center-of-mass energy s=4.009GeV
NASA Astrophysics Data System (ADS)
Ablikim, M.; Achasov, M. N.; Ambrose, D. J.; An, F. F.; An, Q.; An, Z. H.; Bai, J. Z.; Ban, Y.; Becker, J.; Bennett, J. V.; Bertani, M.; Bian, J. M.; Boger, E.; Bondarenko, O.; Boyko, I.; Briere, R. A.; Bytev, V.; Cai, X.; Cakir, O.; Calcaterra, A.; Cao, G. F.; Cetin, S. A.; Chang, J. F.; Chelkov, G.; Chen, G.; Chen, H. S.; Chen, J. C.; Chen, M. L.; Chen, S. J.; Chen, Y. B.; Cheng, H. P.; Chu, Y. P.; Cronin-Hennessy, D.; Dai, H. L.; Dai, J. P.; Dedovich, D.; Deng, Z. Y.; Denig, A.; Denysenko, I.; Destefanis, M.; Ding, W. M.; Ding, Y.; Dong, L. Y.; Dong, M. Y.; Du, S. X.; Fang, J.; Fang, S. S.; Fava, L.; Feldbauer, F.; Feng, C. Q.; Ferroli, R. B.; Fu, C. D.; Fu, J. L.; Gao, Y.; Geng, C.; Goetzen, K.; Gong, W. X.; Gradl, W.; Greco, M.; Gu, M. H.; Gu, Y. T.; Guan, Y. H.; Guo, A. Q.; Guo, L. B.; Guo, Y. P.; Han, Y. L.; Harris, F. A.; He, K. L.; He, M.; He, Z. Y.; Held, T.; Heng, Y. K.; Hou, Z. L.; Hu, H. M.; Hu, J. F.; Hu, T.; Huang, G. M.; Huang, J. S.; Huang, X. T.; Huang, Y. P.; Hussain, T.; Ji, C. S.; Ji, Q.; Ji, X. B.; Ji, X. L.; Jiang, L. L.; Jiang, X. S.; Jiao, J. B.; Jiao, Z.; Jin, D. P.; Jin, S.; Jing, F. F.; Kalantar-Nayestanaki, N.; Kavatsyuk, M.; Kuehn, W.; Lai, W.; Lange, J. S.; Li, C. H.; Li, Cheng; Li, Cui; Li, D. M.; Li, F.; Li, G.; Li, H. B.; Li, J. C.; Li, K.; Li, Lei; Li, Q. J.; Li, S. L.; Li, W. D.; Li, W. G.; Li, X. L.; Li, X. N.; Li, X. Q.; Li, X. R.; Li, Z. B.; Liang, H.; Liang, Y. F.; Liang, Y. T.; Liao, G. R.; Liao, X. T.; Liu, B. J.; Liu, C. L.; Liu, C. X.; Liu, C. Y.; Liu, F. H.; Liu, Fang; Liu, Feng; Liu, H.; Liu, H. B.; Liu, H. H.; Liu, H. M.; Liu, H. W.; Liu, J. P.; Liu, K. Y.; Liu, Kai; Liu, P. L.; Liu, Q.; Liu, S. B.; Liu, X.; Liu, X. H.; Liu, Y. B.; Liu, Z. A.; Liu, Zhiqiang; Liu, Zhiqing; Loehner, H.; Lu, G. R.; Lu, H. J.; Lu, J. G.; Lu, Q. W.; Lu, X. R.; Lu, Y. P.; Luo, C. L.; Luo, M. X.; Luo, T.; Luo, X. L.; Lv, M.; Ma, C. L.; Ma, F. C.; Ma, H. L.; Ma, Q. M.; Ma, S.; Ma, T.; Ma, X. Y.; Ma, Y.; Maas, F. E.; Maggiora, M.; Malik, Q. A.; Mao, Y. J.; Mao, Z. P.; Messchendorp, J. G.; Min, J.; Min, T. J.; Mitchell, R. E.; Mo, X. H.; Morales, C. Morales; Motzko, C.; Muchnoi, N. Yu.; Muramatsu, H.; Nefedov, Y.; Nicholson, C.; Nikolaev, I. B.; Ning, Z.; Olsen, S. L.; Ouyang, Q.; Pacetti, S.; Park, J. W.; Pelizaeus, M.; Peng, H. P.; Peters, K.; Ping, J. L.; Ping, R. G.; Poling, R.; Prencipe, E.; Qi, M.; Qian, S.; Qiao, C. F.; Qin, X. S.; Qin, Y.; Qin, Z. H.; Qiu, J. F.; Rashid, K. H.; Rong, G.; Ruan, X. D.; Sarantsev, A.; Schaefer, B. D.; Schulze, J.; Shao, M.; Shen, C. P.; Shen, X. Y.; Sheng, H. Y.; Shepherd, M. R.; Song, W. M.; Song, X. Y.; Spataro, S.; Spruck, B.; Sun, D. H.; Sun, G. X.; Sun, J. F.; Sun, S. S.; Sun, Y. J.; Sun, Y. Z.; Sun, Z. J.; Sun, Z. T.; Tang, C. J.; Tang, X.; Tapan, I.; Thorndike, E. H.; Toth, D.; Ullrich, M.; Varner, G. S.; Wang, B.; Wang, B. Q.; Wang, K.; Wang, L. L.; Wang, L. S.; Wang, M.; Wang, P.; Wang, P. L.; Wang, Q.; Wang, Q. J.; Wang, S. G.; Wang, X. L.; Wang, Y. D.; Wang, Y. F.; Wang, Y. Q.; Wang, Z.; Wang, Z. G.; Wang, Z. Y.; Wei, D. H.; Weidenkaff, P.; Wen, Q. G.; Wen, S. P.; Werner, M.; Wiedner, U.; Wu, L. H.; Wu, N.; Wu, S. X.; Wu, W.; Wu, Z.; Xia, L. G.; Xiao, Z. J.; Xie, Y. G.; Xiu, Q. L.; Xu, G. F.; Xu, G. M.; Xu, H.; Xu, Q. J.; Xu, X. P.; Xu, Z. R.; Xue, F.; Xue, Z.; Yan, L.; Yan, W. B.; Yan, Y. H.; Yang, H. X.; Yang, Y.; Yang, Y. X.; Ye, H.; Ye, M.; Ye, M. H.; Yu, B. X.; Yu, C. X.; Yu, J. S.; Yu, S. P.; Yuan, C. Z.; Yuan, Y.; Zafar, A. A.; Zallo, A.; Zeng, Y.; Zhang, B. X.; Zhang, B. Y.; Zhang, C. C.; Zhang, D. H.; Zhang, H. H.; Zhang, H. Y.; Zhang, J. Q.; Zhang, J. W.; Zhang, J. Y.; Zhang, J. Z.; Zhang, S. H.; Zhang, X. J.; Zhang, X. Y.; Zhang, Y.; Zhang, Y. H.; Zhang, Y. S.; Zhang, Z. P.; Zhang, Z. Y.; Zhao, G.; Zhao, H. S.; Zhao, J. W.; Zhao, K. X.; Zhao, Lei; Zhao, Ling; Zhao, M. G.; Zhao, Q.; Zhao, S. J.; Zhao, T. C.; Zhao, X. H.; Zhao, Y. B.; Zhao, Z. G.; Zhemchugov, A.; Zheng, B.; Zheng, J. P.; Zheng, Y. H.; Zhong, B.; Zhong, J.; Zhou, L.; Zhou, X. K.; Zhou, X. R.; Zhu, C.; Zhu, K.; Zhu, K. J.; Zhu, S. H.; Zhu, X. L.; Zhu, X. W.; Zhu, Y. C.; Zhu, Y. M.; Zhu, Y. S.; Zhu, Z. A.; Zhuang, J.; Zou, B. S.; Zou, J. H.
2012-10-01
Using a 478pb-1 data sample collected with the BESIII detector operating at the Beijing Electron Positron Collider storage ring at a center-of-mass energy of s=4.009GeV, the production of e+e-→ηJ/ψ is observed for the first time with a statistical significance of greater than 10σ. The Born cross section is measured to be (32.1±2.8±1.3)pb, where the first error is statistical and the second systematic. Assuming the ηJ/ψ signal is from a hadronic transition of the ψ(4040), the fractional transition rate is determined to be B(ψ(4040)→ηJ/ψ)=(5.2±0.5±0.2±0.5)×10-3, where the first, second, and third errors are statistical, systematic, and the uncertainty from the ψ(4040) resonant parameters, respectively. The production of e+e-→π0J/ψ is searched for, but no significant signal is observed, and B(ψ(4040)→π0J/ψ)<2.8×10-4 is obtained at the 90% confidence level.
Imprints of dynamical interactions on brown dwarf pairing statistics and kinematics
NASA Astrophysics Data System (ADS)
Sterzik, M. F.; Durisen, R. H.
2003-03-01
We present statistically robust predictions of brown dwarf properties arising from dynamical interactions during their early evolution in small clusters. Our conclusions are based on numerical calculations of the internal cluster dynamics as well as on Monte-Carlo models. Accounting for recent observational constraints on the sub-stellar mass function and initial properties in fragmenting star forming clumps, we derive multiplicity fractions, mass ratios, separation distributions, and velocity dispersions. We compare them with observations of brown dwarfs in the field and in young clusters. Observed brown dwarf companion fractions around 15 +/- 7% for very low-mass stars as reported recently by Close et al. (\\cite{CSFB03}) are consistent with certain dynamical decay models. A significantly smaller mean separation distribution for brown dwarf binaries than for binaries of late-type stars can be explained by similar specific energy at the time of cluster formation for all cluster masses. Due to their higher velocity dispersions, brown-dwarfs and low-mass single stars will undergo time-dependent spatial segregation from higher-mass stars and multiple systems. This will cause mass functions and binary statistics in star forming regions to vary with the age of the region and the volume sampled.
Danish Passage Graves, "Spring/Summer/Fall full Moons" and Lunar Standstills
NASA Astrophysics Data System (ADS)
Clausen, Claus Jørgen
2015-05-01
The author proposes and discusses a model for azimuth distribution which involves the criterion of a 'spring full moon' (or a 'fall full moon') proposed by Marciano Da Silva (Da Silva 2004). The model is based on elements of the rising pattern of the summer full moon combined with directions pointing towards full moonrises which occur immediately prior to lunar standstill eclipses and directions aimed at the points at which these eclipses begin. An observed sample of 153 directions has been compared with the proposed model, which has been named the lunar 'season pointer'. Statistical tests show that the model fits well with the observed sample within the azimuth interval of 54.5° to 156.5°. The conclusion made is that at least the 'season pointer' section of the model used could very well explain the observed distribution.
Observation of [Formula: see text] and [Formula: see text] decays.
Aaij, R; Adeva, B; Adinolfi, M; Ajaltouni, Z; Akar, S; Albrecht, J; Alessio, F; Alexander, M; Ali, S; Alkhazov, G; Alvarez Cartelle, P; Alves, A A; Amato, S; Amerio, S; Amhis, Y; An, L; Anderlini, L; Andreassi, G; Andreotti, M; Andrews, J E; Appleby, R B; Archilli, F; d'Argent, P; Arnau Romeu, J; Artamonov, A; Artuso, M; Aslanides, E; Auriemma, G; Baalouch, M; Babuschkin, I; Bachmann, S; Back, J J; Badalov, A; Baesso, C; Baker, S; Baldini, W; Barlow, R J; Barschel, C; Barsuk, S; Barter, W; Baszczyk, M; Batozskaya, V; Batsukh, B; Battista, V; Bay, A; Beaucourt, L; Beddow, J; Bedeschi, F; Bediaga, I; Bel, L J; Bellee, V; Belloli, N; Belous, K; Belyaev, I; Ben-Haim, E; Bencivenni, G; Benson, S; Benton, J; Berezhnoy, A; Bernet, R; Bertolin, A; Betancourt, C; Betti, F; Bettler, M-O; van Beuzekom, M; Bezshyiko, Ia; Bifani, S; Billoir, P; Bird, T; Birnkraut, A; Bitadze, A; Bizzeti, A; Blake, T; Blanc, F; Blouw, J; Blusk, S; Bocci, V; Boettcher, T; Bondar, A; Bondar, N; Bonivento, W; Bordyuzhin, I; Borgheresi, A; Borghi, S; Borisyak, M; Borsato, M; Bossu, F; Boubdir, M; Bowcock, T J V; Bowen, E; Bozzi, C; Braun, S; Britsch, M; Britton, T; Brodzicka, J; Buchanan, E; Burr, C; Bursche, A; Buytaert, J; Cadeddu, S; Calabrese, R; Calvi, M; Calvo Gomez, M; Camboni, A; Campana, P; Campora Perez, D H; Capriotti, L; Carbone, A; Carboni, G; Cardinale, R; Cardini, A; Carniti, P; Carson, L; Carvalho Akiba, K; Casse, G; Cassina, L; Castillo Garcia, L; Cattaneo, M; Cauet, Ch; Cavallero, G; Cenci, R; Charles, M; Charpentier, Ph; Chatzikonstantinidis, G; Chefdeville, M; Chen, S; Cheung, S-F; Chobanova, V; Chrzaszcz, M; Cid Vidal, X; Ciezarek, G; Clarke, P E L; Clemencic, M; Cliff, H V; Closier, J; Coco, V; Cogan, J; Cogneras, E; Cogoni, V; Cojocariu, L; Collazuol, G; Collins, P; Comerma-Montells, A; Contu, A; Cook, A; Coombs, G; Coquereau, S; Corti, G; Corvo, M; Costa Sobral, C M; Couturier, B; Cowan, G A; Craik, D C; Crocombe, A; Cruz Torres, M; Cunliffe, S; Currie, R; D'Ambrosio, C; Da Cunha Marinho, F; Dall'Occo, E; Dalseno, J; David, P N Y; Davis, A; De Aguiar Francisco, O; De Bruyn, K; De Capua, S; De Cian, M; De Miranda, J M; De Paula, L; De Serio, M; De Simone, P; Dean, C-T; Decamp, D; Deckenhoff, M; Del Buono, L; Demmer, M; Dendek, A; Derkach, D; Deschamps, O; Dettori, F; Dey, B; Di Canto, A; Dijkstra, H; Dordei, F; Dorigo, M; Dosil Suárez, A; Dovbnya, A; Dreimanis, K; Dufour, L; Dujany, G; Dungs, K; Durante, P; Dzhelyadin, R; Dziurda, A; Dzyuba, A; Déléage, N; Easo, S; Ebert, M; Egede, U; Egorychev, V; Eidelman, S; Eisenhardt, S; Eitschberger, U; Ekelhof, R; Eklund, L; Ely, S; Esen, S; Evans, H M; Evans, T; Falabella, A; Farley, N; Farry, S; Fay, R; Fazzini, D; Ferguson, D; Fernandez Prieto, A; Ferrari, F; Ferreira Rodrigues, F; Ferro-Luzzi, M; Filippov, S; Fini, R A; Fiore, M; Fiorini, M; Firlej, M; Fitzpatrick, C; Fiutowski, T; Fleuret, F; Fohl, K; Fontana, M; Fontanelli, F; Forshaw, D C; Forty, R; Franco Lima, V; Frank, M; Frei, C; Fu, J; Furfaro, E; Färber, C; Gallas Torreira, A; Galli, D; Gallorini, S; Gambetta, S; Gandelman, M; Gandini, P; Gao, Y; Garcia Martin, L M; García Pardiñas, J; Garra Tico, J; Garrido, L; Garsed, P J; Gascon, D; Gaspar, C; Gavardi, L; Gazzoni, G; Gerick, D; Gersabeck, E; Gersabeck, M; Gershon, T; Ghez, Ph; Gianì, S; Gibson, V; Girard, O G; Giubega, L; Gizdov, K; Gligorov, V V; Golubkov, D; Golutvin, A; Gomes, A; Gorelov, I V; Gotti, C; Govorkova, E; Grabalosa Gándara, M; Graciani Diaz, R; Granado Cardoso, L A; Graugés, E; Graverini, E; Graziani, G; Grecu, A; Griffith, P; Grillo, L; Gruberg Cazon, B R; Grünberg, O; Gushchin, E; Guz, Yu; Gys, T; Göbel, C; Hadavizadeh, T; Hadjivasiliou, C; Haefeli, G; Haen, C; Haines, S C; Hall, S; Hamilton, B; Han, X; Hansmann-Menzemer, S; Harnew, N; Harnew, S T; Harrison, J; Hatch, M; He, J; Head, T; Heister, A; Hennessy, K; Henrard, P; Henry, L; Hernando Morata, J A; van Herwijnen, E; Heß, M; Hicheur, A; Hill, D; Hombach, C; Hopchev, H; Hulsbergen, W; Humair, T; Hushchyn, M; Hussain, N; Hutchcroft, D; Idzik, M; Ilten, P; Jacobsson, R; Jaeger, A; Jalocha, J; Jans, E; Jawahery, A; Jiang, F; John, M; Johnson, D; Jones, C R; Joram, C; Jost, B; Jurik, N; Kandybei, S; Kanso, W; Karacson, M; Kariuki, J M; Karodia, S; Kecke, M; Kelsey, M; Kenyon, I R; Kenzie, M; Ketel, T; Khairullin, E; Khanji, B; Khurewathanakul, C; Kirn, T; Klaver, S; Klimaszewski, K; Koliiev, S; Kolpin, M; Komarov, I; Koopman, R F; Koppenburg, P; Kosmyntseva, A; Kozachuk, A; Kozeiha, M; Kravchuk, L; Kreplin, K; Kreps, M; Krokovny, P; Kruse, F; Krzemien, W; Kucewicz, W; Kucharczyk, M; Kudryavtsev, V; Kuonen, A K; Kurek, K; Kvaratskheliya, T; Lacarrere, D; Lafferty, G; Lai, A; Lanfranchi, G; Langenbruch, C; Latham, T; Lazzeroni, C; Le Gac, R; van Leerdam, J; Lees, J-P; Leflat, A; Lefrançois, J; Lefèvre, R; Lemaitre, F; Lemos Cid, E; Leroy, O; Lesiak, T; Leverington, B; Li, Y; Likhomanenko, T; Lindner, R; Linn, C; Lionetto, F; Liu, B; Liu, X; Loh, D; Longstaff, I; Lopes, J H; Lucchesi, D; Lucio Martinez, M; Luo, H; Lupato, A; Luppi, E; Lupton, O; Lusiani, A; Lyu, X; Machefert, F; Maciuc, F; Maev, O; Maguire, K; Malde, S; Malinin, A; Maltsev, T; Manca, G; Mancinelli, G; Manning, P; Maratas, J; Marchand, J F; Marconi, U; Marin Benito, C; Marino, P; Marks, J; Martellotti, G; Martin, M; Martinelli, M; Martinez Santos, D; Martinez Vidal, F; Martins Tostes, D; Massacrier, L M; Massafferri, A; Matev, R; Mathad, A; Mathe, Z; Matteuzzi, C; Mauri, A; Maurin, B; Mazurov, A; McCann, M; McCarthy, J; McNab, A; McNulty, R; Meadows, B; Meier, F; Meissner, M; Melnychuk, D; Merk, M; Merli, A; Michielin, E; Milanes, D A; Minard, M-N; Mitzel, D S; Mogini, A; Molina Rodriguez, J; Monroy, I A; Monteil, S; Morandin, M; Morawski, P; Mordà, A; Morello, M J; Moron, J; Morris, A B; Mountain, R; Muheim, F; Mulder, M; Mussini, M; Müller, D; Müller, J; Müller, K; Müller, V; Naik, P; Nakada, T; Nandakumar, R; Nandi, A; Nasteva, I; Needham, M; Neri, N; Neubert, S; Neufeld, N; Neuner, M; Nguyen, A D; Nguyen, T D; Nguyen-Mau, C; Nieswand, S; Niet, R; Nikitin, N; Nikodem, T; Novoselov, A; O'Hanlon, D P; Oblakowska-Mucha, A; Obraztsov, V; Ogilvy, S; Oldeman, R; Onderwater, C J G; Otalora Goicochea, J M; Otto, A; Owen, P; Oyanguren, A; Pais, P R; Palano, A; Palombo, F; Palutan, M; Panman, J; Papanestis, A; Pappagallo, M; Pappalardo, L L; Parker, W; Parkes, C; Passaleva, G; Pastore, A; Patel, G D; Patel, M; Patrignani, C; Pearce, A; Pellegrino, A; Penso, G; Pepe Altarelli, M; Perazzini, S; Perret, P; Pescatore, L; Petridis, K; Petrolini, A; Petrov, A; Petruzzo, M; Picatoste Olloqui, E; Pietrzyk, B; Pikies, M; Pinci, D; Pistone, A; Piucci, A; Playfer, S; Plo Casasus, M; Poikela, T; Polci, F; Poluektov, A; Polyakov, I; Polycarpo, E; Pomery, G J; Popov, A; Popov, D; Popovici, B; Poslavskii, S; Potterat, C; Price, E; Price, J D; Prisciandaro, J; Pritchard, A; Prouve, C; Pugatch, V; Puig Navarro, A; Punzi, G; Qian, W; Quagliani, R; Rachwal, B; Rademacker, J H; Rama, M; Ramos Pernas, M; Rangel, M S; Raniuk, I; Ratnikov, F; Raven, G; Redi, F; Reichert, S; Dos Reis, A C; Remon Alepuz, C; Renaudin, V; Ricciardi, S; Richards, S; Rihl, M; Rinnert, K; Rives Molina, V; Robbe, P; Rodrigues, A B; Rodrigues, E; Rodriguez Lopez, J A; Rodriguez Perez, P; Rogozhnikov, A; Roiser, S; Rollings, A; Romanovskiy, V; Romero Vidal, A; Ronayne, J W; Rotondo, M; Rudolph, M S; Ruf, T; Ruiz Valls, P; Saborido Silva, J J; Sadykhov, E; Sagidova, N; Saitta, B; Salustino Guimaraes, V; Sanchez Mayordomo, C; Sanmartin Sedes, B; Santacesaria, R; Santamarina Rios, C; Santimaria, M; Santovetti, E; Sarti, A; Satriano, C; Satta, A; Saunders, D M; Savrina, D; Schael, S; Schellenberg, M; Schiller, M; Schindler, H; Schlupp, M; Schmelling, M; Schmelzer, T; Schmidt, B; Schneider, O; Schopper, A; Schubert, K; Schubiger, M; Schune, M-H; Schwemmer, R; Sciascia, B; Sciubba, A; Semennikov, A; Sergi, A; Serra, N; Serrano, J; Sestini, L; Seyfert, P; Shapkin, M; Shapoval, I; Shcheglov, Y; Shears, T; Shekhtman, L; Shevchenko, V; Siddi, B G; Silva Coutinho, R; Silva de Oliveira, L; Simi, G; Simone, S; Sirendi, M; Skidmore, N; Skwarnicki, T; Smith, E; Smith, I T; Smith, J; Smith, M; Snoek, H; Sokoloff, M D; Soler, F J P; Souza De Paula, B; Spaan, B; Spradlin, P; Sridharan, S; Stagni, F; Stahl, M; Stahl, S; Stefko, P; Stefkova, S; Steinkamp, O; Stemmle, S; Stenyakin, O; Stevenson, S; Stoica, S; Stone, S; Storaci, B; Stracka, S; Straticiuc, M; Straumann, U; Sun, L; Sutcliffe, W; Swientek, K; Syropoulos, V; Szczekowski, M; Szumlak, T; T'Jampens, S; Tayduganov, A; Tekampe, T; Tellarini, G; Teubert, F; Thomas, E; van Tilburg, J; Tilley, M J; Tisserand, V; Tobin, M; Tolk, S; Tomassetti, L; Tonelli, D; Topp-Joergensen, S; Toriello, F; Tournefier, E; Tourneur, S; Trabelsi, K; Traill, M; Tran, M T; Tresch, M; Trisovic, A; Tsaregorodtsev, A; Tsopelas, P; Tully, A; Tuning, N; Ukleja, A; Ustyuzhanin, A; Uwer, U; Vacca, C; Vagnoni, V; Valassi, A; Valat, S; Valenti, G; Vallier, A; Vazquez Gomez, R; Vazquez Regueiro, P; Vecchi, S; van Veghel, M; Velthuis, J J; Veltri, M; Veneziano, G; Venkateswaran, A; Vernet, M; Vesterinen, M; Viaud, B; Vieira, D; Vieites Diaz, M; Viemann, H; Vilasis-Cardona, X; Vitti, M; Volkov, V; Vollhardt, A; Voneki, B; Vorobyev, A; Vorobyev, V; Voß, C; de Vries, J A; Vázquez Sierra, C; Waldi, R; Wallace, C; Wallace, R; Walsh, J; Wang, J; Ward, D R; Wark, H M; Watson, N K; Websdale, D; Weiden, A; Whitehead, M; Wicht, J; Wilkinson, G; Wilkinson, M; Williams, M; Williams, M P; Williams, M; Williams, T; Wilson, F F; Wimberley, J; Wishahi, J; Wislicki, W; Witek, M; Wormser, G; Wotton, S A; Wraight, K; Wyllie, K; Xie, Y; Xing, Z; Xu, Z; Yang, Z; Yin, H; Yu, J; Yuan, X; Yushchenko, O; Zarebski, K A; Zavertyaev, M; Zhang, L; Zhang, Y; Zhang, Y; Zhelezov, A; Zheng, Y; Zhokhov, A; Zhu, X; Zhukov, V; Zucchelli, S
2017-01-01
The decays [Formula: see text] and [Formula: see text] are observed for the first time using a data sample corresponding to an integrated luminosity of 3.0 fb[Formula: see text], collected by the LHCb experiment in proton-proton collisions at the centre-of-mass energies of 7 and 8[Formula: see text]. The branching fractions relative to that of [Formula: see text] are measured to be [Formula: see text]where the first uncertainties are statistical and the second are systematic.
An audit of the statistics and the comparison with the parameter in the population
NASA Astrophysics Data System (ADS)
Bujang, Mohamad Adam; Sa'at, Nadiah; Joys, A. Reena; Ali, Mariana Mohamad
2015-10-01
The sufficient sample size that is needed to closely estimate the statistics for particular parameters are use to be an issue. Although sample size might had been calculated referring to objective of the study, however, it is difficult to confirm whether the statistics are closed with the parameter for a particular population. All these while, guideline that uses a p-value less than 0.05 is widely used as inferential evidence. Therefore, this study had audited results that were analyzed from various sub sample and statistical analyses and had compared the results with the parameters in three different populations. Eight types of statistical analysis and eight sub samples for each statistical analysis were analyzed. Results found that the statistics were consistent and were closed to the parameters when the sample study covered at least 15% to 35% of population. Larger sample size is needed to estimate parameter that involve with categorical variables compared with numerical variables. Sample sizes with 300 to 500 are sufficient to estimate the parameters for medium size of population.
NASA Astrophysics Data System (ADS)
Deperas-Standylo, Joanna; Lee, Ryonfa; Nasonova, Elena; Ritter, Sylvia; Gudowska-Nowak, Ewa; Kac, M.; Smoluchowski, M.
Differences in the track structure of high LET (Linear Energy Transfer) particles are clearly visible on chromosomal level, in particular in the number of lesions produced by an ion traversal through a cell nucleus and in the distribution of aberrations among the cells. In the present study we focus on the effects of low energy C-and Cr-ions (<10 MeV/u) in comparison with high energy C-ions (90 MeV/u). For the experiments human lymphocytes were exposed to 9.5 MeV/u C-ions, 4.1 MeV/u Cr-ions or 90 MeV/u C-ions with LET values of 175 keV/µm, 3160 keV/µm and 29 keV/µm, respectively. Chromosome aberrations were measured at several post-irradiation sampling times (48, 60, 72 and 84h) in first cycle metaphases following Giemsa-staining. For 90 MeV/u C-ions, where the track radius is larger than the cell nucleus, the distribution of aberrations did not change significantly with sampling time and has been well described by Poisson statistics. In contrast, for low energy C-ions, where the track radius is smaller than the cell nucleus, distribution of aberration strongly deviates from uni-modal and displays two peaks representative for subpopulations of non-hit and hit cells, respectively. Following this pattern, also damage-dependent cell cycle delay was observed. At 48 h after irradiation a high number of undamaged and probably unhit cells was found to reach mitosis. This number of undamaged cells decreased further with sampling time, while the frequencies of cells carrying aberrations (1-11 per cell) were increasing. All distributions were found to conform a compound Poisson (Neyman-type A) statistics which allows estimating the average number of particle traversals through a cell nucleus and the average number of aberrations induced by one particle traversal. Similar response has also been observed at 48h after Cr-ion exposure. In this case, however, non-aberrant cells have been found to dominate in the population even at later sampling times and a low number of heavily damaged cells up to 24 aberrations have been detected. Accordingly, the distribution of aberrations in cells collected at >48 h could not be then described by a standard Neyman statistics. Obtained results suggest that most cells hit by more than one Cr-ion do not reach mitosis. This observation was confirmed by parallel measurements showing that Cr-ion exposure produces a high fraction of apoptotic cells.
Using venlafaxine to treat behavioral disorders in patients with autism spectrum disorder.
Carminati, Giuliana Galli; Gerber, Fabienne; Darbellay, Barbara; Kosel, Markus Mathaus; Deriaz, Nicolas; Chabert, Jocelyne; Fathi, Marc; Bertschy, Gilles; Ferrero, François; Carminati, Federico
2016-02-04
To test the efficacy of venlafaxine at a dose of 18.75 mg/day on the reduction of behavioral problems such as irritability and hyperactivity/noncompliance in patients with intellectual disabilities and autism spectrum disorder (ASD). Our secondary hypothesis was that the usual doses of zuclopenthixol and/or clonazepam would decrease in the venlafaxine-treated group. In a randomized double-blind study, we compared six patients who received venlafaxine along with their usual treatment (zuclopenthixol and/or clonazepam) with seven patients who received placebo plus usual care. Irritability, hyperactivity/noncompliance, and overall clinical improvement were measured after 2 and 8 weeks, using validated clinical scales. Univariate analyses showed that the symptom of irritability improved in the entire sample (p = 0.023 after 2 weeks, p = 0.061 at study endpoint), although no difference was observed between the venlafaxine and placebo groups. No significant decrease in hyperactivity/noncompliance was observed during the study. At the end of the study, global improvement was observed in 33% of participants treated with venlafaxine and in 71% of participants in the placebo group (p = 0.29). The study found that decreased cumulative doses of clonazepam and zuclopenthixol were required for the venlafaxine group. Multivariate analyses (principal component analyses) with at least three combinations of variables showed that the two populations could be clearly separated (p b 0.05). Moreover, in all cases, the venlafaxine population had lower values for the Aberrant Behavior Checklist (ABC), Behavior Problems Inventory (BPI), and levels of urea with respect to the placebo group. In one case, a reduction in the dosage of clonazepam was also suggested. For an additional set of variables (ABC factor 2, BPI frequency of aggressive behaviors, hematic ammonia at Day 28, and zuclopenthixol and clonazepam intake), the separation between the two samples was statistically significant as was the Bartlett's test, but the Kaiser–Meyer–Olkin Measure of Sampling Adequacy was below the accepted threshold. This set of variables showed a reduction in the cumulative intake of both zuclopenthixol and clonazepam. Despite the small sample sizes, this study documented a statistically significant effect of venlafaxine. Moreover, we showed that lower doses of zuclopenthixol and clonazepam were needed in the venlafaxine group, although this difference was not statistically significant. This was confirmed by multivariate analyses, where this difference reached statistical significance when using a combination of variables involving zuclopenthixol. Larger-scale studies are recommended to better investigate the effectiveness of venlafaxine treatment in patients with intellectual disabilities and ASD.
Lucijanic, Marko; Petrovecki, Mladen
2012-01-01
Analyzing events over time is often complicated by incomplete, or censored, observations. Special non-parametric statistical methods were developed to overcome difficulties in summarizing and comparing censored data. Life-table (actuarial) method and Kaplan-Meier method are described with an explanation of survival curves. For the didactic purpose authors prepared a workbook based on most widely used Kaplan-Meier method. It should help the reader understand how Kaplan-Meier method is conceptualized and how it can be used to obtain statistics and survival curves needed to completely describe a sample of patients. Log-rank test and hazard ratio are also discussed.
Gender discrimination and prediction on the basis of facial metric information.
Fellous, J M
1997-07-01
Horizontal and vertical facial measurements are statistically independent. Discriminant analysis shows that five of such normalized distances explain over 95% of the gender differences of "training" samples and predict the gender of 90% novel test faces exhibiting various facial expressions. The robustness of the method and its results are assessed. It is argued that these distances (termed fiducial) are compatible with those found experimentally by psychophysical and neurophysiological studies. In consequence, partial explanations for the effects observed in these experiments can be found in the intrinsic statistical nature of the facial stimuli used.
NASA Astrophysics Data System (ADS)
Pawlowsky-Glahn, Vera; Buccianti, Antonella
In the investigation of fluid samples of a volcanic system, collected during a given period of time, one of the main goals is to discover cause-effect relationships that allow us to explain changes in the chemical composition. They might be caused by physicochemical factors, such as temperature, pressure, or non-conservative behavior of some chemical constituents (addition or subtraction of material), among others. The presence of subgroups of observations showing different behavior is evidence of unusually complex situations, which might render even more difficult the analysis and interpretation of observed phenomena. These cases require appropriate statistical techniques as well as sound a priori hypothesis concerning underlying geological processes. The purpose of this article is to present the state of the art in the methodology for a better visualization of compositional data, as well as for detecting statistically significant sub-populations. The scheme of this article is to present first the application, and then the underlying methodology, with the aim of the first motivating the second. Thus, the first part has the goal to illustrate how to understand and interpret results, whereas the second is devoted to expose how to perform a study of this kind. The case study is related to the chemical composition of a fumarole of Vulcano Island (southern Italy), called F14. The volcanic activity at Vulcano Island is subject to a continuous program of geochemical surveillance from 1978 up to now and the large data set of observations contains the main chemical composition of volcanic gases as well as trace element concentrations in the condensates of fumarolic gases. Out of the complete set of measured components, the variables H2S, HF and As, determined in samples collected from 1978 to 1993 (As is not available in recent samples) are used to characterize two groups in the original population, which proved to be statistically distinct. The choice of the variables is motivated by the importance of investigating the behavior of well-known toxicity elements, which show, like As, a significant mobility under hydrothermal conditions. The statistical methodology used for this study is based on models devised for compositional data. They include (1) the perturbation approach for a better visualization; (2) cluster analysis to detect groups; (3) confidence regions for the center of the groups to obtain graphical evidence of differences between groups; and (4) tests of hypothesis about centers and covariance structures to obtain statistical evidence about differences between groups. The fact that only three components are used allows us to illustrate the results using ternary diagrams.
Smoking and Cancers: Case-Robust Analysis of a Classic Data Set
ERIC Educational Resources Information Center
Bentler, Peter M.; Satorra, Albert; Yuan, Ke-Hai
2009-01-01
A typical structural equation model is intended to reproduce the means, variances, and correlations or covariances among a set of variables based on parameter estimates of a highly restricted model. It is not widely appreciated that the sample statistics being modeled can be quite sensitive to outliers and influential observations, leading to bias…
ERIC Educational Resources Information Center
Lord, Frederic M.; Stocking, Martha
A general Computer program is described that will compute asymptotic standard errors and carry out significance tests for an endless variety of (standard and) nonstandard large-sample statistical problems, without requiring the statistician to derive asymptotic standard error formulas. The program assumes that the observations have a multinormal…
Sample Size Determination for Rasch Model Tests
ERIC Educational Resources Information Center
Draxler, Clemens
2010-01-01
This paper is concerned with supplementing statistical tests for the Rasch model so that additionally to the probability of the error of the first kind (Type I probability) the probability of the error of the second kind (Type II probability) can be controlled at a predetermined level by basing the test on the appropriate number of observations.…
Statistical characterization of carbon phenolic prepreg materials, volume 1
NASA Technical Reports Server (NTRS)
Beckley, Don A.; Stites, John, Jr.
1988-01-01
The objective was to characterize several lots of materials used for carbon/carbon and carbon/phenol product manufacture. Volume one is organized into testing categories based on raw material of product form. Each category contains a discussion of the sampling plan, comments and observations on each test method utilized, and a summary of the results obtained each category.
Bluhmki, Tobias; Bramlage, Peter; Volk, Michael; Kaltheuner, Matthias; Danne, Thomas; Rathmann, Wolfgang; Beyersmann, Jan
2017-02-01
Complex longitudinal sampling and the observational structure of patient registers in health services research are associated with methodological challenges regarding data management and statistical evaluation. We exemplify common pitfalls and want to stimulate discussions on the design, development, and deployment of future longitudinal patient registers and register-based studies. For illustrative purposes, we use data from the prospective, observational, German DIabetes Versorgungs-Evaluation register. One aim was to explore predictors for the initiation of a basal insulin supported therapy in patients with type 2 diabetes initially prescribed to glucose-lowering drugs alone. Major challenges are missing mortality information, time-dependent outcomes, delayed study entries, different follow-up times, and competing events. We show that time-to-event methodology is a valuable tool for improved statistical evaluation of register data and should be preferred to simple case-control approaches. Patient registers provide rich data sources for health services research. Analyses are accompanied with the trade-off between data availability, clinical plausibility, and statistical feasibility. Cox' proportional hazards model allows for the evaluation of the outcome-specific hazards, but prediction of outcome probabilities is compromised by missing mortality information. Copyright © 2016 Elsevier Inc. All rights reserved.
Statistics and classification of the microwave zebra patterns associated with solar flares
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tan, Baolin; Tan, Chengming; Zhang, Yin
2014-01-10
The microwave zebra pattern (ZP) is the most interesting, intriguing, and complex spectral structure frequently observed in solar flares. A comprehensive statistical study will certainly help us to understand the formation mechanism, which is not exactly clear now. This work presents a comprehensive statistical analysis of a big sample with 202 ZP events collected from observations at the Chinese Solar Broadband Radio Spectrometer at Huairou and the Ondŕejov Radiospectrograph in the Czech Republic at frequencies of 1.00-7.60 GHz from 2000 to 2013. After investigating the parameter properties of ZPs, such as the occurrence in flare phase, frequency range, polarization degree,more » duration, etc., we find that the variation of zebra stripe frequency separation with respect to frequency is the best indicator for a physical classification of ZPs. Microwave ZPs can be classified into three types: equidistant ZPs, variable-distant ZPs, and growing-distant ZPs, possibly corresponding to mechanisms of the Bernstein wave model, whistler wave model, and double plasma resonance model, respectively. This statistical classification may help us to clarify the controversies between the existing various theoretical models and understand the physical processes in the source regions.« less
The effect of rare variants on inflation of the test statistics in case-control analyses.
Pirie, Ailith; Wood, Angela; Lush, Michael; Tyrer, Jonathan; Pharoah, Paul D P
2015-02-20
The detection of bias due to cryptic population structure is an important step in the evaluation of findings of genetic association studies. The standard method of measuring this bias in a genetic association study is to compare the observed median association test statistic to the expected median test statistic. This ratio is inflated in the presence of cryptic population structure. However, inflation may also be caused by the properties of the association test itself particularly in the analysis of rare variants. We compared the properties of the three most commonly used association tests: the likelihood ratio test, the Wald test and the score test when testing rare variants for association using simulated data. We found evidence of inflation in the median test statistics of the likelihood ratio and score tests for tests of variants with less than 20 heterozygotes across the sample, regardless of the total sample size. The test statistics for the Wald test were under-inflated at the median for variants below the same minor allele frequency. In a genetic association study, if a substantial proportion of the genetic variants tested have rare minor allele frequencies, the properties of the association test may mask the presence or absence of bias due to population structure. The use of either the likelihood ratio test or the score test is likely to lead to inflation in the median test statistic in the absence of population structure. In contrast, the use of the Wald test is likely to result in under-inflation of the median test statistic which may mask the presence of population structure.
NASA Technical Reports Server (NTRS)
Lien, Guo-Yuan; Kalnay, Eugenia; Miyoshi, Takemasa; Huffman, George J.
2016-01-01
Assimilation of satellite precipitation data into numerical models presents several difficulties, with two of the most important being the non-Gaussian error distributions associated with precipitation, and large model and observation errors. As a result, improving the model forecast beyond a few hours by assimilating precipitation has been found to be difficult. To identify the challenges and propose practical solutions to assimilation of precipitation, statistics are calculated for global precipitation in a low-resolution NCEP Global Forecast System (GFS) model and the TRMM Multisatellite Precipitation Analysis (TMPA). The samples are constructed using the same model with the same forecast period, observation variables, and resolution as in the follow-on GFSTMPA precipitation assimilation experiments presented in the companion paper.The statistical results indicate that the T62 and T126 GFS models generally have positive bias in precipitation compared to the TMPA observations, and that the simulation of the marine stratocumulus precipitation is not realistic in the T62 GFS model. It is necessary to apply to precipitation either the commonly used logarithm transformation or the newly proposed Gaussian transformation to obtain a better relationship between the model and observational precipitation. When the Gaussian transformations are separately applied to the model and observational precipitation, they serve as a bias correction that corrects the amplitude-dependent biases. In addition, using a spatially andor temporally averaged precipitation variable, such as the 6-h accumulated precipitation, should be advantageous for precipitation assimilation.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Reiser, I; Lu, Z
2014-06-01
Purpose: Recently, task-based assessment of diagnostic CT systems has attracted much attention. Detection task performance can be estimated using human observers, or mathematical observer models. While most models are well established, considerable bias can be introduced when performance is estimated from a limited number of image samples. Thus, the purpose of this work was to assess the effect of sample size on bias and uncertainty of two channelized Hotelling observers and a template-matching observer. Methods: The image data used for this study consisted of 100 signal-present and 100 signal-absent regions-of-interest, which were extracted from CT slices. The experimental conditions includedmore » two signal sizes and five different x-ray beam current settings (mAs). Human observer performance for these images was determined in 2-alternative forced choice experiments. These data were provided by the Mayo clinic in Rochester, MN. Detection performance was estimated from three observer models, including channelized Hotelling observers (CHO) with Gabor or Laguerre-Gauss (LG) channels, and a template-matching observer (TM). Different sample sizes were generated by randomly selecting a subset of image pairs, (N=20,40,60,80). Observer performance was quantified as proportion of correct responses (PC). Bias was quantified as the relative difference of PC for 20 and 80 image pairs. Results: For n=100, all observer models predicted human performance across mAs and signal sizes. Bias was 23% for CHO (Gabor), 7% for CHO (LG), and 3% for TM. The relative standard deviation, σ(PC)/PC at N=20 was highest for the TM observer (11%) and lowest for the CHO (Gabor) observer (5%). Conclusion: In order to make image quality assessment feasible in the clinical practice, a statistically efficient observer model, that can predict performance from few samples, is needed. Our results identified two observer models that may be suited for this task.« less
Eye-gaze determination of user intent at the computer interface
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goldberg, J.H.; Schryver, J.C.
1993-12-31
Determination of user intent at the computer interface through eye-gaze monitoring can significantly aid applications for the disabled, as well as telerobotics and process control interfaces. Whereas current eye-gaze control applications are limited to object selection and x/y gazepoint tracking, a methodology was developed here to discriminate a more abstract interface operation: zooming-in or out. This methodology first collects samples of eve-gaze location looking at controlled stimuli, at 30 Hz, just prior to a user`s decision to zoom. The sample is broken into data frames, or temporal snapshots. Within a data frame, all spatial samples are connected into a minimummore » spanning tree, then clustered, according to user defined parameters. Each cluster is mapped to one in the prior data frame, and statistics are computed from each cluster. These characteristics include cluster size, position, and pupil size. A multiple discriminant analysis uses these statistics both within and between data frames to formulate optimal rules for assigning the observations into zooming, zoom-out, or no zoom conditions. The statistical procedure effectively generates heuristics for future assignments, based upon these variables. Future work will enhance the accuracy and precision of the modeling technique, and will empirically test users in controlled experiments.« less
Huber, Stefan; Klein, Elise; Moeller, Korbinian; Willmes, Klaus
2015-10-01
In neuropsychological research, single-cases are often compared with a small control sample. Crawford and colleagues developed inferential methods (i.e., the modified t-test) for such a research design. In the present article, we suggest an extension of the methods of Crawford and colleagues employing linear mixed models (LMM). We first show that a t-test for the significance of a dummy coded predictor variable in a linear regression is equivalent to the modified t-test of Crawford and colleagues. As an extension to this idea, we then generalized the modified t-test to repeated measures data by using LMMs to compare the performance difference in two conditions observed in a single participant to that of a small control group. The performance of LMMs regarding Type I error rates and statistical power were tested based on Monte-Carlo simulations. We found that starting with about 15-20 participants in the control sample Type I error rates were close to the nominal Type I error rate using the Satterthwaite approximation for the degrees of freedom. Moreover, statistical power was acceptable. Therefore, we conclude that LMMs can be applied successfully to statistically evaluate performance differences between a single-case and a control sample. Copyright © 2015 Elsevier Ltd. All rights reserved.
Development of the NASA MCAT Auxiliary Telescope for Orbital Debris Research
NASA Technical Reports Server (NTRS)
Frith, James; Lederer, Susan; Cowardin, Heather; Buckalew, Brent; Hickson, Paul; Anz-Meador, Phillip
2016-01-01
The National Aeronautical and Space Administration (NASA) has recently deployed the Meter Class Autonomous Telescope (MCAT) to Ascension Island. MCAT will provide NASA with a dedicated optical sensor for observations of orbital debris with the goal of statistically sampling the orbital and photometric characteristics of the population from low Earth to Geosynchronous orbits. Additionally, a small auxiliary telescope, co-located with MCAT, is being deployed to augment its observations by providing near-simultaneous photometry and astrometry, as well as offloading low priority targets from MCAT's observing queue. It will also be available to provide observational measurements to the Space Surveillance Network for the United States Air Force.
Mixing of thawed coagulation samples prior to testing: Is any technique better than another?
Lima-Oliveira, Gabriel; Adcock, Dorothy M; Salvagno, Gian Luca; Favaloro, Emmanuel J; Lippi, Giuseppe
2016-12-01
Thus study was aimed to investigate whether the mixing technique could influence the results of routine and specialized clotting tests on post-thawed specimens. The sample population consisted of 13 healthy volunteers. Venous blood was collected by evacuated system into three 3.5mL tubes containing 0.109mmol/L buffered sodium citrate. The three blood tubes of each subject were pooled immediately after collection inside a Falcon 15mL tube, then mixed by 6 gentle end-over-end inversions, and centrifuged at 1500g for 15min. Plasma-pool of each subject was then divided in 4 identical aliquots. All aliquots were thawed after 2-day freezing -70°C. Immediately afterwards, the plasma of the four paired aliquots were treated using four different techniques: (a) reference procedure, entailing 6 gentle end-over-end inversions; (b) placing the sample on a blood tube rocker (i.e., rotor mixing) for 5min to induce agitation and mixing; (c) use of a vortex mixer for 20s to induce agitation and mixing; and (d) no mixing. The significance of differences against the reference technique for mixing thawed plasma specimens (i.e., 6 gentle end-over-end inversions) were assessed with paired Student's t-test. The statistical significance was set at p<0.05. As compared to the reference 6-time gentle inversion technique, statistically significant differences were only observed for fibrinogen, and factor VIII in plasma mixed on tube rocker. Some trends were observed in the remaining other cases, but the bias did not achieve statistical significance. We hence suggest that each laboratory should standardize the procedures for mixing of thawed plasma according to a single technique. Copyright © 2016 The Canadian Society of Clinical Chemists. Published by Elsevier Inc. All rights reserved.
Montgomery, Eric; Gao, Chen; de Luca, Julie; Bower, Jessie; Attwood, Kristropher; Ylagan, Lourdes
2014-12-01
The Cellient(®) cell block system has become available as an alternative, partially automated method to create cell blocks in cytology. We sought to show a validation method for immunohistochemical (IHC) staining on the Cellient cell block system (CCB) in comparison with the formalin fixed paraffin embedded traditional cell block (TCB). Immunohistochemical staining was performed using 31 antibodies on 38 patient samples for a total of 326 slides. Split samples were processed using both methods by following the Cellient(®) manufacturer's recommendations for the Cellient cell block (CCB) and the Histogel method for preparing the traditional cell block (TCB). Interpretation was performed by three pathologists and two cytotechnologists. Immunohistochemical stains were scored as: 0/1+ (negative) and 2/3+ (positive). Inter-rater agreement for each antibody was evaluated for CCB and TCB, as well as the intra-rater agreement between TCB and CCB between observers. Interobserver staining concordance for the TCB was obtained with statistical significance (P < 0.05) in 24 of 31 antibodies. Interobserver staining concordance for the CCB was obtained with statistical significance in 27 of 31 antibodies. Intra-observer staining concordance between TCB and CCB was obtained with statistical significance in 24 of 31 antibodies tested. In conclusions, immunohistochemical stains on cytologic specimens processed by the Cellient system are reliable and concordant with stains performed on the same split samples processed via a formalin fixed-paraffin embedded (FFPE) block. The Cellient system is a welcome adjunct to cytology work-flow by producing cell block material of sufficient quality to allow the use of routine IHC. © 2014 Wiley Periodicals, Inc.
BATSE Observations of the Large-Scale Isotropy of Gamma-Ray Bursts
NASA Technical Reports Server (NTRS)
Briggs, Michael S.; Paciesas, William S.; Pendleton, Geoffrey N.; Meegan, Charles A.; Fishman, Gerald J.; Horack, John M.; Brock, Martin N.; Kouveliotou, Chryssa; Hartmann, Dieter H.; Hakkila, Jon
1996-01-01
We use dipole and quadrupole statistics to test the large-scale isotropy of the first 1005 gamma-ray bursts observed by the Burst and Transient Source Experiment (BATSE). In addition to the entire sample of 1005 gamma-ray bursts, many subsets are examined. We use a variety of dipole and quadrupole statistics to search for Galactic and other predicted anisotropies and for anisotropies in a coordinate-system independent manner. We find the gamma-ray burst locations to be consistent with isotropy, e.g., for the total sample the observed Galactic dipole moment (cos theta) differs from the value predicted for isotropy by 0.9 sigma and the observed Galactic quadrupole moment (sin(exp 2) b - 1/3) by 0.3 sigma. We estimate for various models the anisotropies that could have been detected. If one-half of the locations were within 86 deg of the Galactic center, or within 28 deg of the Galactic plane, the ensuing dipole or quadrupole moment would have typically been detected at the 99% confidence level. We compare the observations with the dipole and quadrupole moments of various Galactic models. Several Galactic gamma-ray bursts models have moments within 2 sigma of the observations; most of the Galactic models proposed to date are no longer in acceptable agreement with the data. Although a spherical dark matter halo distribution could be consistent with the data, the required core radius is larger than the core radius of the dark matter halo used to explain the Galaxy's rotation curve. Gamma-ray bursts are much more isotropic than any observed Galactic population, strongly favoring but not requiring an origin at cosmological distances.
Verhagen, Simone J. W.; Simons, Claudia J. P.; van Zelst, Catherine; Delespaul, Philippe A. E. G.
2017-01-01
Background: Mental healthcare needs person-tailored interventions. Experience Sampling Method (ESM) can provide daily life monitoring of personal experiences. This study aims to operationalize and test a measure of momentary reward-related Quality of Life (rQoL). Intuitively, quality of life improves by spending more time on rewarding experiences. ESM clinical interventions can use this information to coach patients to find a realistic, optimal balance of positive experiences (maximize reward) in daily life. rQoL combines the frequency of engaging in a relevant context (a ‘behavior setting’) with concurrent (positive) affect. High rQoL occurs when the most frequent behavior settings are combined with positive affect or infrequent behavior settings co-occur with low positive affect. Methods: Resampling procedures (Monte Carlo experiments) were applied to assess the reliability of rQoL using various behavior setting definitions under different sampling circumstances, for real or virtual subjects with low-, average- and high contextual variability. Furthermore, resampling was used to assess whether rQoL is a distinct concept from positive affect. Virtual ESM beep datasets were extracted from 1,058 valid ESM observations for virtual and real subjects. Results: Behavior settings defined by Who-What contextual information were most informative. Simulations of at least 100 ESM observations are needed for reliable assessment. Virtual ESM beep datasets of a real subject can be defined by Who-What-Where behavior setting combinations. Large sample sizes are necessary for reliable rQoL assessments, except for subjects with low contextual variability. rQoL is distinct from positive affect. Conclusion: rQoL is a feasible concept. Monte Carlo experiments should be used to assess the reliable implementation of an ESM statistic. Future research in ESM should asses the behavior of summary statistics under different sampling situations. This exploration is especially relevant in clinical implementation, where often only small datasets are available. PMID:29163294
Verhagen, Simone J W; Simons, Claudia J P; van Zelst, Catherine; Delespaul, Philippe A E G
2017-01-01
Background: Mental healthcare needs person-tailored interventions. Experience Sampling Method (ESM) can provide daily life monitoring of personal experiences. This study aims to operationalize and test a measure of momentary reward-related Quality of Life (rQoL). Intuitively, quality of life improves by spending more time on rewarding experiences. ESM clinical interventions can use this information to coach patients to find a realistic, optimal balance of positive experiences (maximize reward) in daily life. rQoL combines the frequency of engaging in a relevant context (a 'behavior setting') with concurrent (positive) affect. High rQoL occurs when the most frequent behavior settings are combined with positive affect or infrequent behavior settings co-occur with low positive affect. Methods: Resampling procedures (Monte Carlo experiments) were applied to assess the reliability of rQoL using various behavior setting definitions under different sampling circumstances, for real or virtual subjects with low-, average- and high contextual variability. Furthermore, resampling was used to assess whether rQoL is a distinct concept from positive affect. Virtual ESM beep datasets were extracted from 1,058 valid ESM observations for virtual and real subjects. Results: Behavior settings defined by Who-What contextual information were most informative. Simulations of at least 100 ESM observations are needed for reliable assessment. Virtual ESM beep datasets of a real subject can be defined by Who-What-Where behavior setting combinations. Large sample sizes are necessary for reliable rQoL assessments, except for subjects with low contextual variability. rQoL is distinct from positive affect. Conclusion: rQoL is a feasible concept. Monte Carlo experiments should be used to assess the reliable implementation of an ESM statistic. Future research in ESM should asses the behavior of summary statistics under different sampling situations. This exploration is especially relevant in clinical implementation, where often only small datasets are available.
The bioavailability of manganese in welders in relation to its solubility in welding fumes.
Ellingsen, Dag G; Zibarev, Evgenij; Kusraeva, Zarina; Berlinger, Balazs; Chashchin, Maxim; Bast-Pettersen, Rita; Chashchin, Valery; Thomassen, Yngvar
2013-02-01
Blood and urine samples for determination of manganese (Mn) and iron (Fe) concentrations were collected in a cross-sectional study of 137 currently exposed welders, 137 referents and 34 former welders. Aerosol samples for measurements of personal air exposure to Mn and Fe were also collected. The aerosol samples were assessed for their solubility using a simulated lung lining fluid (Hatch solution). On average 13.8% of the total Mn mass (range 1-49%; N = 237) was soluble (Hatch sol), while only 1.4% (<0.1-10.0%; N = 237) of the total Fe mass was Hatch sol. The welders had statistically significantly higher geometric mean concentrations of Mn in whole blood (B-Mn 12.8 vs. 8.0 μg L (-1)), serum (S-Mn 1.04 vs. 0.77 μg L(-1)) and urine (U-Mn 0.36 vs. 0.07 μg g (-1) cr.) than the referents. Statistically significant univariate correlations were observed between exposure to Hatch sol Mn in the welding aerosol and B-Mn, S-Mn and U-Mn respectively. Pearson's correlation coefficient between mean Hatch sol Mn of two days preceding the collection of biological samples and U-Mn was 0.46 (p < 0.001). The duration of employment as a welder in years was also associated with B-Mn and S-Mn, but not with U-Mn. Statistically significantly higher U-Mn and B-Mn were observed in welders currently exposed to even less than 12 and 6 μg m (-3) Hatchsol Mn, respectively. When using the 95(th) percentile concentration among the referents as a cut-point, 70.0 and 64.5% of the most highly exposed welders exceeded this level with respect to B-Mn and U-Mn. The concentrations of B-Mn, S-Mn and U-Mn were all highly correlated in the welders, but not in the referents.
Secular Extragalactic Parallax and Geometric Distances with Gaia Proper Motions
NASA Astrophysics Data System (ADS)
Paine, Jennie; Darling, Jeremiah K.
2018-06-01
The motion of the Solar System with respect to the cosmic microwave background (CMB) rest frame creates a well measured dipole in the CMB, which corresponds to a linear solar velocity of about 78 AU/yr. This motion causes relatively nearby extragalactic objects to appear to move compared to more distant objects, an effect that can be measured in the proper motions of nearby galaxies. An object at 1 Mpc and perpendicular to the CMB apex will exhibit a secular parallax, observed as a proper motion, of 78 µas/yr. The relatively large peculiar motions of galaxies make the detection of secular parallax challenging for individual objects. Instead, a statistical parallax measurement can be made for a sample of objects with proper motions, where the global parallax signal is modeled as an E-mode dipole that diminishes linearly with distance. We present preliminary results of applying this model to a sample of nearby galaxies with Gaia proper motions to detect the statistical secular parallax signal. The statistical measurement can be used to calibrate the canonical cosmological “distance ladder.”
A search for AGN activity in Infrared-Faint Radio Sources (IFRS)
NASA Astrophysics Data System (ADS)
Lenc, Emil; Middelberg, Enno; Norris, Ray; Mao, Minnie
2009-04-01
We propose to observe a large sample of radio sources from the ATLAS (Australia Telescope Large Area Survey) source catalogue with the LBA, to determine their compactness. The sample consists of 36 sources with no counterpart in the co-located SWIRE survey (3.6 um to 160 um), carried out with the Spitzer Space Telescope. This rare class of sources, dubber Infrared-Faint Radio Sources (IFRS), is inconsistent with current galaxy evolution models. VLBI observations are an essential way to obtain further clues on what these objects are and why they are hidden from infrared observations. We will measure the flux densities on long baselines to determine their compactness. Only five IFRS have been previously targeted with VLBI observations (resulting in two detections). We propose using single baseline (Parkes-ATCA) eVLBI observations with the LBA at 1 Gbps to maximise sensitivity. With the observations proposed here we will increase the number of VLBI-observed IFRS from 5 to 36, allowing us to draw statistical conclusions about this intriguing new class of objects.
A search for AGN activity in Infrared-Faint Radio Sources (IFRS)
NASA Astrophysics Data System (ADS)
Lenc, Emil; Middelberg, Enno; Norris, Ray; Mao, Minnie
2010-04-01
We propose to observe a large sample of radio sources from the ATLAS (Australia Telescope Large Area Survey) source catalogue with the LBA, to determine their compactness. The sample consists of 36 sources with no counterpart in the co-located SWIRE survey (3.6 um to 160 um), carried out with the Spitzer Space Telescope. This rare class of sources, dubber Infrared-Faint Radio Sources (IFRS), is inconsistent with current galaxy evolution models. VLBI observations are an essential way to obtain further clues on what these objects are and why they are hidden from infrared observations. We will measure the flux densities on long baselines to determine their compactness. Only five IFRS have been previously targeted with VLBI observations (resulting in two detections). We propose using single baseline (Parkes-ATCA) eVLBI observations with the LBA at 1 Gbps to maximise sensitivity. With the observations proposed here we will increase the number of VLBI-observed IFRS from 5 to 36, allowing us to draw statistical conclusions about this intriguing new class of objects.
Asteroid orbital inversion using uniform phase-space sampling
NASA Astrophysics Data System (ADS)
Muinonen, K.; Pentikäinen, H.; Granvik, M.; Oszkiewicz, D.; Virtanen, J.
2014-07-01
We review statistical inverse methods for asteroid orbit computation from a small number of astrometric observations and short time intervals of observations. With the help of Markov-chain Monte Carlo methods (MCMC), we present a novel inverse method that utilizes uniform sampling of the phase space for the orbital elements. The statistical orbital ranging method (Virtanen et al. 2001, Muinonen et al. 2001) was set out to resolve the long-lasting challenges in the initial computation of orbits for asteroids. The ranging method starts from the selection of a pair of astrometric observations. Thereafter, the topocentric ranges and angular deviations in R.A. and Decl. are randomly sampled. The two Cartesian positions allow for the computation of orbital elements and, subsequently, the computation of ephemerides for the observation dates. Candidate orbital elements are included in the sample of accepted elements if the χ^2-value between the observed and computed observations is within a pre-defined threshold. The sample orbital elements obtain weights based on a certain debiasing procedure. When the weights are available, the full sample of orbital elements allows the probabilistic assessments for, e.g., object classification and ephemeris computation as well as the computation of collision probabilities. The MCMC ranging method (Oszkiewicz et al. 2009; see also Granvik et al. 2009) replaces the original sampling algorithm described above with a proposal probability density function (p.d.f.), and a chain of sample orbital elements results in the phase space. MCMC ranging is based on a bivariate Gaussian p.d.f. for the topocentric ranges, and allows for the sampling to focus on the phase-space domain with most of the probability mass. In the virtual-observation MCMC method (Muinonen et al. 2012), the proposal p.d.f. for the orbital elements is chosen to mimic the a posteriori p.d.f. for the elements: first, random errors are simulated for each observation, resulting in a set of virtual observations; second, corresponding virtual least-squares orbital elements are derived using the Nelder-Mead downhill simplex method; third, repeating the procedure two times allows for a computation of a difference for two sets of virtual orbital elements; and, fourth, this orbital-element difference constitutes a symmetric proposal in a random-walk Metropolis-Hastings algorithm, avoiding the explicit computation of the proposal p.d.f. In a discrete approximation, the allowed proposals coincide with the differences that are based on a large number of pre-computed sets of virtual least-squares orbital elements. The virtual-observation MCMC method is thus based on the characterization of the relevant volume in the orbital-element phase space. Here we utilize MCMC to map the phase-space domain of acceptable solutions. We can make use of the proposal p.d.f.s from the MCMC ranging and virtual-observation methods. The present phase-space mapping produces, upon convergence, a uniform sampling of the solution space within a pre-defined χ^2-value. The weights of the sampled orbital elements are then computed on the basis of the corresponding χ^2-values. The present method resembles the original ranging method. On one hand, MCMC mapping is insensitive to local extrema in the phase space and efficiently maps the solution space. This is somewhat contrary to the MCMC methods described above. On the other hand, MCMC mapping can suffer from producing a small number of sample elements with small χ^2-values, in resemblance to the original ranging method. We apply the methods to example near-Earth, main-belt, and transneptunian objects, and highlight the utilization of the methods in the data processing and analysis pipeline of the ESA Gaia space mission.
Voids and constraints on nonlinear clustering of galaxies
NASA Technical Reports Server (NTRS)
Vogeley, Michael S.; Geller, Margaret J.; Park, Changbom; Huchra, John P.
1994-01-01
Void statistics of the galaxy distribution in the Center for Astrophysics Redshift Survey provide strong constraints on galaxy clustering in the nonlinear regime, i.e., on scales R equal to or less than 10/h Mpc. Computation of high-order moments of the galaxy distribution requires a sample that (1) densely traces the large-scale structure and (2) covers sufficient volume to obtain good statistics. The CfA redshift survey densely samples structure on scales equal to or less than 10/h Mpc and has sufficient depth and angular coverage to approach a fair sample on these scales. In the nonlinear regime, the void probability function (VPF) for CfA samples exhibits apparent agreement with hierarchical scaling (such scaling implies that the N-point correlation functions for N greater than 2 depend only on pairwise products of the two-point function xi(r)) However, simulations of cosmological models show that this scaling in redshift space does not necessarily imply such scaling in real space, even in the nonlinear regime; peculiar velocities cause distortions which can yield erroneous agreement with hierarchical scaling. The underdensity probability measures the frequency of 'voids' with density rho less than 0.2 -/rho. This statistic reveals a paucity of very bright galaxies (L greater than L asterisk) in the 'voids.' Underdensities are equal to or greater than 2 sigma more frequent in bright galaxy samples than in samples that include fainter galaxies. Comparison of void statistics of CfA samples with simulations of a range of cosmological models favors models with Gaussian primordial fluctuations and Cold Dark Matter (CDM)-like initial power spectra. Biased models tend to produce voids that are too empty. We also compare these data with three specific models of the Cold Dark Matter cosmogony: an unbiased, open universe CDM model (omega = 0.4, h = 0.5) provides a good match to the VPF of the CfA samples. Biasing of the galaxy distribution in the 'standard' CDM model (omega = 1, b = 1.5; see below for definitions) and nonzero cosmological constant CDM model (omega = 0.4, h = 0.6 lambda(sub 0) = 0.6, b = 1.3) produce voids that are too empty. All three simulations match the observed VPF and underdensity probability for samples of very bright (M less than M asterisk = -19.2) galaxies, but produce voids that are too empty when compared with samples that include fainter galaxies.
NASA Astrophysics Data System (ADS)
Cheng, Hok-Chuen
This thesis summaries the measurements of correlations between Lambda 0Lambda0, Lambda0Lambda 0, and Lambda0Lambda 0 hyperon pairs produced inclusively at the LHC, which are useful for a better understanding of the quark-antiquark pair production and jet fragmentation and hadronization processes. The analysis is based on hyperon pairs selected using the muon and minimum bias data samples collected at the ATLAS experiment from proton-proton collisions at a center-of-mass energy of 7 TeV in 2010. Excess Lambda0Lambda 0 are observed near the production threshold and are identified to be originated from the parton system in the string model in the MC sample, decaying either directly or through heavy strange resonances such as Sigma0 and Sigma*(1385). Dynamical correlations have been explored through a correlation function defined as the ratio of two-particle to single-particle densities. Positive correlation is observed for Lambda0Lambda0 and anticorrelation is observed for Lambda0Lambda 0 and Lambda0Lambda 0 for Q in [0,2] GeV. The structure replicates similar correlations in pp, pp, and pppp events in PYTHIA generator as predicted by the Lund string fragmentation model. Parameters of the "popcorn" mechanism implemented in the PYTHIA generator are tuned and are found to have little impact on the structure observed. The spin composition of the sample is extracted using a data-driven reference sample built by event mixing. Appropriate corrections have been made to the kinematic distributions in the reference sample by kinematic weighting to make sure that the detector effects are well modeled. A modified Pearson's chi2 test statistics is calculated for the costheta* distribution to determine the best-fitted A-value for data. The results are consistent with zero for both like-type and unlike-type hyperon pairs in Q ∈ [0,10] GeV and Q ∈ [1,10] GeV respectively. The data statistics in the range of Q ∈ [0, 1] GeV is currently too low for the estimation of the emitter size for Fermi-Dirac correlation.
Dynamic heterogeneity and non-Gaussian statistics for acetylcholine receptors on live cell membrane
NASA Astrophysics Data System (ADS)
He, W.; Song, H.; Su, Y.; Geng, L.; Ackerson, B. J.; Peng, H. B.; Tong, P.
2016-05-01
The Brownian motion of molecules at thermal equilibrium usually has a finite correlation time and will eventually be randomized after a long delay time, so that their displacement follows the Gaussian statistics. This is true even when the molecules have experienced a complex environment with a finite correlation time. Here, we report that the lateral motion of the acetylcholine receptors on live muscle cell membranes does not follow the Gaussian statistics for normal Brownian diffusion. From a careful analysis of a large volume of the protein trajectories obtained over a wide range of sampling rates and long durations, we find that the normalized histogram of the protein displacements shows an exponential tail, which is robust and universal for cells under different conditions. The experiment indicates that the observed non-Gaussian statistics and dynamic heterogeneity are inherently linked to the slow-active remodelling of the underlying cortical actin network.
An astronomer's guide to period searching
NASA Astrophysics Data System (ADS)
Schwarzenberg-Czerny, A.
2003-03-01
We concentrate on analysis of unevenly sampled time series, interrupted by periodic gaps, as often encountered in astronomy. While some of our conclusions may appear surprising, all are based on classical statistical principles of Fisher & successors. Except for discussion of the resolution issues, it is best for the reader to forget temporarily about Fourier transforms and to concentrate on problems of fitting of a time series with a model curve. According to their statistical content we divide the issues into several sections, consisting of: (ii) statistical numerical aspects of model fitting, (iii) evaluation of fitted models as hypotheses testing, (iv) the role of the orthogonal models in signal detection (v) conditions for equivalence of periodograms (vi) rating sensitivity by test power. An experienced observer working with individual objects would benefit little from formalized statistical approach. However, we demonstrate the usefulness of this approach in evaluation of performance of periodograms and in quantitative design of large variability surveys.
Design-based Sample and Probability Law-Assumed Sample: Their Role in Scientific Investigation.
ERIC Educational Resources Information Center
Ojeda, Mario Miguel; Sahai, Hardeo
2002-01-01
Discusses some key statistical concepts in probabilistic and non-probabilistic sampling to provide an overview for understanding the inference process. Suggests a statistical model constituting the basis of statistical inference and provides a brief review of the finite population descriptive inference and a quota sampling inferential theory.…
Comparative Financial Statistics for Public Two-Year Colleges: FY 1991 National Sample.
ERIC Educational Resources Information Center
Dickmeyer, Nathan; Cirino, Anna Marie
This report provides comparative financial information derived from a national sample of 503 public two-year colleges. The report includes space for colleges to compare their institutional statistics with data provided on national sample medians; quartile data for the national sample; and statistics presented in various formats, including tables,…
Implications of the Observed Ultraluminous X-Ray Source Luminosity Function
NASA Technical Reports Server (NTRS)
Swartz, Douglas A.; Tennant, Allyn; Soria, Roberto; Yukita, Mihoko
2012-01-01
We present the X-ray luminosity function (XLF) of ultraluminous X-ray (ULX) sources with 0.3-10.0 keV luminosities in excess of 10(sup 39) erg/s in a complete sample of nearby galaxies. The XLF shows a break or cut-off at high luminosities that deviates from its pure power law distribution at lower luminosities. The cut-off is at roughly the Eddington luminosity for a 90-140 solar mass accretor. We examine the effects on the observed XLF of sample biases, of small-number statistics (at the high luminosity end) and of measurement uncertainties. We consider the physical implications of the shape and normalization of the XLF. The XLF is also compared and contrasted to results of other recent surveys.
Boiano, J M; Wallace, M E; Sieber, W K; Groff, J H; Wang, J; Ashley, K
2000-08-01
A field study was conducted with the goal of comparing the performance of three recently developed or modified sampling and analytical methods for the determination of airborne hexavalent chromium (Cr(VI)). The study was carried out in a hard chrome electroplating facility and in a jet engine manufacturing facility where airborne Cr(VI) was expected to be present. The analytical methods evaluated included two laboratory-based procedures (OSHA Method ID-215 and NIOSH Method 7605) and a field-portable method (NIOSH Method 7703). These three methods employ an identical sampling methodology: collection of Cr(VI)-containing aerosol on a polyvinyl chloride (PVC) filter housed in a sampling cassette, which is connected to a personal sampling pump calibrated at an appropriate flow rate. The basis of the analytical methods for all three methods involves extraction of the PVC filter in alkaline buffer solution, chemical isolation of the Cr(VI) ion, complexation of the Cr(VI) ion with 1,5-diphenylcarbazide, and spectrometric measurement of the violet chromium diphenylcarbazone complex at 540 nm. However, there are notable specific differences within the sample preparation procedures used in three methods. To assess the comparability of the three measurement protocols, a total of 20 side-by-side air samples were collected, equally divided between a chromic acid electroplating operation and a spray paint operation where water soluble forms of Cr(VI) were used. A range of Cr(VI) concentrations from 0.6 to 960 microg m(-3), with Cr(VI) mass loadings ranging from 0.4 to 32 microg, was measured at the two operations. The equivalence of the means of the log-transformed Cr(VI) concentrations obtained from the different analytical methods was compared. Based on analysis of variance (ANOVA) results, no statistically significant differences were observed between mean values measured using each of the three methods. Small but statistically significant differences were observed between results obtained from performance evaluation samples for the NIOSH field method and the OSHA laboratory method.
Significance levels for studies with correlated test statistics.
Shi, Jianxin; Levinson, Douglas F; Whittemore, Alice S
2008-07-01
When testing large numbers of null hypotheses, one needs to assess the evidence against the global null hypothesis that none of the hypotheses is false. Such evidence typically is based on the test statistic of the largest magnitude, whose statistical significance is evaluated by permuting the sample units to simulate its null distribution. Efron (2007) has noted that correlation among the test statistics can induce substantial interstudy variation in the shapes of their histograms, which may cause misleading tail counts. Here, we show that permutation-based estimates of the overall significance level also can be misleading when the test statistics are correlated. We propose that such estimates be conditioned on a simple measure of the spread of the observed histogram, and we provide a method for obtaining conditional significance levels. We justify this conditioning using the conditionality principle described by Cox and Hinkley (1974). Application of the method to gene expression data illustrates the circumstances when conditional significance levels are needed.
Ganna, Andrea; Lee, Donghwan; Ingelsson, Erik; Pawitan, Yudi
2015-07-01
It is common and advised practice in biomedical research to validate experimental or observational findings in a population different from the one where the findings were initially assessed. This practice increases the generalizability of the results and decreases the likelihood of reporting false-positive findings. Validation becomes critical when dealing with high-throughput experiments, where the large number of tests increases the chance to observe false-positive results. In this article, we review common approaches to determine statistical thresholds for validation and describe the factors influencing the proportion of significant findings from a 'training' sample that are replicated in a 'validation' sample. We refer to this proportion as rediscovery rate (RDR). In high-throughput studies, the RDR is a function of false-positive rate and power in both the training and validation samples. We illustrate the application of the RDR using simulated data and real data examples from metabolomics experiments. We further describe an online tool to calculate the RDR using t-statistics. We foresee two main applications. First, if the validation study has not yet been collected, the RDR can be used to decide the optimal combination between the proportion of findings taken to validation and the size of the validation study. Secondly, if a validation study has already been done, the RDR estimated using the training data can be compared with the observed RDR from the validation data; hence, the success of the validation study can be assessed. © The Author 2014. Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.
Joint Adaptive Mean-Variance Regularization and Variance Stabilization of High Dimensional Data
Dazard, Jean-Eudes; Rao, J. Sunil
2012-01-01
The paper addresses a common problem in the analysis of high-dimensional high-throughput “omics” data, which is parameter estimation across multiple variables in a set of data where the number of variables is much larger than the sample size. Among the problems posed by this type of data are that variable-specific estimators of variances are not reliable and variable-wise tests statistics have low power, both due to a lack of degrees of freedom. In addition, it has been observed in this type of data that the variance increases as a function of the mean. We introduce a non-parametric adaptive regularization procedure that is innovative in that : (i) it employs a novel “similarity statistic”-based clustering technique to generate local-pooled or regularized shrinkage estimators of population parameters, (ii) the regularization is done jointly on population moments, benefiting from C. Stein's result on inadmissibility, which implies that usual sample variance estimator is improved by a shrinkage estimator using information contained in the sample mean. From these joint regularized shrinkage estimators, we derived regularized t-like statistics and show in simulation studies that they offer more statistical power in hypothesis testing than their standard sample counterparts, or regular common value-shrinkage estimators, or when the information contained in the sample mean is simply ignored. Finally, we show that these estimators feature interesting properties of variance stabilization and normalization that can be used for preprocessing high-dimensional multivariate data. The method is available as an R package, called ‘MVR’ (‘Mean-Variance Regularization’), downloadable from the CRAN website. PMID:22711950
Information-Theoretic Assessment of Sample Imaging Systems
NASA Technical Reports Server (NTRS)
Huck, Friedrich O.; Alter-Gartenberg, Rachel; Park, Stephen K.; Rahman, Zia-ur
1999-01-01
By rigorously extending modern communication theory to the assessment of sampled imaging systems, we develop the formulations that are required to optimize the performance of these systems within the critical constraints of image gathering, data transmission, and image display. The goal of this optimization is to produce images with the best possible visual quality for the wide range of statistical properties of the radiance field of natural scenes that one normally encounters. Extensive computational results are presented to assess the performance of sampled imaging systems in terms of information rate, theoretical minimum data rate, and fidelity. Comparisons of this assessment with perceptual and measurable performance demonstrate that (1) the information rate that a sampled imaging system conveys from the captured radiance field to the observer is closely correlated with the fidelity, sharpness and clarity with which the observed images can be restored and (2) the associated theoretical minimum data rate is closely correlated with the lowest data rate with which the acquired signal can be encoded for efficient transmission.
Epidemiological Observations on Cryptosporidiosis in Diarrheic Goat Kids in Greece.
Giadinis, Nektarios D; Papadopoulos, Elias; Lafi, Shawkat Q; Papanikolopoulou, Vasiliki; Karanikola, Sofia; Diakou, Anastasia; Vergidis, Vergos; Xiao, Lihua; Ioannidou, Evi; Karatzias, Harilaos
2015-01-01
This study aimed at investigating the occurrence of Cryptosporidium spp. in diarrheic goat kids in Greece and the risk factors associated with cryptosporidiosis. Altogether, 292 diarrheic 4-15-day-old goat kids from 54 dairy goat herds of Northern Greece were examined. Oocysts of Cryptosporidium spp. were detected in 223 of 292 (76.4%) goat kids and the intensity of infection was scored as "high" in 142 samples, "moderate" in 45 samples, and "low" in 36 samples. Larger herds (>200 animals) had higher infection rates than smaller ones, although this difference was not statistically significant. Significantly higher infection rates were observed in herds during late kidding season (1 January to 30 April) compared to the early one (1 September to 31 December). These results suggest that cryptosporidiosis is very common in diarrheic goat kids in Greece, especially in large herds during the late parturition season.
Flux-limited sample of Galactic carbon stars
DOE Office of Scientific and Technical Information (OSTI.GOV)
Claussen, M.J.; Kleinmann, S.G.; Joyce, R.R.
Published observational data (including IRAS observations) for a flux-limited sample of 215 Galactic carbon stars (CSs) selected from the 2-micron sky survey of Neugebauer and Leighton (1969) are compiled in extensive tables and graphs and analyzed statistically. The sample is found to penetrate a volume of radius 1.5 kpc, and the local CS space density and surface density are calculated as log rho0 (per cu kpc) = 2.0 + or - 0.4 and log N (per sq kpc) = 1.6 + or - 0.2, respectively. The total Galactic mass-return rate from these CSs is estimated as 0.013 solar mass/yr, implyingmore » a time scale of 0.1-1 Myr for the CS evolutionary phase and a mass of 1.2-1.6 solar mass for the (probably F-type) main-seqence progenitors of CSs. 81 references.« less
BATSE analysis techniques for probing the GRB spatial and luminosity distributions
NASA Technical Reports Server (NTRS)
Hakkila, Jon; Meegan, Charles A.
1992-01-01
The Burst And Transient Source Experiment (BATSE) has measured homogeneity and isotropy parameters from an increasingly large sample of observed gamma-ray bursts (GRBs), while also maintaining a summary of the way in which the sky has been sampled. Measurement of both of these are necessary for any study of the BATSE data statistically, as they take into account the most serious observational selection effects known in the study of GRBs: beam-smearing and inhomogeneous, anisotropic sky sampling. Knowledge of these effects is important to analysis of GRB angular and intensity distributions. In addition to determining that the bursts are local, it is hoped that analysis of such distributions will allow boundaries to be placed on the true GRB spatial distribution and luminosity function. The technique for studying GRB spatial and luminosity distributions is direct. Results of BATSE analyses are compared to Monte Carlo models parameterized by a variety of spatial and luminosity characteristics.
Lange, J H; Lange, P R; Reinhard, T K; Thomulka, K W
1996-08-01
Data were collected and analysed on airborne concentrations of asbestos generated by abatement of different asbestos-containing materials using various removal practices. Airborne concentrations of asbestos are dramatically variable among the types of asbestos-containing material being abated. Abatement practices evaluated in this study were removal of boiler/pipe insulation in a crawl space, ceiling tile, transite, floor tile/mastic with traditional methods, and mastic removal with a high-efficiency particulate air filter blast track (shot-blast) machine. In general, abatement of boiler and pipe insulation produces the highest airborne fibre levels, while abatement of floor tile and mastic was observed to be the lowest. A comparison of matched personal and area samples was not significantly different, and exhibited a good correlation using regression analysis. After adjusting data for outliers, personal sample fibre concentrations were greater than area sample fibre concentrations. Statistical analysis and sample distribution of airborne asbestos concentrations appear to be best represented in a logarithmic form. Area sample fibre concentrations were shown in this study to have a larger variability than personal measurements. Evaluation of outliers in fibre concentration data and the ability of these values to skew sample populations is presented. The use of personal and area samples in determining exposure, selecting personal protective equipment and its historical relevance as related to future abatement projects is discussed.
An asymptotic analysis of the logrank test.
Strawderman, R L
1997-01-01
Asymptotic expansions for the null distribution of the logrank statistic and its distribution under local proportional hazards alternatives are developed in the case of iid observations. The results, which are derived from the work of Gu (1992) and Taniguchi (1992), are easy to interpret, and provide some theoretical justification for many behavioral characteristics of the logrank test that have been previously observed in simulation studies. We focus primarily upon (i) the inadequacy of the usual normal approximation under treatment group imbalance; and, (ii) the effects of treatment group imbalance on power and sample size calculations. A simple transformation of the logrank statistic is also derived based on results in Konishi (1991) and is found to substantially improve the standard normal approximation to its distribution under the null hypothesis of no survival difference when there is treatment group imbalance.
RCT: Module 2.03, Counting Errors and Statistics, Course 8768
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hillmer, Kurt T.
2017-04-01
Radiological sample analysis involves the observation of a random process that may or may not occur and an estimation of the amount of radioactive material present based on that observation. Across the country, radiological control personnel are using the activity measurements to make decisions that may affect the health and safety of workers at those facilities and their surrounding environments. This course will present an overview of measurement processes, a statistical evaluation of both measurements and equipment performance, and some actions to take to minimize the sources of error in count room operations. This course will prepare the student withmore » the skills necessary for radiological control technician (RCT) qualification by passing quizzes, tests, and the RCT Comprehensive Phase 1, Unit 2 Examination (TEST 27566) and by providing in the field skills.« less
Arm structure in normal spiral galaxies, 1: Multivariate data for 492 galaxies
NASA Technical Reports Server (NTRS)
Magri, Christopher
1994-01-01
Multivariate data have been collected as part of an effort to develop a new classification system for spiral galaxies, one which is not necessarily based on subjective morphological properties. A sample of 492 moderately bright northern Sa and Sc spirals was chosen for future statistical analysis. New observations were made at 20 and 21 cm; the latter data are described in detail here. Infrared Astronomy Satellite (IRAS) fluxes were obtained from archival data. Finally, new estimates of arm pattern radomness and of local environmental harshness were compiled for most sample objects.
Wang, Yunpeng; Thompson, Wesley K.; Schork, Andrew J.; Holland, Dominic; Chen, Chi-Hua; Bettella, Francesco; Desikan, Rahul S.; Li, Wen; Witoelar, Aree; Zuber, Verena; Devor, Anna; Nöthen, Markus M.; Rietschel, Marcella; Chen, Qiang; Werge, Thomas; Cichon, Sven; Weinberger, Daniel R.; Djurovic, Srdjan; O’Donovan, Michael; Visscher, Peter M.; Andreassen, Ole A.; Dale, Anders M.
2016-01-01
Most of the genetic architecture of schizophrenia (SCZ) has not yet been identified. Here, we apply a novel statistical algorithm called Covariate-Modulated Mixture Modeling (CM3), which incorporates auxiliary information (heterozygosity, total linkage disequilibrium, genomic annotations, pleiotropy) for each single nucleotide polymorphism (SNP) to enable more accurate estimation of replication probabilities, conditional on the observed test statistic (“z-score”) of the SNP. We use a multiple logistic regression on z-scores to combine information from auxiliary information to derive a “relative enrichment score” for each SNP. For each stratum of these relative enrichment scores, we obtain nonparametric estimates of posterior expected test statistics and replication probabilities as a function of discovery z-scores, using a resampling-based approach that repeatedly and randomly partitions meta-analysis sub-studies into training and replication samples. We fit a scale mixture of two Gaussians model to each stratum, obtaining parameter estimates that minimize the sum of squared differences of the scale-mixture model with the stratified nonparametric estimates. We apply this approach to the recent genome-wide association study (GWAS) of SCZ (n = 82,315), obtaining a good fit between the model-based and observed effect sizes and replication probabilities. We observed that SNPs with low enrichment scores replicate with a lower probability than SNPs with high enrichment scores even when both they are genome-wide significant (p < 5x10-8). There were 693 and 219 independent loci with model-based replication rates ≥80% and ≥90%, respectively. Compared to analyses not incorporating relative enrichment scores, CM3 increased out-of-sample yield for SNPs that replicate at a given rate. This demonstrates that replication probabilities can be more accurately estimated using prior enrichment information with CM3. PMID:26808560
Tipton, John; Hooten, Mevin B.; Goring, Simon
2017-01-01
Scientific records of temperature and precipitation have been kept for several hundred years, but for many areas, only a shorter record exists. To understand climate change, there is a need for rigorous statistical reconstructions of the paleoclimate using proxy data. Paleoclimate proxy data are often sparse, noisy, indirect measurements of the climate process of interest, making each proxy uniquely challenging to model statistically. We reconstruct spatially explicit temperature surfaces from sparse and noisy measurements recorded at historical United States military forts and other observer stations from 1820 to 1894. One common method for reconstructing the paleoclimate from proxy data is principal component regression (PCR). With PCR, one learns a statistical relationship between the paleoclimate proxy data and a set of climate observations that are used as patterns for potential reconstruction scenarios. We explore PCR in a Bayesian hierarchical framework, extending classical PCR in a variety of ways. First, we model the latent principal components probabilistically, accounting for measurement error in the observational data. Next, we extend our method to better accommodate outliers that occur in the proxy data. Finally, we explore alternatives to the truncation of lower-order principal components using different regularization techniques. One fundamental challenge in paleoclimate reconstruction efforts is the lack of out-of-sample data for predictive validation. Cross-validation is of potential value, but is computationally expensive and potentially sensitive to outliers in sparse data scenarios. To overcome the limitations that a lack of out-of-sample records presents, we test our methods using a simulation study, applying proper scoring rules including a computationally efficient approximation to leave-one-out cross-validation using the log score to validate model performance. The result of our analysis is a spatially explicit reconstruction of spatio-temporal temperature from a very sparse historical record.
NASA Astrophysics Data System (ADS)
Oliva-Altamirano, P.; Fisher, D. B.; Glazebrook, K.; Wisnioski, E.; Bekiaris, G.; Bassett, R.; Obreschkow, D.; Abraham, R.
2018-02-01
We present Keck/OSIRIS adaptive optics observations with 150-400 pc spatial sampling of 7 turbulent, clumpy disc galaxies from the DYNAMO sample ($0.07
A critical analysis of high-redshift, massive, galaxy clusters. Part I
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hoyle, Ben; Jimenez, Raul; Verde, Licia
2012-02-01
We critically investigate current statistical tests applied to high redshift clusters of galaxies in order to test the standard cosmological model and describe their range of validity. We carefully compare a sample of high-redshift, massive, galaxy clusters with realistic Poisson sample simulations of the theoretical mass function, which include the effect of Eddington bias. We compare the observations and simulations using the following statistical tests: the distributions of ensemble and individual existence probabilities (in the > M, > z sense), the redshift distributions, and the 2d Kolmogorov-Smirnov test. Using seemingly rare clusters from Hoyle et al. (2011), and Jee etmore » al. (2011) and assuming the same survey geometry as in Jee et al. (2011, which is less conservative than Hoyle et al. 2011), we find that the ( > M, > z) existence probabilities of all clusters are fully consistent with ΛCDM. However assuming the same survey geometry, we use the 2d K-S test probability to show that the observed clusters are not consistent with being the least probable clusters from simulations at > 95% confidence, and are also not consistent with being a random selection of clusters, which may be caused by the non-trivial selection function and survey geometry. Tension can be removed if we examine only a X-ray selected sub sample, with simulations performed assuming a modified survey geometry.« less
Tupinambá, Rogerio Amaral; Claro, Cristiane Aparecida de Assis; Pereira, Cristiane Aparecida; Nobrega, Celestino José Prudente; Claro, Ana Paula Rosifini Alves
2017-01-01
Plasma-polymerized film deposition was created to modify metallic orthodontic brackets surface properties in order to inhibit bacterial adhesion. Hexamethyldisiloxane (HMDSO) polymer films were deposited on conventional (n = 10) and self-ligating (n = 10) stainless steel orthodontic brackets using the Plasma-Enhanced Chemical Vapor Deposition (PECVD) radio frequency technique. The samples were divided into two groups according to the kind of bracket and two subgroups after surface treatment. Scanning Electron Microscopy (SEM) analysis was performed to assess the presence of bacterial adhesion over samples surfaces (slot and wings region) and film layer integrity. Surface roughness was assessed by Confocal Interferometry (CI) and surface wettability, by goniometry. For bacterial adhesion analysis, samples were exposed for 72 hours to a Streptococcus mutans solution for biofilm formation. The values obtained for surface roughness were analyzed using the Mann-Whitney test while biofilm adhesion were assessed by Kruskal-Wallis and SNK test. Significant statistical differences (p< 0.05) for surface roughness and bacterial adhesion reduction were observed on conventional brackets after surface treatment and between conventional and self-ligating brackets; no significant statistical differences were observed between self-ligating groups (p> 0.05). Plasma-polymerized film deposition was only effective on reducing surface roughness and bacterial adhesion in conventional brackets. It was also noted that conventional brackets showed lower biofilm adhesion than self-ligating brackets despite the absence of film.
Ariza-Miguel, Jaime; Oniciuc, Elena-Alexandra; Sanz, Iván; Fernández-Natal, Isabel; Hernández, Marta; Rodríguez-Lázaro, David
2015-09-16
We compared the diagnostic performance of two chromogenic media, Brilliance MRSA 2 agar (Thermo Fisher Scientific) and ChromID MRSA agar (bioMérieux), for MRSA confirmation of 239 Staphylococcus aureus isolates from clinical, animal and food samples. Statistically significant differences were not observed between MRSA confirmation by mecA/mecC PCR, and by culture in both chromogenic media. However, a statistically significant difference was observed between the results obtained by both chromogenic media (p = 0.003). Segregated analysis of the results depending on the origin of the isolates (clinical, animal, and food) revealed a significant lower performance in the MRSA confirmation of food-derived isolates by using Brilliance MRSA 2 agar in comparison to PCR confirmation (p = 0.003) or ChromID MRSA agar (p<0.001). Both chromogenic media provided a good diagnostic performance for detection of MRSA isolates of human and animal origin. In conclusion, the use of chromogenic agar plates for MRSA confirmation of S. aureus isolates can provide a good diagnostic performance (sensitivity >92% and specificity >89%) regardless of the type of chromogenic media used or the origin of the S. aureus isolates. However, our results revealed a lower diagnostic performance for MRSA confirmation of S. aureus isolates from food samples by using Brilliance MRSA 2 agar. Copyright © 2015 Elsevier B.V. All rights reserved.
Reuter, Jon D; Nelson, Suzanne L
2018-05-24
A population of white-tailed deer ( Odocoileus virginianus) resides throughout the island of St John, US Virgin Islands, predominately in the Virgin Islands National Park. Adult deer ( n=23), ranging from 1 yr to 8 yr old, were assessed to characterize body condition and health. Serologic samples were screened for important viral pathogens in the area, including Zika, chikungunya, bluetongue, and epizootic hemorrhagic disease viruses. Samples were collected in July 2016; males were in velvet and all females were in diestrus. Deer had recovered from a severe drought the previous year but were generally healthy, with a low-level but high incidence of tick parasitism. Marked statistically significant changes in hematocrit and hemoglobin levels were associated with the effects of the anesthetic mixture used for capture. No other statistically significant differences were observed. Serum from four deer induced reduction in Zika virus plaques, suggesting possible exposure. No serum was reactive for chikungunya virus. Bluetongue and epizootic hemorrhagic disease antibodies were present in 50% of the sampled deer, but no clinical signs associated with disease were observed during the study period. These data will be valuable for future dynamic health assessment and may help assess changes to the population, such as those induced by climate change, infectious disease, or other demographic events.
NASA Astrophysics Data System (ADS)
Faucher-Giguere, Claude-Andre
2016-10-01
HST has invested thousands of orbits to complete multi-wavelength surveys of high-redshift galaxies including the Deep Fields, COSMOS, 3D-HST and CANDELS. Over the next few years, JWST will undertake complementary, spatially-resolved infrared observations. Cosmological simulations are the most powerful tool to make detailed predictions for the properties of galaxy populations and to interpret these surveys. We will leverage recent major advances in the predictive power of cosmological hydrodynamic simulations to produce the first statistical sample of hundreds of galaxies simulated with 10 pc resolution and with explicit interstellar medium and stellar feedback physics proved to simultaneously reproduce the galaxy stellar mass function, the chemical enrichment of galaxies, and the neutral hydrogen content of galaxy halos. We will process our new set of full-volume cosmological simulations, called FIREBOX, with a mock imaging and spectral synthesis pipeline to produce realistic mock HST and JWST observations, including spatially-resolved photometry and spectroscopy. By comparing FIREBOX with recent high-redshift HST surveys, we will study the stellar build up of galaxies, the evolution massive star-forming clumps, their contribution to bulge growth, the connection of bulges to star formation quenching, and the triggering mechanisms of AGN activity. Our mock data products will also enable us to plan future JWST observing programs. We will publicly release all our mock data products to enable HST and JWST science beyond our own analysis, including with the Frontier Fields.
Real, Jordi; Forné, Carles; Roso-Llorach, Albert; Martínez-Sánchez, Jose M
2016-05-01
Controlling for confounders is a crucial step in analytical observational studies, and multivariable models are widely used as statistical adjustment techniques. However, the validation of the assumptions of the multivariable regression models (MRMs) should be made clear in scientific reporting. The objective of this study is to review the quality of statistical reporting of the most commonly used MRMs (logistic, linear, and Cox regression) that were applied in analytical observational studies published between 2003 and 2014 by journals indexed in MEDLINE.Review of a representative sample of articles indexed in MEDLINE (n = 428) with observational design and use of MRMs (logistic, linear, and Cox regression). We assessed the quality of reporting about: model assumptions and goodness-of-fit, interactions, sensitivity analysis, crude and adjusted effect estimate, and specification of more than 1 adjusted model.The tests of underlying assumptions or goodness-of-fit of the MRMs used were described in 26.2% (95% CI: 22.0-30.3) of the articles and 18.5% (95% CI: 14.8-22.1) reported the interaction analysis. Reporting of all items assessed was higher in articles published in journals with a higher impact factor.A low percentage of articles indexed in MEDLINE that used multivariable techniques provided information demonstrating rigorous application of the model selected as an adjustment method. Given the importance of these methods to the final results and conclusions of observational studies, greater rigor is required in reporting the use of MRMs in the scientific literature.
Hu, Zonghui; Qin, Jing
2018-05-20
Many observational studies adopt what we call retrospective convenience sampling (RCS). With the sample size in each arm prespecified, RCS randomly selects subjects from the treatment-inclined subpopulation into the treatment arm and those from the control-inclined into the control arm. Samples in each arm are representative of the respective subpopulation, but the proportion of the 2 subpopulations is usually not preserved in the sample data. We show in this work that, under RCS, existing causal effect estimators actually estimate the treatment effect over the sample population instead of the underlying study population. We investigate how to correct existing methods for consistent estimation of the treatment effect over the underlying population. Although RCS is adopted in medical studies for ethical and cost-effective purposes, it also has a big advantage for statistical inference: When the tendency to receive treatment is low in a study population, treatment effect estimators under RCS, with proper correction, are more efficient than their parallels under random sampling. These properties are investigated both theoretically and through numerical demonstration. Published 2018. This article is a U.S. Government work and is in the public domain in the USA.
Detection of semi-volatile organic compounds in permeable ...
Abstract The Edison Environmental Center (EEC) has a research and demonstration permeable parking lot comprised of three different permeable systems: permeable asphalt, porous concrete and interlocking concrete permeable pavers. Water quality and quantity analysis has been ongoing since January, 2010. This paper describes a subset of the water quality analysis, analysis of semivolatile organic compounds (SVOCs) to determine if hydrocarbons were in water infiltrated through the permeable surfaces. SVOCs were analyzed in samples collected from 11 dates over a 3 year period, from 2/8/2010 to 4/1/2013.Results are broadly divided into three categories: 42 chemicals were never detected; 12 chemicals (11 chemical test) were detected at a rate of less than 10% or less; and 22 chemicals were detected at a frequency of 10% or greater (ranging from 10% to 66.5% detections). Fundamental and exploratory statistical analyses were performed on these latter analyses results by grouping results by surface type. The statistical analyses were limited due to low frequency of detections and dilutions of samples which impacted detection limits. The infiltrate data through three permeable surfaces were analyzed as non-parametric data by the Kaplan-Meier estimation method for fundamental statistics; there were some statistically observable difference in concentration between pavement types when using Tarone-Ware Comparison Hypothesis Test. Additionally Spearman Rank order non-parame
Schares, G; Langenmayer, M C; Majzoub-Altweck, M; Scharr, J C; Gentile, A; Maksimov, A; Schares, S; Conraths, F J; Gollnick, N S
2016-01-30
Bovine besnoitiosis is caused by Besnoitia besnoiti, an apicomplexan parasite closely related to Toxoplasma gondii and Neospora caninum. In the acute stage of besnoitiosis, cattle suffer from pyrexia, swollen lymph nodes, anorexia and subcutaneous edema. In the chronic stage, tissue cysts are formed in a variety of tissues including the skin. Knowledge about the distribution of tissue cysts of different parts of the skin of infected animals is scarce. Four chronically infected cattle were euthanized and skin samples were taken from a total of 77 standardized cutaneous locations per animal. Portions of the dermis were taken, from which DNA was extracted and examined by real-time PCR. Cycle of transition (Ct) values reflecting the amount of parasite DNA in the samples were determined. For statistical analysis, samples were attributed to 11 larger skin regions ('OuterHindlegDistal', 'Rump, ForelegMiddle', 'NoseFrontEars', 'CheekEye', 'SideLowerPart', 'ForelegDistal', 'SideUpperPart', 'LegsInner', 'VentralHeadNeck', 'DorsalNeckWithersBackTail'). While all samples revealed a positive result in three female cattle, only 63.6% (49/77) of the samples of a bull showed positive results. For statistical analysis, a Ct value of 45 was assumed for samples with a negative result. The dams showed median Ct values of 16.1, 17.5 and 19.4, while in skin samples of the bull a median Ct value of 37.6 was observed. To determine the differences in DNA concentrations between different locations of the skin of the animals, a relative Ct (relCt) was determined by subtracting for each animal indv the MedianCtindv from each sample Ct. Analyses of the relCt values showed that the highest relative parasite DNA concentrations were observed in the categories 'OuterHindlegDistal', 'Rump', 'ForelegMiddle' and 'NoseFrontEars'. The relCt values in these categories differed statistically significantly from those determined for the categories 'VentralHeadNeck' and 'DorsalNeckWithersBackTail'. The analysis showed clear differences in the distribution and the detectability of parasite DNA in the skin of cattle infected with B. besnoiti. In all four animals, samples from the 'Rump' region (Regio fermoris) showed high parasite DNA concentrations. Because this region is also easily accessible for veterinarians, this skin location appears to be optimal for taking skin biopsies for detection or isolation of B. besnoiti. Copyright © 2015 Elsevier B.V. All rights reserved.
Sampling studies to estimate the HIV prevalence rate in female commercial sex workers.
Pascom, Ana Roberta Pati; Szwarcwald, Célia Landmann; Barbosa Júnior, Aristides
2010-01-01
We investigated sampling methods being used to estimate the HIV prevalence rate among female commercial sex workers. The studies were classified according to the adequacy or not of the sample size to estimate HIV prevalence rate and according to the sampling method (probabilistic or convenience). We identified 75 studies that estimated the HIV prevalence rate among female sex workers. Most of the studies employed convenience samples. The sample size was not adequate to estimate HIV prevalence rate in 35 studies. The use of convenience sample limits statistical inference for the whole group. It was observed that there was an increase in the number of published studies since 2005, as well as in the number of studies that used probabilistic samples. This represents a large advance in the monitoring of risk behavior practices and HIV prevalence rate in this group.
Code of Federal Regulations, 2011 CFR
2011-01-01
... 7 Agriculture 2 2011-01-01 2011-01-01 false Statistical sampling procedures for lot inspection of processed fruits and vegetables by attributes. 52.38c Section 52.38c Agriculture Regulations of the... Regulations Governing Inspection and Certification Sampling § 52.38c Statistical sampling procedures for lot...
Code of Federal Regulations, 2011 CFR
2011-01-01
... 7 Agriculture 2 2011-01-01 2011-01-01 false Statistical sampling procedures for on-line inspection by attributes of processed fruits and vegetables. 52.38b Section 52.38b Agriculture Regulations of... Regulations Governing Inspection and Certification Sampling § 52.38b Statistical sampling procedures for on...
Code of Federal Regulations, 2010 CFR
2010-01-01
... 7 Agriculture 2 2010-01-01 2010-01-01 false Statistical sampling procedures for on-line inspection by attributes of processed fruits and vegetables. 52.38b Section 52.38b Agriculture Regulations of... Regulations Governing Inspection and Certification Sampling § 52.38b Statistical sampling procedures for on...
Code of Federal Regulations, 2010 CFR
2010-01-01
... 7 Agriculture 2 2010-01-01 2010-01-01 false Statistical sampling procedures for lot inspection of processed fruits and vegetables by attributes. 52.38c Section 52.38c Agriculture Regulations of the... Regulations Governing Inspection and Certification Sampling § 52.38c Statistical sampling procedures for lot...
75 FR 38871 - Proposed Collection; Comment Request for Revenue Procedure 2004-29
Federal Register 2010, 2011, 2012, 2013, 2014
2010-07-06
... comments concerning Revenue Procedure 2004-29, Statistical Sampling in Sec. 274 Context. DATES: Written... Internet, at [email protected] . SUPPLEMENTARY INFORMATION: Title: Statistical Sampling in Sec...: Revenue Procedure 2004-29 prescribes the statistical sampling methodology by which taxpayers under...
Statistical inference for extended or shortened phase II studies based on Simon's two-stage designs.
Zhao, Junjun; Yu, Menggang; Feng, Xi-Ping
2015-06-07
Simon's two-stage designs are popular choices for conducting phase II clinical trials, especially in the oncology trials to reduce the number of patients placed on ineffective experimental therapies. Recently Koyama and Chen (2008) discussed how to conduct proper inference for such studies because they found that inference procedures used with Simon's designs almost always ignore the actual sampling plan used. In particular, they proposed an inference method for studies when the actual second stage sample sizes differ from planned ones. We consider an alternative inference method based on likelihood ratio. In particular, we order permissible sample paths under Simon's two-stage designs using their corresponding conditional likelihood. In this way, we can calculate p-values using the common definition: the probability of obtaining a test statistic value at least as extreme as that observed under the null hypothesis. In addition to providing inference for a couple of scenarios where Koyama and Chen's method can be difficult to apply, the resulting estimate based on our method appears to have certain advantage in terms of inference properties in many numerical simulations. It generally led to smaller biases and narrower confidence intervals while maintaining similar coverages. We also illustrated the two methods in a real data setting. Inference procedures used with Simon's designs almost always ignore the actual sampling plan. Reported P-values, point estimates and confidence intervals for the response rate are not usually adjusted for the design's adaptiveness. Proper statistical inference procedures should be used.
GOIATO, Marcelo Coelho; dos SANTOS, Daniela Micheline; MORENO, Amália; GENNARI-FILHO, Humberto; PELLIZZER, Eduardo Piza
2011-01-01
The use of ocular prostheses for ophthalmic patients aims to rebuild facial aesthetics and provide an artificial substitute to the visual organ. Natural intemperate conditions promote discoloration of artificial irides and many studies have attempted to produce irides with greater chromatic paint durability using different paint materials. Objectives The present study evaluated the color stability of artificial irides obtained with two techniques (oil painting and digital image) and submitted to microwave polymerization. Material and Methods Forty samples were fabricated simulating ocular prostheses. Each sample was constituted by one disc of acrylic resin N1 and one disc of colorless acrylic resin with the iris interposed between the discs. The irides in brown and blue color were obtained by oil painting or digital image. The color stability was determined by a reflection spectrophotometer and measurements were taken before and after microwave polymerization. Statistical analysis of the techniques for reproducing artificial irides was performed by applying the normal data distribution test followed by 2-way ANOVA and Tukey HSD test (α=.05). Results Chromatic alterations occurred in all specimens and statistically significant differences were observed between the oil-painted samples and those obtained by digital imaging. There was no statistical difference between the brown and blue colors. Independently of technique, all samples suffered color alterations after microwave polymerization. Conclusion The digital imaging technique for reproducing irides presented better color stability after microwave polymerization. PMID:21625733
Unbiased estimation of oceanic mean rainfall from satellite borne radiometer measurements
NASA Technical Reports Server (NTRS)
Mittal, M. C.
1981-01-01
The statistical properties of the radar derived rainfall obtained during the GARP Atlantic Tropical Experiment (GATE) are used to derive quantitative estimates of the spatial and temporal sampling errors associated with estimating rainfall from brightness temperature measurements such as would be obtained from a satelliteborne microwave radiometer employing a practical size antenna aperture. A basis for a method of correcting the so called beam filling problem, i.e., for the effect of nonuniformity of rainfall over the radiometer beamwidth is provided. The method presented employs the statistical properties of the observations themselves without need for physical assumptions beyond those associated with the radiative transfer model. The simulation results presented offer a validation of the estimated accuracy that can be achieved and the graphs included permit evaluation of the effect of the antenna resolution on both the temporal and spatial sampling errors.
Time Series Analysis Based on Running Mann Whitney Z Statistics
USDA-ARS?s Scientific Manuscript database
A sensitive and objective time series analysis method based on the calculation of Mann Whitney U statistics is described. This method samples data rankings over moving time windows, converts those samples to Mann-Whitney U statistics, and then normalizes the U statistics to Z statistics using Monte-...
Cloud encounter statistics in the 28.5-43.5 KFT altitude region from four years of GASP observations
NASA Technical Reports Server (NTRS)
Jasperson, W. H.; Nastrom, G. D.; Davis, R. E.; Holdeman, J. D.
1983-01-01
The results of an analysis of cloud encounter measurements taken at aircraft flight altitudes as part of the Global Atmospheric Sampling Program are summarized. The results can be used in estimating the probability of cloud encounter and in assessing the economic feasibility of laminar flow control aircraft along particular routes. The data presented clearly show the tropical circulation and its seasonal migration; characteristics of the mid-latitude regime, such as the large-scale traveling cyclones in the winter and increased convective activity in the summer, can be isolated in the data. The cloud encounter statistics are shown to be consistent with the mid-latitude cyclone model. A model for TIC (time-in-clouds), a cloud encounter statistic, is presented for several common airline routes.
Evaluation and application of summary statistic imputation to discover new height-associated loci.
Rüeger, Sina; McDaid, Aaron; Kutalik, Zoltán
2018-05-01
As most of the heritability of complex traits is attributed to common and low frequency genetic variants, imputing them by combining genotyping chips and large sequenced reference panels is the most cost-effective approach to discover the genetic basis of these traits. Association summary statistics from genome-wide meta-analyses are available for hundreds of traits. Updating these to ever-increasing reference panels is very cumbersome as it requires reimputation of the genetic data, rerunning the association scan, and meta-analysing the results. A much more efficient method is to directly impute the summary statistics, termed as summary statistics imputation, which we improved to accommodate variable sample size across SNVs. Its performance relative to genotype imputation and practical utility has not yet been fully investigated. To this end, we compared the two approaches on real (genotyped and imputed) data from 120K samples from the UK Biobank and show that, genotype imputation boasts a 3- to 5-fold lower root-mean-square error, and better distinguishes true associations from null ones: We observed the largest differences in power for variants with low minor allele frequency and low imputation quality. For fixed false positive rates of 0.001, 0.01, 0.05, using summary statistics imputation yielded a decrease in statistical power by 9, 43 and 35%, respectively. To test its capacity to discover novel associations, we applied summary statistics imputation to the GIANT height meta-analysis summary statistics covering HapMap variants, and identified 34 novel loci, 19 of which replicated using data in the UK Biobank. Additionally, we successfully replicated 55 out of the 111 variants published in an exome chip study. Our study demonstrates that summary statistics imputation is a very efficient and cost-effective way to identify and fine-map trait-associated loci. Moreover, the ability to impute summary statistics is important for follow-up analyses, such as Mendelian randomisation or LD-score regression.
Evaluation and application of summary statistic imputation to discover new height-associated loci
2018-01-01
As most of the heritability of complex traits is attributed to common and low frequency genetic variants, imputing them by combining genotyping chips and large sequenced reference panels is the most cost-effective approach to discover the genetic basis of these traits. Association summary statistics from genome-wide meta-analyses are available for hundreds of traits. Updating these to ever-increasing reference panels is very cumbersome as it requires reimputation of the genetic data, rerunning the association scan, and meta-analysing the results. A much more efficient method is to directly impute the summary statistics, termed as summary statistics imputation, which we improved to accommodate variable sample size across SNVs. Its performance relative to genotype imputation and practical utility has not yet been fully investigated. To this end, we compared the two approaches on real (genotyped and imputed) data from 120K samples from the UK Biobank and show that, genotype imputation boasts a 3- to 5-fold lower root-mean-square error, and better distinguishes true associations from null ones: We observed the largest differences in power for variants with low minor allele frequency and low imputation quality. For fixed false positive rates of 0.001, 0.01, 0.05, using summary statistics imputation yielded a decrease in statistical power by 9, 43 and 35%, respectively. To test its capacity to discover novel associations, we applied summary statistics imputation to the GIANT height meta-analysis summary statistics covering HapMap variants, and identified 34 novel loci, 19 of which replicated using data in the UK Biobank. Additionally, we successfully replicated 55 out of the 111 variants published in an exome chip study. Our study demonstrates that summary statistics imputation is a very efficient and cost-effective way to identify and fine-map trait-associated loci. Moreover, the ability to impute summary statistics is important for follow-up analyses, such as Mendelian randomisation or LD-score regression. PMID:29782485
Hansen, John P
2003-01-01
Healthcare quality improvement professionals need to understand and use inferential statistics to interpret sample data from their organizations. In quality improvement and healthcare research studies all the data from a population often are not available, so investigators take samples and make inferences about the population by using inferential statistics. This three-part series will give readers an understanding of the concepts of inferential statistics as well as the specific tools for calculating confidence intervals for samples of data. This article, Part 2, describes probability, populations, and samples. The uses of descriptive and inferential statistics are outlined. The article also discusses the properties and probability of normal distributions, including the standard normal distribution.
ERIC Educational Resources Information Center
Landsberger, Betty H.
To locate possible causes for the gender and race differences observed in adolescent health status, an analysis was made of the relationship between the scores of a national sample of 12- to 17-year-old adolescents on selected items of the National Center for Health Statistics' Health Examination Survey. Thirty survey items indicating social…
VizieR Online Data Catalog: REFLEX Galaxy Cluster Survey catalogue (Boehringer+, 2004)
NASA Astrophysics Data System (ADS)
Boehringer, H.; Schuecker, P.; Guzzo, L.; Collins, C. A.; Voges, W.; Cruddace, R. G.; Ortiz-Gil, A.; Chincarini, G.; de Grandi, S.; Edge, A. C.; MacGillivray, H. T.; Neumann, D. M.; Schindler, S.; Shaver, P.
2004-05-01
The following tables provide the catalogue as well as several data files necessary to reproduce the sample preparation. These files are also required for the cosmological modeling of these observations in e.g. the study of the statistics of the large-scale structure of the matter distribution in the Universe and related cosmological tests. (13 data files).
ERIC Educational Resources Information Center
Kelly, Philip L.
Observations and recollections of Mexican workers, smuggled illegally to farms in southern Colorado, resulted in this account of their attitudes toward work conditions, pay and benefits, leisure activities, feelings of insecurity, and their aspirations. Backgrounds of the 30 aliens interviewed coincided with available statistics on illegal Mexican…
Microwave dielectric spectrum of rocks
NASA Technical Reports Server (NTRS)
Ulaby, F. T.; Bengal, T.; East, J.; Dobson, M. C.; Garvin, J.; Evans, D.
1988-01-01
A combination of several measurement techniques was used to investigate the dielectric properties of 80 rock samples in the microwave region. The real part of the dielectric constant, epsilon', was measured in 0.1 GHz steps from 0.5 to 18 GHz, and the imaginary part, epsilon'', was measured at five frequencies extending between 1.6 and 16 GHz. In addition to the dielectric measurements, the bulk density was measured for all the samples and the bulk chemical composition was determined for 56 of the samples. The study shows that epsilon' is frequency-dependent over the 0.5 to 18 GHz range for all rock samples, and that the bulk density rho accounts for about 50 percent of the observed variance of epsilon'. For individual rock types (by genesis), about 90 percent of the observed variance may be explained by the combination of density and the fractional contents of SiO2, Fe2O3, MgO, and TiO2. For the loss factor epsilon'', it was not possible to establish statistically significant relationships between it and the measured properties of the rock samples (density and chemical composition).
A new model of physical evolution of Jupiter-family comets
NASA Astrophysics Data System (ADS)
Rickman, H.; Szutowicz, S.; Wójcikowski, K.
2014-07-01
We aim to find the statistical physical lifetimes of Jupiter Family comets. For this purpose, we try to model the processes that govern the dynamical and physical evolution of comets. We pay special attention to physical evolution; attempts at such modelling have been made before, but we propose a more accurate model, which will include more physical effects. The model is tested on a sample of fictitious comets based on real Jupiter Family comets with some orbital elements changed to a state before the capture by Jupiter. We model four different physical effects: erosion by sublimation, dust mantling, rejuvenation (mantle blow-off), and splitting. While for sublimation and splitting there already are some models, like di Sisto et. al. (2009), and we only wish to make them more accurate, dust mantling and rejuvenation have not been included in previous, statistical physical evolution models. Each of these effects depends on one or more tunable parameters, which we establish by choosing the model that best fits the observed comet sample in a way similar to di Sisto et. al. (2009). In contrast to di Sisto et. al., our comparison also involves the observed active fractions vs. nuclear radii.
Possible influence of the environmental pollutant bisphenol A on the cardiometabolic risk factors.
Milošević, Nataša; Jakšić, Vladimir; Sudji, Jan; Vuković, Bojan; Ičin, Tijana; Milić, Nataša; Medić Stojanoska, Milica
2017-02-01
Bisphenol A (BPA) is a ubiquitous environmental pollutant which is often associated with various health issues. In this study 103 healthy female volunteers in reproductive age from Serbian north province Vojvodina were enrolled and examined for the BPA exposure in the urine samples after 12 h of fasting. BPA was found in 35.92 % (37/103) of subjects. Statistically significant increment in waist circumference (p = 0.045) and waist-to-height ratio (p = 0.037) was observed among the BPA positive women in comparison with the women who had the same energetic balance and had not been exposed to BPA. Linear correlation was obtained between the BPA concentration in urine samples and body mass index (r 2 = 0.35, p = 0.003) waist circumference (r 2 = 0.21, p = 0.02) and waist-to-height ratio (r 2 = 0.25, p = 0.01) among the obese. High energetic intake and reduced physical activity additionally pronounced BPA positive association with obesity. No statistically significant difference was observed in triglycerides, HDL and LDL cholesterol levels between the BPA exposed and BPA non-exposed female volunteers.
An Extension of the EDGES Survey: Stellar Populations in Dark Matter Halos
NASA Astrophysics Data System (ADS)
van Zee, Liese
The formation and evolution of galactic disks is one of the key questions in extragalactic astronomy today. We plan to use archival data from GALEX, Spitzer, and WISE to investigate the growth and evolution of the stellar component in a statistical sample of nearby galaxies. Data covering a broad wavelength range are critical for measurement of current star formation activity, stellar populations, and stellar distributions in nearby galaxies. In order to investigate the timescales associated with the growth of galactic disks, we will (1) investigate the structure of the underlying stellar distribution, (2) measure the ratio of current-to-past star formation activity as a function of radius, and (3) investigate the growth of the stellar disk as a function of baryon fraction and total dynamical mass. The proposed projects leverage the existing deep wide field-of-view near infrared imaging observations obtained with the Spitzer Space Telescope as part of the EDGES Survey, a Cycle 8 Exploration Science Program. The proposed analysis of multiwavelength imaging observations of a well-defined statistical sample will place strong constraints on hierarchical models of galaxy formation and evolution and will further our understanding of the stellar component of nearby galaxies.
Fisher statistics for analysis of diffusion tensor directional information.
Hutchinson, Elizabeth B; Rutecki, Paul A; Alexander, Andrew L; Sutula, Thomas P
2012-04-30
A statistical approach is presented for the quantitative analysis of diffusion tensor imaging (DTI) directional information using Fisher statistics, which were originally developed for the analysis of vectors in the field of paleomagnetism. In this framework, descriptive and inferential statistics have been formulated based on the Fisher probability density function, a spherical analogue of the normal distribution. The Fisher approach was evaluated for investigation of rat brain DTI maps to characterize tissue orientation in the corpus callosum, fornix, and hilus of the dorsal hippocampal dentate gyrus, and to compare directional properties in these regions following status epilepticus (SE) or traumatic brain injury (TBI) with values in healthy brains. Direction vectors were determined for each region of interest (ROI) for each brain sample and Fisher statistics were applied to calculate the mean direction vector and variance parameters in the corpus callosum, fornix, and dentate gyrus of normal rats and rats that experienced TBI or SE. Hypothesis testing was performed by calculation of Watson's F-statistic and associated p-value giving the likelihood that grouped observations were from the same directional distribution. In the fornix and midline corpus callosum, no directional differences were detected between groups, however in the hilus, significant (p<0.0005) differences were found that robustly confirmed observations that were suggested by visual inspection of directionally encoded color DTI maps. The Fisher approach is a potentially useful analysis tool that may extend the current capabilities of DTI investigation by providing a means of statistical comparison of tissue structural orientation. Copyright © 2012 Elsevier B.V. All rights reserved.
75 FR 53738 - Proposed Collection; Comment Request for Rev. Proc. 2007-35
Federal Register 2010, 2011, 2012, 2013, 2014
2010-09-01
... Revenue Procedure Revenue Procedure 2007-35, Statistical Sampling for purposes of Section 199. DATES... through the Internet, at [email protected] . SUPPLEMENTARY INFORMATION: Title: Statistical Sampling...: This revenue procedure provides for determining when statistical sampling may be used in purposes of...
Framework for making better predictions by directly estimating variables' predictivity.
Lo, Adeline; Chernoff, Herman; Zheng, Tian; Lo, Shaw-Hwa
2016-12-13
We propose approaching prediction from a framework grounded in the theoretical correct prediction rate of a variable set as a parameter of interest. This framework allows us to define a measure of predictivity that enables assessing variable sets for, preferably high, predictivity. We first define the prediction rate for a variable set and consider, and ultimately reject, the naive estimator, a statistic based on the observed sample data, due to its inflated bias for moderate sample size and its sensitivity to noisy useless variables. We demonstrate that the [Formula: see text]-score of the PR method of VS yields a relatively unbiased estimate of a parameter that is not sensitive to noisy variables and is a lower bound to the parameter of interest. Thus, the PR method using the [Formula: see text]-score provides an effective approach to selecting highly predictive variables. We offer simulations and an application of the [Formula: see text]-score on real data to demonstrate the statistic's predictive performance on sample data. We conjecture that using the partition retention and [Formula: see text]-score can aid in finding variable sets with promising prediction rates; however, further research in the avenue of sample-based measures of predictivity is much desired.
Airborne desert dust and aeromicrobiology over the Turkish Mediterranean coastline
Griffin, Dale W.; Kubilay, Nilgün; Kocak, Mustafa; Gray, Mike A.; Borden, Timothy C.; Shinn, Eugene A.
2007-01-01
Between 18 March and 27 October 2002, 220 air samples were collected on 209 of 224 calendar days, on top of a coastal atmospheric research tower in Erdemli, Turkey. The volume of air filtered for each sample was 340 liters. Two hundred fifty-seven bacterial and 2598 fungal colony forming units (CFU) were enumerated from the samples using a low-nutrient agar. Ground-based dust measurements demonstrated that the region is routinely impacted by dust generated regionally and from North Africa and that the highest combined percent recovery of total CFU and African dust deposition occurred in the month of April (93.4% of CFU recovery and 91.1% of dust deposition occurred during African dust days versus no African dust present, for that month). A statistically significant correlation was observed (peak regional African dust months of March, April and May; rs=0.576, P=0.000) between an increase in the prevalence of microorganisms recovered from atmospheric samples on dust days (regional and African as determined by ground-based dust measurements), versus that observed on non-dust days. Given the prevalence of atmospherically suspended desert dust and microorganisms observed in this study, and that culture-based studies typically only recover a small fraction (
2011-01-01
Background There is substantial variation in reported reference intervals for canine plasma creatinine among veterinary laboratories, thereby influencing the clinical assessment of analytical results. The aims of the study was to determine the inter- and intra-laboratory variation in plasma creatinine among 10 veterinary laboratories, and to compare results from each laboratory with the upper limit of its reference interval. Methods Samples were collected from 10 healthy dogs, 10 dogs with expected intermediate plasma creatinine concentrations, and 10 dogs with azotemia. Overlap was observed for the first two groups. The 30 samples were divided into 3 batches and shipped in random order by postal delivery for plasma creatinine determination. Statistical testing was performed in accordance with ISO standard methodology. Results Inter- and intra-laboratory variation was clinically acceptable as plasma creatinine values for most samples were usually of the same magnitude. A few extreme outliers caused three laboratories to fail statistical testing for consistency. Laboratory sample means above or below the overall sample mean, did not unequivocally reflect high or low reference intervals in that laboratory. Conclusions In spite of close analytical results, further standardization among laboratories is warranted. The discrepant reference intervals seem to largely reflect different populations used in establishing the reference intervals, rather than analytical variation due to different laboratory methods. PMID:21477356
NASA Technical Reports Server (NTRS)
Levy, Gad; Tiu, Felice S.
1990-01-01
Statistical tests are performed on the Seasat scatterometer observations to examine if and to what degree thermal advection and stratification effects manifest themselves in these remotely sensed measurements of mean wind and wind stress over the ocean. On the basis of a two layer baroclinic boundary layer model which is presented, it is shown that the thermal advection and stratification of the entire boundary layer as well as the geostrophic forcing influence the modeled near surface wind and wind stress profiles. Evidence of diurnal variation in the stratification under barotropic conditions is found in the data, with the daytime marine boundary layer being more convective than its nighttime counterpart. The temporal and spacial sampling pattern of the satellite makes it impossible to recover the full diurnal cycle, however. The observed effects of the thermal advection are shown to be statistically significant during the day (and presumed more convective) hours, causing a systematic increase in the poleward transport of mass and heat. The statistical results are in a qualitative agreement with the model simulations and cannot be reproduced in randomized control tests.
Statistical relationship between the succeeding solar flares detected by the RHESSI satellite
NASA Astrophysics Data System (ADS)
Balázs, L. G.; Gyenge, N.; Korsós, M. B.; Baranyi, T.; Forgács-Dajka, E.; Ballai, I.
2014-06-01
The Reuven Ramaty High Energy Solar Spectroscopic Imager has observed more than 80 000 solar energetic events since its launch on 2002 February 12. Using this large sample of observed flares, we studied the spatiotemporal relationship between succeeding flares. Our results show that the statistical relationship between the temporal and spatial differences of succeeding flares can be described as a power law of the form R(t) ˜ tp with p = 0.327 ± 0.007. We discuss the possible interpretations of this result as a characteristic function of a supposed underlying physics. Different scenarios are considered to explain this relation, including the case where the connectivity between succeeding events is realized through a shock wave in the post Sedov-Taylor phase or where the spatial and temporal relationship between flares is supposed to be provided by an expanding flare area in the sub-diffusive regime. Furthermore, we cannot exclude the possibility that the physical process behind the statistical relationship is the reordering of the magnetic field by the flare or it is due to some unknown processes.
NASA Astrophysics Data System (ADS)
Gottwald, Georg A.; Wormell, J. P.; Wouters, Jeroen
2016-09-01
Using a sensitive statistical test we determine whether or not one can detect the breakdown of linear response given observations of deterministic dynamical systems. A goodness-of-fit statistics is developed for a linear statistical model of the observations, based on results for central limit theorems for deterministic dynamical systems, and used to detect linear response breakdown. We apply the method to discrete maps which do not obey linear response and show that the successful detection of breakdown depends on the length of the time series, the magnitude of the perturbation and on the choice of the observable. We find that in order to reliably reject the assumption of linear response for typical observables sufficiently large data sets are needed. Even for simple systems such as the logistic map, one needs of the order of 106 observations to reliably detect the breakdown with a confidence level of 95 %; if less observations are available one may be falsely led to conclude that linear response theory is valid. The amount of data required is larger the smaller the applied perturbation. For judiciously chosen observables the necessary amount of data can be drastically reduced, but requires detailed a priori knowledge about the invariant measure which is typically not available for complex dynamical systems. Furthermore we explore the use of the fluctuation-dissipation theorem (FDT) in cases with limited data length or coarse-graining of observations. The FDT, if applied naively to a system without linear response, is shown to be very sensitive to the details of the sampling method, resulting in erroneous predictions of the response.
Phenotypic Association Analyses With Copy Number Variation in Recurrent Depressive Disorder.
Rucker, James J H; Tansey, Katherine E; Rivera, Margarita; Pinto, Dalila; Cohen-Woods, Sarah; Uher, Rudolf; Aitchison, Katherine J; Craddock, Nick; Owen, Michael J; Jones, Lisa; Jones, Ian; Korszun, Ania; Barnes, Michael R; Preisig, Martin; Mors, Ole; Maier, Wolfgang; Rice, John; Rietschel, Marcella; Holsboer, Florian; Farmer, Anne E; Craig, Ian W; Scherer, Stephen W; McGuffin, Peter; Breen, Gerome
2016-02-15
Defining the molecular genomic basis of the likelihood of developing depressive disorder is a considerable challenge. We previously associated rare, exonic deletion copy number variants (CNV) with recurrent depressive disorder (RDD). Sex chromosome abnormalities also have been observed to co-occur with RDD. In this reanalysis of our RDD dataset (N = 3106 cases; 459 screened control samples and 2699 population control samples), we further investigated the role of larger CNVs and chromosomal abnormalities in RDD and performed association analyses with clinical data derived from this dataset. We found an enrichment of Turner's syndrome among cases of depression compared with the frequency observed in a large population sample (N = 34,910) of live-born infants collected in Denmark (two-sided p = .023, odds ratio = 7.76 [95% confidence interval = 1.79-33.6]), a case of diploid/triploid mosaicism, and several cases of uniparental isodisomy. In contrast to our previous analysis, large deletion CNVs were no more frequent in cases than control samples, although deletion CNVs in cases contained more genes than control samples (two-sided p = .0002). After statistical correction for multiple comparisons, our data do not support a substantial role for CNVs in RDD, although (as has been observed in similar samples) occasional cases may harbor large variants with etiological significance. Genetic pleiotropy and sample heterogeneity suggest that very large sample sizes are required to study conclusively the role of genetic variation in mood disorders. Copyright © 2016 Society of Biological Psychiatry. Published by Elsevier Inc. All rights reserved.
78 FR 43002 - Proposed Collection; Comment Request for Revenue Procedure 2004-29
Federal Register 2010, 2011, 2012, 2013, 2014
2013-07-18
... comments concerning statistical sampling in Sec. 274 Context. DATES: Written comments should be received on... INFORMATION: Title: Statistical Sampling in Sec. 274 Contest. OMB Number: 1545-1847. Revenue Procedure Number: Revenue Procedure 2004-29. Abstract: Revenue Procedure 2004-29 prescribes the statistical sampling...
42 CFR 1003.133 - Statistical sampling.
Code of Federal Regulations, 2014 CFR
2014-10-01
... 42 Public Health 5 2014-10-01 2014-10-01 false Statistical sampling. 1003.133 Section 1003.133 Public Health OFFICE OF INSPECTOR GENERAL-HEALTH CARE, DEPARTMENT OF HEALTH AND HUMAN SERVICES OIG AUTHORITIES CIVIL MONEY PENALTIES, ASSESSMENTS AND EXCLUSIONS § 1003.133 Statistical sampling. (a) In meeting...
Data assimilation and bathymetric inversion in a two-dimensional horizontal surf zone model
NASA Astrophysics Data System (ADS)
Wilson, G. W.; Ã-Zkan-Haller, H. T.; Holman, R. A.
2010-12-01
A methodology is described for assimilating observations in a steady state two-dimensional horizontal (2-DH) model of nearshore hydrodynamics (waves and currents), using an ensemble-based statistical estimator. In this application, we treat bathymetry as a model parameter, which is subject to a specified prior uncertainty. The statistical estimator uses state augmentation to produce posterior (inverse, updated) estimates of bathymetry, wave height, and currents, as well as their posterior uncertainties. A case study is presented, using data from a 2-D array of in situ sensors on a natural beach (Duck, NC). The prior bathymetry is obtained by interpolation from recent bathymetric surveys; however, the resulting prior circulation is not in agreement with measurements. After assimilating data (significant wave height and alongshore current), the accuracy of modeled fields is improved, and this is quantified by comparing with observations (both assimilated and unassimilated). Hence, for the present data, 2-DH bathymetric uncertainty is an important source of error in the model and can be quantified and corrected using data assimilation. Here the bathymetric uncertainty is ascribed to inadequate temporal sampling; bathymetric surveys were conducted on a daily basis, but bathymetric change occurred on hourly timescales during storms, such that hydrodynamic model skill was significantly degraded. Further tests are performed to analyze the model sensitivities used in the assimilation and to determine the influence of different observation types and sampling schemes.
EVALUATION OF A NEW MEAN SCALED AND MOMENT ADJUSTED TEST STATISTIC FOR SEM.
Tong, Xiaoxiao; Bentler, Peter M
2013-01-01
Recently a new mean scaled and skewness adjusted test statistic was developed for evaluating structural equation models in small samples and with potentially nonnormal data, but this statistic has received only limited evaluation. The performance of this statistic is compared to normal theory maximum likelihood and two well-known robust test statistics. A modification to the Satorra-Bentler scaled statistic is developed for the condition that sample size is smaller than degrees of freedom. The behavior of the four test statistics is evaluated with a Monte Carlo confirmatory factor analysis study that varies seven sample sizes and three distributional conditions obtained using Headrick's fifth-order transformation to nonnormality. The new statistic performs badly in most conditions except under the normal distribution. The goodness-of-fit χ(2) test based on maximum-likelihood estimation performed well under normal distributions as well as under a condition of asymptotic robustness. The Satorra-Bentler scaled test statistic performed best overall, while the mean scaled and variance adjusted test statistic outperformed the others at small and moderate sample sizes under certain distributional conditions.
Pullin, A N; Pairis-Garcia, M D; Campbell, B J; Campler, M R; Proudfoot, K L
2017-11-01
When considering methodologies for collecting behavioral data, continuous sampling provides the most complete and accurate data set whereas instantaneous sampling can provide similar results and also increase the efficiency of data collection. However, instantaneous time intervals require validation to ensure accurate estimation of the data. Therefore, the objective of this study was to validate scan sampling intervals for lambs housed in a feedlot environment. Feeding, lying, standing, drinking, locomotion, and oral manipulation were measured on 18 crossbred lambs housed in an indoor feedlot facility for 14 h (0600-2000 h). Data from continuous sampling were compared with data from instantaneous scan sampling intervals of 5, 10, 15, and 20 min using a linear regression analysis. Three criteria determined if a time interval accurately estimated behaviors: 1) ≥ 0.90, 2) slope not statistically different from 1 ( > 0.05), and 3) intercept not statistically different from 0 ( > 0.05). Estimations for lying behavior were accurate up to 20-min intervals, whereas feeding and standing behaviors were accurate only at 5-min intervals (i.e., met all 3 regression criteria). Drinking, locomotion, and oral manipulation demonstrated poor associations () for all tested intervals. The results from this study suggest that a 5-min instantaneous sampling interval will accurately estimate lying, feeding, and standing behaviors for lambs housed in a feedlot, whereas continuous sampling is recommended for the remaining behaviors. This methodology will contribute toward the efficiency, accuracy, and transparency of future behavioral data collection in lamb behavior research.
Design-based and model-based inference in surveys of freshwater mollusks
Dorazio, R.M.
1999-01-01
Well-known concepts in statistical inference and sampling theory are used to develop recommendations for planning and analyzing the results of quantitative surveys of freshwater mollusks. Two methods of inference commonly used in survey sampling (design-based and model-based) are described and illustrated using examples relevant in surveys of freshwater mollusks. The particular objectives of a survey and the type of information observed in each unit of sampling can be used to help select the sampling design and the method of inference. For example, the mean density of a sparsely distributed population of mollusks can be estimated with higher precision by using model-based inference or by using design-based inference with adaptive cluster sampling than by using design-based inference with conventional sampling. More experience with quantitative surveys of natural assemblages of freshwater mollusks is needed to determine the actual benefits of different sampling designs and inferential procedures.
NASA Technical Reports Server (NTRS)
Bernacca, P. L.
1971-01-01
The correlation between the equatorial velocities of the components of double stars is studied from a statistical standpoint. A theory of rotational correlation is developed and discussed with regard to its applicability to existing observations. The theory is then applied to a sample of visual binaries which are the least studied for rotational coupling. Consideration of eclipsing systems and spectroscopic binaries is limited to show how the degrees of freedom in the spin parallelism problem can be reduced. The analysis lends support to the existence of synchronism in closely spaced binaries.
Imaging Extended Emission-Line Regions of Obscured AGN with the Subaru Hyper Suprime-Cam Survey
NASA Astrophysics Data System (ADS)
Sun, Ai-Lei; Greene, Jenny E.; Zakamska, Nadia L.; Goulding, Andy; Strauss, Michael A.; Huang, Song; Johnson, Sean; Kawaguchi, Toshihiro; Matsuoka, Yoshiki; Marsteller, Alisabeth A.; Nagao, Tohru; Toba, Yoshiki
2018-05-01
Narrow-line regions excited by active galactic nuclei (AGN) are important for studying AGN photoionization and feedback. Their strong [O III] lines can be detected with broadband images, allowing morphological studies of these systems with large-area imaging surveys. We develop a new broad-band imaging technique to reconstruct the images of the [O III] line using the Subaru Hyper Suprime-Cam (HSC) Survey aided with spectra from the Sloan Digital Sky Survey (SDSS). The technique involves a careful subtraction of the galactic continuum to isolate emission from the [O III]λ5007 and [O III]λ4959 lines. Compared to traditional targeted observations, this technique is more efficient at covering larger samples without dedicated observational resources. We apply this technique to an SDSS spectroscopically selected sample of 300 obscured AGN at redshifts 0.1 - 0.7, uncovering extended emission-line region candidates with sizes up to tens of kpc. With the largest sample of uniformly derived narrow-line region sizes, we revisit the narrow-line region size - luminosity relation. The area and radii of the [O III] emission-line regions are strongly correlated with the AGN luminosity inferred from the mid-infrared (15 μm rest-frame) with a power-law slope of 0.62^{+0.05}_{-0.06}± 0.10 (statistical and systematic errors), consistent with previous spectroscopic findings. We discuss the implications for the physics of AGN emission-line regions and future applications of this technique, which should be useful for current and next-generation imaging surveys to study AGN photoionization and feedback with large statistical samples.
Uranium hydrogeochemical and stream sediment reconnaissance of the Solomon NTMS quadrangle, Alaska
DOE Office of Scientific and Technical Information (OSTI.GOV)
Langfeldt, S.L.; Youngquist, C.A.; D'Andrea, R.F. Jr.
This report presents results of a Hydrogeochemical and Stream Sediment Reconnaissance (HSSR) of the Solomon NTMS quadrangle, Alaska. In addition to this abbreviated data release, more complete data are available to the public in machine-readable form through the Grand Junction Office Information System at Oak Ridge National Laboratory. Presented in this data release are location data, field analyses, and laboratory analyses of several different sample media. For the sake of brevity, many field site observations have not been included in this volume. These data are, however, available on the magnetic tape. Appendices A and B describe the sample media andmore » summarize the analytical results for each medium. The data were subdivided by one of the Los Alamos National Laboratory (LANL) sorting programs of Zinkl and others into groups of stream sediment and stream water samples. For each group which contains a sufficient number of observations, statistical tables, tables of raw data, and 1:1000000 scale maps of pertinent elements have been included in this report. In addition, maps showing results of multivariate statistical analyses have been included. Further information about the HSSR program in general, or about the LANL portion of the program in particular, can be obtained in quarterly or semiannual program progress reports on open-file at DOE's Technical Library in Grand Junction. Information about the field and analytical procedures used by LANL during sample collection and analysis may be found in any HSSR data release prepared by the LANL and will not be included in this report.« less
Petry, Patrícia; Polli, Janaina B; Mattos, Vinícius F; Rosa, Rosana C M; Zen, Paulo R G; Graziadio, Carla; Paskulin, Giorgio A; Rosa, Rafael F M
2013-06-01
Trisomy 13 or Patau syndrome (PS) is a chromosomal disorder characterized by a well known presentation of multiple congenital anomalies. Our objective was to determine the clinical features and prognosis observed in a sample of patients with PS. The series was composed of patients with diagnosis of PS consecutively evaluated by a Clinical Genetics Service from a reference hospital of southern Brazil, in the period between 1975 and 2012. Statistical analysis was performed using PEPI program (version 4.0), with two-tailed Fisher's exact test for comparison of frequencies (P<0.05). The sample consisted of 30 patients, 60% male, median age at first evaluation of 9 days. Full trisomy of chromosome 13 was the main cytogenetic alteration (73%). The major clinical findings included: cryptorchidism (78%), abnormal auricles (77%), congenital heart defects (76%), polydactyly (63%), microphthalmia (60%) and micrognathia (50%). Four patients (13%) simultaneously had micro/anophthalmia, oral clefts and polydactyly. Some findings were only observed in our sample and included, among others, preauricular tags (10%), duplication of the hallux (3%) and spots following the lines of Blaschko (3%). Mosaicism (20% of cases) had a statistically significant association only with absence of cryptorchidism. The median of survival was 26 days. Patients with and without mosaicism had similar median of survival. Our findings, in agreement with the literature, show that the anomalies in patients with PS can be quite variable, sometimes even atypical. There is no pathognomonic finding, which may make the early identification of these patients challenging. Copyright © 2013 Wiley Periodicals, Inc.
Nursing Care as Perceived by Nurses Working in Disability Community Settings in Greece
Fotiadou, Elpida; Malliarou, Maria; Zetta, Stella; Gouva, Mary; Kotrotsiou, Evaggelia
2016-01-01
Introduction-Aim: The concept of nursing care in learning disability community settings has not been investigated in Greece. The aim of this paper is to investigate how nurses working in learning disability community settings perceive the meaning of nursing care. Material and Methods: The sample consisted of 100 nurses and nursing assistants working in a social care hospice. Participants were asked to answer questions about socio- demographic characteristics of the sample and fill in a questionnaire of care (GR-NDI-24), the “Job-Communication-Satisfaction-Importance” (JCSI) questionnaire and the altruism scale of Ahmed and Jackson. The data analysis was realized with statistical methods of descriptive and inductive statistics. The analysis was made with the use of SPSS (version 19). Results: The majority of the sample was women (78%). The majority of participants were married (66 %), DE graduates (66%) without postgraduate studies (96.7%). The mean age of respondents was 36.98±6.70 years. On the scales of caring and altruism, the mean values were 40.89±15.87 and 28.12±4.16 respectively. Very or fully satisfied with his work was 72% of the sample. The scope of work emerges as the most important factor influencing job satisfaction. The wages and working conditions (73% and 40% respectively) are the parameters of work which gathers the most dissatisfaction, while the salary is emerging as the most important parameter, the improvement of which would provide the highest satisfaction. Marginally statistically significant difference was observed in the range between TE graduates (d=40) and those of the DE grade (d=37), p=0.053. No statistically significant differences were observed in relation to other working and demographic characteristics (p>0.05). Greater care importance was associated with greater job satisfaction (p<0.01), while the latter was associated with high levels of altruism (p<0.05). Conclusion: The scope of work provides high satisfaction to nurses working in social care hospices, while the salary is not satisfactory. Nurses’ aides appeared highly sensitive to care issues. A multidimensional approach to the materiality of care and job satisfaction in future research will allow to further highlight all the aspects affecting job satisfaction and performance of nurses. This will identify critical parameters of nursing care in healthcare centers for the chronically ill. PMID:26383223
Nursing Care as Perceived by Nurses Working in Disability Community Settings in Greece.
Fotiadou, Elpida; Malliarou, Maria; Zetta, Stella; Gouva, Mary; Kotrotsiou, Evaggelia
2015-06-25
The concept of nursing care in learning disability community settings has not been investigated in Greece. The aim of this paper is to investigate how nurses working in learning disability community settings perceive the meaning of nursing care. The sample consisted of 100 nurses and nursing assistants working in a social care hospice. Participants were asked to answer questions about socio- demographic characteristics of the sample and fill in a questionnaire of care (GR-NDI-24), the "Job-Communication-Satisfaction-Importance" (JCSI) questionnaire and the altruism scale of Ahmed and Jackson. The data analysis was realized with statistical methods of descriptive and inductive statistics. The analysis was made with the use of SPSS (version 19). The majority of the sample was women (78%). The majority of participants were married (66 %), DE graduates (66%) without postgraduate studies (96.7%). The mean age of respondents was 36.98±6.70 years. On the scales of caring and altruism, the mean values were 40.89±15.87 and 28.12±4.16 respectively. Very or fully satisfied with his work was 72% of the sample. The scope of work emerges as the most important factor influencing job satisfaction. The wages and working conditions (73% and 40% respectively) are the parameters of work which gathers the most dissatisfaction, while the salary is emerging as the most important parameter, the improvement of which would provide the highest satisfaction. Marginally statistically significant difference was observed in the range between TE graduates (d=40) and those of the DE grade (d=37), p=0.053. No statistically significant differences were observed in relation to other working and demographic characteristics (p>0.05). Greater care importance was associated with greater job satisfaction (p<0.01), while the latter was associated with high levels of altruism (p<0.05). The scope of work provides high satisfaction to nurses working in social care hospices, while the salary is not satisfactory. Nurses' aides appeared highly sensitive to care issues. A multidimensional approach to the materiality of care and job satisfaction in future research will allow to further highlight all the aspects affecting job satisfaction and performance of nurses. This will identify critical parameters of nursing care in healthcare centers for the chronically ill.
78 FR 63568 - Proposed Collection; Comment Request for Rev. Proc. 2007-35
Federal Register 2010, 2011, 2012, 2013, 2014
2013-10-24
... Revenue Procedure 2007-35, Statistical Sampling for purposes of Section 199. DATES: Written comments... . SUPPLEMENTARY INFORMATION: Title: Statistical Sampling for purposes of Section 199. OMB Number: 1545-2072... statistical sampling may be used in purposes of section 199, which provides a deduction for income...
Laube, Norbert; Zimmermann, Diana J
2004-01-01
This study was performed to quantify the effect of a 1-week freezer storage of urine on its calcium oxalate crystallization risk. Calcium oxalate is the most common urinary stone material observed in urolithiasis patients in western and affluent countries. The BONN-Risk-Index of calcium oxalate crystallization risk in human urine is determined from a crystallization experiment performed on untreated native urine samples. We tested the influence of a 1-week freezing on the BONN-Risk-Index value as well as the effect of the sample freezing on the urinary osmolality. In vitro crystallization experiments in 49 native urine samples from stone-forming and non-stone forming individuals were performed in order to determine their calcium oxalate crystallization risk according to the BONN-Risk-Index approach. Comparison of the results derived from original sample investigations with those obtained from the thawed aliquots by statistical evaluation shows that i) no significant deviation from linearity between both results exists and ii) both results are identical by statistical means. This is valid for both, the BONN-Risk-Index and the osmolality data. The differences in the BONN-Risk-Index results of both procedures of BONN-Risk-Index determination, however, exceed the clinically acceptable difference. Thus, determination of the urinary calcium oxalate crystallization risk from thawed urine samples cannot be recommended.
NASA Astrophysics Data System (ADS)
Orović, Irena; Stanković, Srdjan; Amin, Moeness
2013-05-01
A modified robust two-dimensional compressive sensing algorithm for reconstruction of sparse time-frequency representation (TFR) is proposed. The ambiguity function domain is assumed to be the domain of observations. The two-dimensional Fourier bases are used to linearly relate the observations to the sparse TFR, in lieu of the Wigner distribution. We assume that a set of available samples in the ambiguity domain is heavily corrupted by an impulsive type of noise. Consequently, the problem of sparse TFR reconstruction cannot be tackled using standard compressive sensing optimization algorithms. We introduce a two-dimensional L-statistics based modification into the transform domain representation. It provides suitable initial conditions that will produce efficient convergence of the reconstruction algorithm. This approach applies sorting and weighting operations to discard an expected amount of samples corrupted by noise. The remaining samples serve as observations used in sparse reconstruction of the time-frequency signal representation. The efficiency of the proposed approach is demonstrated on numerical examples that comprise both cases of monocomponent and multicomponent signals.
NASA Astrophysics Data System (ADS)
Lartizien, Carole; Kinahan, Paul E.; Comtat, Claude; Lin, Michael; Swensson, Richard G.; Trebossen, Regine; Bendriem, Bernard
2000-04-01
This work presents initial results from observer detection performance studies using the same volume visualization software tools that are used in clinical PET oncology imaging. Research into the FORE+OSEM and FORE+AWOSEM statistical image reconstruction methods tailored to whole- body 3D PET oncology imaging have indicated potential improvements in image SNR compared to currently used analytic reconstruction methods (FBP). To assess the resulting impact of these reconstruction methods on the performance of human observers in detecting and localizing tumors, we use a non- Monte Carlo technique to generate multiple statistically accurate realizations of 3D whole-body PET data, based on an extended MCAT phantom and with clinically realistic levels of statistical noise. For each realization, we add a fixed number of randomly located 1 cm diam. lesions whose contrast is varied among pre-calibrated values so that the range of true positive fractions is well sampled. The observer is told the number of tumors and, similar to the AFROC method, asked to localize all of them. The true positive fraction for the three algorithms (FBP, FORE+OSEM, FORE+AWOSEM) as a function of lesion contrast is calculated, although other protocols could be compared. A confidence level for each tumor is also recorded for incorporation into later AFROC analysis.
NASA Astrophysics Data System (ADS)
Graham, Wendy; Destouni, Georgia; Demmy, George; Foussereau, Xavier
1998-07-01
The methodology developed in Destouni and Graham [Destouni, G., Graham, W.D., 1997. The influence of observation method on local concentration statistics in the subsurface. Water Resour. Res. 33 (4) 663-676.] for predicting locally measured concentration statistics for solute transport in heterogeneous porous media under saturated flow conditions is applied to the prediction of conservative nonreactive solute transport in the vadose zone where observations are obtained by soil coring. Exact analytical solutions are developed for both the mean and variance of solute concentrations measured in discrete soil cores using a simplified physical model for vadose-zone flow and solute transport. Theoretical results show that while the ensemble mean concentration is relatively insensitive to the length-scale of the measurement, predictions of the concentration variance are significantly impacted by the sampling interval. Results also show that accounting for vertical heterogeneity in the soil profile results in significantly less spreading in the mean and variance of the measured solute breakthrough curves, indicating that it is important to account for vertical heterogeneity even for relatively small travel distances. Model predictions for both the mean and variance of locally measured solute concentration, based on independently estimated model parameters, agree well with data from a field tracer test conducted in Manatee County, Florida.
[Effect sizes, statistical power and sample sizes in "the Japanese Journal of Psychology"].
Suzukawa, Yumi; Toyoda, Hideki
2012-04-01
This study analyzed the statistical power of research studies published in the "Japanese Journal of Psychology" in 2008 and 2009. Sample effect sizes and sample statistical powers were calculated for each statistical test and analyzed with respect to the analytical methods and the fields of the studies. The results show that in the fields like perception, cognition or learning, the effect sizes were relatively large, although the sample sizes were small. At the same time, because of the small sample sizes, some meaningful effects could not be detected. In the other fields, because of the large sample sizes, meaningless effects could be detected. This implies that researchers who could not get large enough effect sizes would use larger samples to obtain significant results.
NASA Astrophysics Data System (ADS)
Adushkin, V. V.
- A statistical procedure is described for estimating the yields of underground nuclear tests at the former Soviet Semipalatinsk test site using the peak amplitudes of short-period surface waves observed at near-regional distances (Δ < 150 km) from these explosions. This methodology is then applied to data recorded from a large sample of the Semipalatinsk explosions, including the Soviet JVE explosion of September 14, 1988, and it is demonstrated that it provides seismic estimates of explosion yield which are typically within 20% of the yields determined for these same explosions using more accurate, non-seismic techniques based on near-source observations.
First Observation of the Doubly Cabibbo-Suppressed Decay of a Charmed Baryon: Λ_{c}^{+}→pK^{+}π^{-}.
Yang, S B; Tanida, K; Kim, B H; Adachi, I; Aihara, H; Asner, D M; Aulchenko, V; Aushev, T; Babu, V; Badhrees, I; Bakich, A M; Barberio, E; Bhardwaj, V; Bhuyan, B; Biswal, J; Bonvicini, G; Bozek, A; Bračko, M; Browder, T E; Červenkov, D; Chekelian, V; Chen, A; Cheon, B G; Chilikin, K; Chistov, R; Cho, K; Chobanova, V; Choi, Y; Cinabro, D; Dalseno, J; Danilov, M; Dash, N; Doležal, Z; Drásal, Z; Dutta, D; Eidelman, S; Farhat, H; Fast, J E; Ferber, T; Fulsom, B G; Gabyshev, N; Garmash, A; Gaur, V; Gillard, R; Goh, Y M; Goldenzweig, P; Greenwald, D; Grygier, J; Haba, J; Hamer, P; Hara, T; Hayasaka, K; Hayashii, H; Hou, W-S; Iijima, T; Inami, K; Inguglia, G; Ishikawa, A; Itoh, R; Iwasaki, Y; Jacobs, W W; Jaegle, I; Jeon, H B; Joo, K K; Julius, T; Kang, K H; Kato, E; Katrenko, P; Kiesling, C; Kim, D Y; Kim, H J; Kim, J B; Kim, K T; Kim, M J; Kim, S H; Kim, S K; Kim, Y J; Kinoshita, K; Kobayashi, N; Kodyš, P; Korpar, S; Križan, P; Krokovny, P; Kuhr, T; Kuzmin, A; Kwon, Y-J; Lange, J S; Lee, I S; Li, C H; Li, H; Li, L; Li, Y; Li Gioi, L; Libby, J; Liventsev, D; Lubej, M; Masuda, M; Matvienko, D; Miyabayashi, K; Miyata, H; Mizuk, R; Mohanty, G B; Moll, A; Moon, H K; Mussa, R; Nakano, E; Nakao, M; Nanut, T; Nath, K J; Nayak, M; Negishi, K; Niiyama, M; Nisar, N K; Nishida, S; Ogawa, S; Okuno, S; Olsen, S L; Pakhlova, G; Pal, B; Park, C W; Park, H; Pedlar, T K; Pestotnik, R; Petrič, M; Piilonen, L E; Pulvermacher, C; Rauch, J; Ritter, M; Rostomyan, A; Ryu, S; Sahoo, H; Sakai, Y; Sandilya, S; Santelj, L; Sanuki, T; Sato, Y; Savinov, V; Schlüter, T; Schneider, O; Schnell, G; Schwanda, C; Schwartz, A J; Seino, Y; Senyo, K; Seon, O; Seong, I S; Sevior, M E; Shebalin, V; Shibata, T-A; Shiu, J-G; Shwartz, B; Simon, F; Sohn, Y-S; Sokolov, A; Stanič, S; Starič, M; Stypula, J; Sumihama, M; Sumiyoshi, T; Takizawa, M; Tamponi, U; Teramoto, Y; Trabelsi, K; Trusov, V; Uchida, M; Uglov, T; Unno, Y; Uno, S; Urquijo, P; Usov, Y; Vanhoefer, P; Varner, G; Varvell, K E; Vinokurova, A; Vossen, A; Wagner, M N; Wang, C H; Wang, M-Z; Wang, P; Wang, X L; Watanabe, Y; Williams, K M; Won, E; Yamaoka, J; Yashchenko, S; Ye, H; Yelton, J; Yuan, C Z; Yusa, Y; Zhang, Z P; Zhilich, V; Zhulanov, V; Zupanc, A
2016-07-01
We report the first observation of the decay Λ_{c}^{+}→pK^{+}π^{-} using a 980 fb^{-1} data sample collected by the Belle detector at the KEKB asymmetric-energy e^{+}e^{-} collider. This is the first observation of a doubly Cabibbo-suppressed decay of a charmed baryon. We measure the branching ratio of this decay with respect to its Cabibbo-favored counterpart to be B(Λ_{c}^{+}→pK^{+}π^{-})/B(Λ_{c}^{+}→pK^{-}π^{+})=(2.35±0.27±0.21)×10^{-3}, where the uncertainties are statistical and systematic, respectively.
First Observation of the Doubly Cabibbo-Suppressed Decay of a Charmed Baryon: Λ c + → p K + π -
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yang, S. B.; Tanida, K.; Kim, B. H.
We report the first observation of the decay Λ + c→pK +π - using a 980 fb -1 data sample collected by the Belle detector at the KEKB asymmetric-energy e +e - collider. This is the first observation of a doubly Cabibbo-suppressed decay of a charmed baryon. We measure the branching ratio of this decay with respect to its Cabibbo-favored counterpart to be B(Λ +c→pK +π -)/B(Λ + c→pK -π +)=(2.35±0.27±0.21)×10 -3, where the uncertainties are statistical and systematic, respectively.
NASA Astrophysics Data System (ADS)
ten Veldhuis, Marie-Claire; Schleiss, Marc
2017-04-01
Urban catchments are typically characterised by a more flashy nature of the hydrological response compared to natural catchments. Predicting flow changes associated with urbanisation is not straightforward, as they are influenced by interactions between impervious cover, basin size, drainage connectivity and stormwater management infrastructure. In this study, we present an alternative approach to statistical analysis of hydrological response variability and basin flashiness, based on the distribution of inter-amount times. We analyse inter-amount time distributions of high-resolution streamflow time series for 17 (semi-)urbanised basins in North Carolina, USA, ranging from 13 to 238 km2 in size. We show that in the inter-amount-time framework, sampling frequency is tuned to the local variability of the flow pattern, resulting in a different representation and weighting of high and low flow periods in the statistical distribution. This leads to important differences in the way the distribution quantiles, mean, coefficient of variation and skewness vary across scales and results in lower mean intermittency and improved scaling. Moreover, we show that inter-amount-time distributions can be used to detect regulation effects on flow patterns, identify critical sampling scales and characterise flashiness of hydrological response. The possibility to use both the classical approach and the inter-amount-time framework to identify minimum observable scales and analyse flow data opens up interesting areas for future research.
Seven ways to increase power without increasing N.
Hansen, W B; Collins, L M
1994-01-01
Many readers of this monograph may wonder why a chapter on statistical power was included. After all, by now the issue of statistical power is in many respects mundane. Everyone knows that statistical power is a central research consideration, and certainly most National Institute on Drug Abuse grantees or prospective grantees understand the importance of including a power analysis in research proposals. However, there is ample evidence that, in practice, prevention researchers are not paying sufficient attention to statistical power. If they were, the findings observed by Hansen (1992) in a recent review of the prevention literature would not have emerged. Hansen (1992) examined statistical power based on 46 cohorts followed longitudinally, using nonparametric assumptions given the subjects' age at posttest and the numbers of subjects. Results of this analysis indicated that, in order for a study to attain 80-percent power for detecting differences between treatment and control groups, the difference between groups at posttest would need to be at least 8 percent (in the best studies) and as much as 16 percent (in the weakest studies). In order for a study to attain 80-percent power for detecting group differences in pre-post change, 22 of the 46 cohorts would have needed relative pre-post reductions of greater than 100 percent. Thirty-three of the 46 cohorts had less than 50-percent power to detect a 50-percent relative reduction in substance use. These results are consistent with other review findings (e.g., Lipsey 1990) that have shown a similar lack of power in a broad range of research topics. Thus, it seems that, although researchers are aware of the importance of statistical power (particularly of the necessity for calculating it when proposing research), they somehow are failing to end up with adequate power in their completed studies. This chapter argues that the failure of many prevention studies to maintain adequate statistical power is due to an overemphasis on sample size (N) as the only, or even the best, way to increase statistical power. It is easy to see how this overemphasis has come about. Sample size is easy to manipulate, has the advantage of being related to power in a straight-forward way, and usually is under the direct control of the researcher, except for limitations imposed by finances or subject availability. Another option for increasing power is to increase the alpha used for hypothesis-testing but, as very few researchers seriously consider significance levels much larger than the traditional .05, this strategy seldom is used. Of course, sample size is important, and the authors of this chapter are not recommending that researchers cease choosing sample sizes carefully. Rather, they argue that researchers should not confine themselves to increasing N to enhance power. It is important to take additional measures to maintain and improve power over and above making sure the initial sample size is sufficient. The authors recommend two general strategies. One strategy involves attempting to maintain the effective initial sample size so that power is not lost needlessly. The other strategy is to take measures to maximize the third factor that determines statistical power: effect size.
Information retrieval from wide-band meteorological data - An example
NASA Technical Reports Server (NTRS)
Adelfang, S. I.; Smith, O. E.
1983-01-01
The methods proposed by Smith and Adelfang (1981) and Smith et al. (1982) are used to calculate probabilities over rectangles and sectors of the gust magnitude-gust length plane; probabilities over the same regions are also calculated from the observed distributions and a comparison is also presented to demonstrate the accuracy of the statistical model. These and other statistical results are calculated from samples of Jimsphere wind profiles at Cape Canaveral. The results are presented for a variety of wavelength bands, altitudes, and seasons. It is shown that wind perturbations observed in Jimsphere wind profiles in various wavelength bands can be analyzed by using digital filters. The relationship between gust magnitude and gust length is modeled with the bivariate gamma distribution. It is pointed out that application of the model to calculate probabilities over specific areas of the gust magnitude-gust length plane can be useful in aerospace design.
Evolution of high-mass star-forming regions .
NASA Astrophysics Data System (ADS)
Giannetti, A.; Leurini, S.; Wyrowski, F.; Urquhart, J.; König, C.; Csengeri, T.; Güsten, R.; Menten, K. M.
Observational identification of a coherent evolutionary sequence for high-mass star-forming regions is still missing. We use the progressive heating of the gas caused by the feedback of high-mass young stellar objects to prove the statistical validity of the most common schemes used to observationally define an evolutionary sequence for high-mass clumps, and identify which physical process dominates in the different phases. From the spectroscopic follow-ups carried out towards the TOP100 sample between 84 and 365 km s^-1 giga hertz, we selected several multiplets of CH3CN, CH3CCH, and CH3OH lines to derive the physical properties of the gas in the clumps along the evolutionary sequence. We demonstrate that the evolutionary sequence is statistically valid, and we define intervals in L/M separating the compression, collapse and accretion, and disruption phases. The first hot cores and ZAMS stars appear at L/M≈10usk {L_ȯ}msun-1
A method to estimate the effect of deformable image registration uncertainties on daily dose mapping
Murphy, Martin J.; Salguero, Francisco J.; Siebers, Jeffrey V.; Staub, David; Vaman, Constantin
2012-01-01
Purpose: To develop a statistical sampling procedure for spatially-correlated uncertainties in deformable image registration and then use it to demonstrate their effect on daily dose mapping. Methods: Sequential daily CT studies are acquired to map anatomical variations prior to fractionated external beam radiotherapy. The CTs are deformably registered to the planning CT to obtain displacement vector fields (DVFs). The DVFs are used to accumulate the dose delivered each day onto the planning CT. Each DVF has spatially-correlated uncertainties associated with it. Principal components analysis (PCA) is applied to measured DVF error maps to produce decorrelated principal component modes of the errors. The modes are sampled independently and reconstructed to produce synthetic registration error maps. The synthetic error maps are convolved with dose mapped via deformable registration to model the resulting uncertainty in the dose mapping. The results are compared to the dose mapping uncertainty that would result from uncorrelated DVF errors that vary randomly from voxel to voxel. Results: The error sampling method is shown to produce synthetic DVF error maps that are statistically indistinguishable from the observed error maps. Spatially-correlated DVF uncertainties modeled by our procedure produce patterns of dose mapping error that are different from that due to randomly distributed uncertainties. Conclusions: Deformable image registration uncertainties have complex spatial distributions. The authors have developed and tested a method to decorrelate the spatial uncertainties and make statistical samples of highly correlated error maps. The sample error maps can be used to investigate the effect of DVF uncertainties on daily dose mapping via deformable image registration. An initial demonstration of this methodology shows that dose mapping uncertainties can be sensitive to spatial patterns in the DVF uncertainties. PMID:22320766
Testing the Isotropic Universe Using the Gamma-Ray Burst Data of Fermi/GBM
NASA Astrophysics Data System (ADS)
Řípa, Jakub; Shafieloo, Arman
2017-12-01
The sky distribution of gamma-ray bursts (GRBs) has been intensively studied by various groups for more than two decades. Most of these studies test the isotropy of GRBs based on their sky number density distribution. In this work, we propose an approach to test the isotropy of the universe through inspecting the isotropy of the properties of GRBs such as their duration, fluences, and peak fluxes at various energy bands and different timescales. We apply this method on the Fermi/Gamma-ray Burst Monitor (GBM) data sample containing 1591 GRBs. The most noticeable feature we found is near the Galactic coordinates l≈ 30^\\circ , b≈ 15^\\circ , and radius r≈ 20^\\circ {--}40^\\circ . The inferred probability for the occurrence of such an anisotropic signal (in a random isotropic sample) is derived to be less than a percent in some of the tests while the other tests give results consistent with isotropy. These are based on the comparison of the results from the real data with the randomly shuffled data samples. Considering the large number of statistics we used in this work (some of which are correlated with each other), we can anticipate that the detected feature could be a result of statistical fluctuations. Moreover, we noticed a considerably low number of GRBs in this particular patch, which might be due to some instrumentation or observational effects that can consequently affect our statistics through some systematics. Further investigation is highly desirable in order to clarify this result, e.g., utilizing a larger future Fermi/GBM data sample as well as data samples of other GRB missions and also looking for possible systematics.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ackermann, M.; Ajello, M.; Allafort, A.
We present a detailed statistical analysis of the correlation between radio and gamma-ray emission of the active galactic nuclei (AGNs) detected by Fermi during its first year of operation, with the largest data sets ever used for this purpose. We use both archival interferometric 8.4 GHz data (from the Very Large Array and ATCA, for the full sample of 599 sources) and concurrent single-dish 15 GHz measurements from the Owens Valley Radio Observatory (OVRO, for a sub sample of 199 objects). Our unprecedentedly large sample permits us to assess with high accuracy the statistical significance of the correlation, using amore » surrogate data method designed to simultaneously account for common-distance bias and the effect of a limited dynamical range in the observed quantities. We find that the statistical significance of a positive correlation between the centimeter radio and the broadband (E > 100 MeV) gamma-ray energy flux is very high for the whole AGN sample, with a probability of <10{sup -7} for the correlation appearing by chance. Using the OVRO data, we find that concurrent data improve the significance of the correlation from 1.6 x 10{sup -6} to 9.0 x 10{sup -8}. Our large sample size allows us to study the dependence of correlation strength and significance on specific source types and gamma-ray energy band. We find that the correlation is very significant (chance probability < 10{sup -7}) for both flat spectrum radio quasars and BL Lac objects separately; a dependence of the correlation strength on the considered gamma-ray energy band is also present, but additional data will be necessary to constrain its significance.« less
Ackermann, M.; Ajello, M.; Allafort, A.; ...
2011-10-12
We present a detailed statistical analysis of the correlation between radio and gamma-ray emission of the active galactic nuclei (AGNs) detected by Fermi during its first year of operation, with the largest data sets ever used for this purpose. We use both archival interferometric 8.4 GHz data (from the Very Large Array and ATCA, for the full sample of 599 sources) and concurrent single-dish 15 GHz measurements from the Owens Valley Radio Observatory (OVRO, for a sub sample of 199 objects). Our unprecedentedly large sample permits us to assess with high accuracy the statistical significance of the correlation, using amore » surrogate data method designed to simultaneously account for common-distance bias and the effect of a limited dynamical range in the observed quantities. We find that the statistical significance of a positive correlation between the centimeter radio and the broadband (E > 100 MeV) gamma-ray energy flux is very high for the whole AGN sample, with a probability of <10 –7 for the correlation appearing by chance. Using the OVRO data, we find that concurrent data improve the significance of the correlation from 1.6 × 10 –6 to 9.0 × 10 –8. Our large sample size allows us to study the dependence of correlation strength and significance on specific source types and gamma-ray energy band. As a result, we find that the correlation is very significant (chance probability < 10 –7) for both flat spectrum radio quasars and BL Lac objects separately; a dependence of the correlation strength on the considered gamma-ray energy band is also present, but additional data will be necessary to constrain its significance.« less
NASA Technical Reports Server (NTRS)
Ackermann, M.; Ajello, M.; Allafort, A.; Angelakis, E.; Axelsson, M.; Baldini, L.; Ballet, J.; Barbiellini, G.; Bastieri, D.; Bellazzini, R.;
2011-01-01
We present a detailed statistical analysis of the correlation between radio and gamma-ray emission of the active galactic nuclei (AGNs) detected by Fermi during its first year of operation, with the largest data sets ever used for this purpose.We use both archival interferometric 8.4 GHz data (from the Very Large Array and ATCA, for the full sample of 599 sources) and concurrent single-dish 15 GHz measurements from the OwensValley RadioObservatory (OVRO, for a sub sample of 199 objects). Our unprecedentedly large sample permits us to assess with high accuracy the statistical significance of the correlation, using a surrogate data method designed to simultaneously account for common-distance bias and the effect of a limited dynamical range in the observed quantities. We find that the statistical significance of a positive correlation between the centimeter radio and the broadband (E > 100 MeV) gamma-ray energy flux is very high for the whole AGN sample, with a probability of <10(exp -7) for the correlation appearing by chance. Using the OVRO data, we find that concurrent data improve the significance of the correlation from 1.6 10(exp -6) to 9.0 10(exp -8). Our large sample size allows us to study the dependence of correlation strength and significance on specific source types and gamma-ray energy band. We find that the correlation is very significant (chance probability < 10(exp -7)) for both flat spectrum radio quasars and BL Lac objects separately; a dependence of the correlation strength on the considered gamma-ray energy band is also present, but additional data will be necessary to constrain its significance.
ERIC Educational Resources Information Center
Ali, Sundus Muhsin; Hussein, Khalid Shakir
2014-01-01
This paper presents an attempt to verify the comparative power of two statistical features: Type/Token, and Hapax legomena/Token ratios (henceforth TTR and HTR). A corpus of ten novels is compiled. Then sixteen samples (each is 5,000 tokens in length) are taken randomly out of these novels as representative blocks. The researchers observe the way…
George L. Farnsworth; James D. Nichols; John R. Sauer; Steven G. Fancy; Kenneth H. Pollock; Susan A. Shriner; Theodore R. Simons
2005-01-01
Point counts are a standard sampling procedure for many bird species, but lingering concerns still exist about the quality of information produced from the method. It is well known that variation in observer ability and environmental conditions can influence the detection probability of birds in point counts, but many biologists have been reluctant to abandon point...
A Fatigue Management System for Sustained Military Operations
2008-03-31
Bradford, Brenda Jones, Margaret Campbell, Heather McCrory, Amy Campbell, Linda Mendez, Juan Cardenas, Beckie Moise, Samuel Cardenas, Fernando...always under the direct observation of research personnel or knowingly monitored from a central control station by closed circuit television, excluding...blocks 1, 5, 9, and 13). Statistical Analyses To determine the appropriate sample size for this study, a power analysis was based on the post
Kalra, Pinky; Suman, Ethel; Shenoy, Ramya; Suprabha, Baranya-Shrikrishna
2017-01-01
Background Endodontic instrumentation carries the risk of over extrusion of debris and bacteria. The technique used and the type of instrumentation influences this risk. Aim The purpose of this study was to evaluate and compare the K-file, ProTaper hand and ProTaper rotary instrumentation systems for the amount of apically extruded debris, irrigant solution and intracanal bacteria. Design Experimental single blinded randomized type of in vitro study with sample of 30 single rooted teeth. Endodontic access cavities were prepared and the root canals were filled with the suspension of E. faecalis. Myers and Montogomery Model was used to collect apically extruded debris and irrigant. Canals were prepared using K files, Hand protapers and Protaper rotary files. Statistical analysis Non Parametric test like Kruskal-Wallis and Mann-Whitney U test were applied to determine the significant differences among the group. Results Tests revealed statistically significant difference between the amount of debris and number of bacteria extruded by the ProTaper hand and the K-files. No statistically significant difference was observed between the amounts of irrigant extruded by the ProTaper hand and the K-file system. Statistically significant differences were observed between the amounts of bacteria and irrigant extruded by the ProTaper rotary and the Protaper hand. No statistically significant difference was observed between the amounts of debris extruded by the ProTaper hand and the K-file system. Conclusions Amount of apical extrusion of irrigant solution, bacteria and debris are significantly greater with K File instruments and least with Protaper rotary instruments. Key words:Protaper, rotary, periapical extrusion. PMID:28210445
A comprehensive review of arsenic levels in the semiconductor manufacturing industry.
Park, Donguk; Yang, Haengsun; Jeong, Jeeyeon; Ha, Kwonchul; Choi, Sangjun; Kim, Chinyon; Yoon, Chungsik; Park, Dooyong; Paek, Domyung
2010-11-01
This paper presents a summary of arsenic level statistics from air and wipe samples taken from studies conducted in fabrication operations. The main objectives of this study were not only to describe arsenic measurement data but also, through a literature review, to categorize fabrication workers in accordance with observed arsenic levels. All airborne arsenic measurements reported were included in the summary statistics for analysis of the measurement data. The arithmetic mean was estimated assuming a lognormal distribution from the geometric mean and the geometric standard deviation or the range. In addition, weighted arithmetic means (WAMs) were calculated based on the number of measurements reported for each mean. Analysis of variance (ANOVA) was employed to compare arsenic levels classified according to several categories such as the year, sampling type, location sampled, operation type, and cleaning technique. Nine papers were found reporting airborne arsenic measurement data from maintenance workers or maintenance areas in semiconductor chip-making plants. A total of 40 statistical summaries from seven articles were identified that represented a total of 423 airborne arsenic measurements. Arsenic exposure levels taken during normal operating activities in implantation operations (WAM = 1.6 μg m⁻³, no. of samples = 77, no. of statistical summaries = 2) were found to be lower than exposure levels of engineers who were involved in maintenance works (7.7 μg m⁻³, no. of samples = 181, no. of statistical summaries = 19). The highest level (WAM = 218.6 μg m⁻³) was associated with various maintenance works performed inside an ion implantation chamber. ANOVA revealed no significant differences in the WAM arsenic levels among the categorizations based on operation and sampling characteristics. Arsenic levels (56.4 μg m⁻³) recorded during maintenance works performed in dry conditions were found to be much higher than those from maintenance works in wet conditions (0.6 μg m⁻³). Arsenic levels from wipe samples in process areas after maintenance activities ranged from non-detectable to 146 μg cm⁻², indicating the potential for dispersion into the air and hence inhalation. We conclude that workers who are regularly or occasionally involved in maintenance work have higher potential for occupational exposure than other employees who are in charge of routine production work. In addition, fabrication workers can be classified into two groups based on the reviewed arsenic exposure levels: operators with potential for low levels of exposure and maintenance engineers with high levels of exposure. These classifications could be used as a basis for a qualitative ordinal ranking of exposure in an epidemiological study.
Boka, V; Arapostathis, K; Karagiannis, V; Kotsanos, N; van Loveren, C; Veerkamp, J
2017-03-01
To present: the normative data on dental fear and caries status; the dental fear cut-off points of young children in the city of Thessaloniki, Greece. Study Design: This is a cross-sectional study with two independent study groups. A first representative sample consisted of 1484 children from 15 primary public schools of Thessaloniki. A second sample consisted of 195 randomly selected age-matched children, all patients of the Postgraduate Paediatric Dental Clinic of Aristotle University of Thessaloniki. First sample: In order to select data on dental fear and caries, dental examination took place in the classroom with disposable mirrors and a penlight. All the children completed the Dental Subscale of the Children's Fear Survey Schedule (CFSS-DS). Second sample: In order to define the cut-off points of the CFSS-DS, dental treatment of the 195 children was performed at the University Clinic. Children⁁s dental fear was assessed using the CFSS-DS and their behaviour during dental treatment was observed by one calibrated examiner using the Venham scale. Statistical analysis of the data was performed with IBM SPSS Statistics 20 at a statistical significance level of <0.05. First sample: The mean CFSS-DS score was 27.1±10.8. Age was significantly (p<0.05) related to dental fear. Mean differences between boys and girls were not significant. Caries was not correlated with dental fear. Second sample: CFSS-DS< 33 was defined as 'no dental fear', scores 33-37 as 'borderline' and scores > 37 as 'dental fear'. In the first sample, 84.6% of the children did not suffer from dental fear (CFSS-DS<33). Dental fear was correlated to age and not to caries and gender. The dental fear cut-off point for the CFSS-DS was estimated at 37 for 6-12 year old children (33-37 borderlines).
Infants' statistical learning: 2- and 5-month-olds' segmentation of continuous visual sequences.
Slone, Lauren Krogh; Johnson, Scott P
2015-05-01
Past research suggests that infants have powerful statistical learning abilities; however, studies of infants' visual statistical learning offer differing accounts of the developmental trajectory of and constraints on this learning. To elucidate this issue, the current study tested the hypothesis that young infants' segmentation of visual sequences depends on redundant statistical cues to segmentation. A sample of 20 2-month-olds and 20 5-month-olds observed a continuous sequence of looming shapes in which unit boundaries were defined by both transitional probability and co-occurrence frequency. Following habituation, only 5-month-olds showed evidence of statistically segmenting the sequence, looking longer to a statistically improbable shape pair than to a probable pair. These results reaffirm the power of statistical learning in infants as young as 5 months but also suggest considerable development of statistical segmentation ability between 2 and 5 months of age. Moreover, the results do not support the idea that infants' ability to segment visual sequences based on transitional probabilities and/or co-occurrence frequencies is functional at the onset of visual experience, as has been suggested previously. Rather, this type of statistical segmentation appears to be constrained by the developmental state of the learner. Factors contributing to the development of statistical segmentation ability during early infancy, including memory and attention, are discussed. Copyright © 2015 Elsevier Inc. All rights reserved.
Inverse statistical physics of protein sequences: a key issues review.
Cocco, Simona; Feinauer, Christoph; Figliuzzi, Matteo; Monasson, Rémi; Weigt, Martin
2018-03-01
In the course of evolution, proteins undergo important changes in their amino acid sequences, while their three-dimensional folded structure and their biological function remain remarkably conserved. Thanks to modern sequencing techniques, sequence data accumulate at unprecedented pace. This provides large sets of so-called homologous, i.e. evolutionarily related protein sequences, to which methods of inverse statistical physics can be applied. Using sequence data as the basis for the inference of Boltzmann distributions from samples of microscopic configurations or observables, it is possible to extract information about evolutionary constraints and thus protein function and structure. Here we give an overview over some biologically important questions, and how statistical-mechanics inspired modeling approaches can help to answer them. Finally, we discuss some open questions, which we expect to be addressed over the next years.
Inverse statistical physics of protein sequences: a key issues review
NASA Astrophysics Data System (ADS)
Cocco, Simona; Feinauer, Christoph; Figliuzzi, Matteo; Monasson, Rémi; Weigt, Martin
2018-03-01
In the course of evolution, proteins undergo important changes in their amino acid sequences, while their three-dimensional folded structure and their biological function remain remarkably conserved. Thanks to modern sequencing techniques, sequence data accumulate at unprecedented pace. This provides large sets of so-called homologous, i.e. evolutionarily related protein sequences, to which methods of inverse statistical physics can be applied. Using sequence data as the basis for the inference of Boltzmann distributions from samples of microscopic configurations or observables, it is possible to extract information about evolutionary constraints and thus protein function and structure. Here we give an overview over some biologically important questions, and how statistical-mechanics inspired modeling approaches can help to answer them. Finally, we discuss some open questions, which we expect to be addressed over the next years.
MANCOVA for one way classification with homogeneity of regression coefficient vectors
NASA Astrophysics Data System (ADS)
Mokesh Rayalu, G.; Ravisankar, J.; Mythili, G. Y.
2017-11-01
The MANOVA and MANCOVA are the extensions of the univariate ANOVA and ANCOVA techniques to multidimensional or vector valued observations. The assumption of a Gaussian distribution has been replaced with the Multivariate Gaussian distribution for the vectors data and residual term variables in the statistical models of these techniques. The objective of MANCOVA is to determine if there are statistically reliable mean differences that can be demonstrated between groups later modifying the newly created variable. When randomization assignment of samples or subjects to groups is not possible, multivariate analysis of covariance (MANCOVA) provides statistical matching of groups by adjusting dependent variables as if all subjects scored the same on the covariates. In this research article, an extension has been made to the MANCOVA technique with more number of covariates and homogeneity of regression coefficient vectors is also tested.
Heidel, R Eric
2016-01-01
Statistical power is the ability to detect a significant effect, given that the effect actually exists in a population. Like most statistical concepts, statistical power tends to induce cognitive dissonance in hepatology researchers. However, planning for statistical power by an a priori sample size calculation is of paramount importance when designing a research study. There are five specific empirical components that make up an a priori sample size calculation: the scale of measurement of the outcome, the research design, the magnitude of the effect size, the variance of the effect size, and the sample size. A framework grounded in the phenomenon of isomorphism, or interdependencies amongst different constructs with similar forms, will be presented to understand the isomorphic effects of decisions made on each of the five aforementioned components of statistical power.
Pupek, Alex; Matthewson, Beverly; Whitman, Erin; Fullarton, Rachel; Chen, Yu
2017-08-28
The pneumatic tube system (PTS) is commonly used in modern clinical laboratories to provide quick specimen delivery. However, its impact on sample integrity and laboratory testing results are still debatable. In addition, each PTS installation and configuration is unique to its institution. We sought to validate our Swisslog PTS by comparing routine chemistry, hematology, coagulation and blood gas test results and sample integrity indices between duplicate samples transported either manually or by PTS. Duplicate samples were delivered to the core laboratory manually by human courier or via the Swisslog PTS. Head-to-head comparisons of 48 routine chemistry, hematology, coagulation and blood gas laboratory tests, and three sample integrity indices were conducted on 41 healthy volunteers and 61 adult patients. The PTS showed no impact on sample hemolysis, lipemia, or icterus indices (all p<0.05). Although alkaline phosphatase, total bilirubin and hemoglobin reached statistical significance (p=0.009, 0.027 and 0.012, respectively), all had very low average bias which ranged from 0.01% to 2%. Potassium, total hemoglobin and percent deoxyhemoglobin were statistically significant for the neonatal capillary tube study (p=0.011, 0.033 and 0.041, respectively) but no biases greater than ±4% were identified for these parameters. All observed differences of these 48 laboratory tests were not clinically significant. The modern PTS investigated in this study is acceptable for reliable sample delivery for routine chemistry, hematology, coagulation and blood gas (in syringe and capillary tube) laboratory tests.
Occurrence of Radio Minihalos in a Mass-Limited Sample of Galaxy Clusters
NASA Technical Reports Server (NTRS)
Giacintucci, Simona; Markevitch, Maxim; Cassano, Rossella; Venturi, Tiziana; Clarke, Tracy E.; Brunetti, Gianfranco
2017-01-01
We investigate the occurrence of radio minihalos-diffuse radio sources of unknown origin observed in the cores of some galaxy clusters-in a statistical sample of 58 clusters drawn from the Planck Sunyaev-Zeldovich cluster catalog using a mass cut (M(sub 500) greater than 6 x 10(exp 14) solar mass). We supplement our statistical sample with a similarly sized nonstatistical sample mostly consisting of clusters in the ACCEPT X-ray catalog with suitable X-ray and radio data, which includes lower-mass clusters. Where necessary (for nine clusters), we reanalyzed the Very Large Array archival radio data to determine whether a minihalo is present. Our total sample includes all 28 currently known and recently discovered radio minihalos, including six candidates. We classify clusters as cool-core or non-cool-core according to the value of the specific entropy floor in the cluster center, rederived or newly derived from the Chandra X-ray density and temperature profiles where necessary (for 27 clusters). Contrary to the common wisdom that minihalos are rare, we find that almost all cool cores-at least 12 out of 15 (80%)-in our complete sample of massive clusters exhibit minihalos. The supplementary sample shows that the occurrence of minihalos may be lower in lower-mass cool-core clusters. No minihalos are found in non-cool cores or "warm cores." These findings will help test theories of the origin of minihalos and provide information on the physical processes and energetics of the cluster cores.
Griffith, J.A.; Stehman, S.V.; Sohl, Terry L.; Loveland, Thomas R.
2003-01-01
Temporal trends in landscape pattern metrics describing texture, patch shape and patch size were evaluated in the US Middle Atlantic Coastal Plain Ecoregion. The landscape pattern metrics were calculated for a sample of land use/cover data obtained for four points in time from 1973-1992. The multiple sampling dates permit evaluation of trend, whereas availability of only two sampling dates allows only evaluation of change. Observed statistically significant trends in the landscape pattern metrics demonstrated that the sampling-based monitoring protocol was able to detect a trend toward a more fine-grained landscape in this ecoregion. This sampling and analysis protocol is being extended spatially to the remaining 83 ecoregions in the US and temporally to the year 2000 to provide a national and regional synthesis of the temporal and spatial dynamics of landscape pattern covering the period 1973-2000.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Surovchak, Scott; Miller, Michele
The 2008 Long-Term Surveillance Plan [LTSP] for the Decommissioned Hallam Nuclear Power Facility, Hallam, Nebraska (http://www.lm.doe.gov/Hallam/Documents.aspx) requires groundwater monitoring once every 2 years. Seventeen monitoring wells at the Hallam site were sampled during this event as specified in the plan. Planned monitoring locations are shown in Attachment 1, Sampling and Analysis Work Order. Water levels were measured at all sampled wells and at two additional wells (6A and 6B) prior to the start of sampling. Additionally, water levels of each sampled well were measured at the beginning of sampling. See Attachment 2, Trip Report, for additional details. Sampling and analysismore » were conducted as specified in Sampling and Analysis Plan for U.S. Department of Energy Office of Legacy Management Sites (LMS/PRO/S04351, continually updated, http://energy.gov/lm/downloads/sampling-and-analysis-plan-us-department- energy-office-legacy-management-sites). Gross alpha and gross beta are the only parameters that were detected at statistically significant concentrations. Time/concentration graphs of the gross alpha and gross beta data are included in Attachment 3, Data Presentation. The gross alpha and gross beta activity concentrations observed are consistent with values previously observed and are attributed to naturally occurring radionuclides (e.g., uranium and uranium decay chain products) in the groundwater.« less
IR Observations of a Complete Unbiased Sample of Bright Seyfert Galaxies
NASA Astrophysics Data System (ADS)
Malkan, Matthew; Bendo, George; Charmandaris, Vassilis; Smith, Howard; Spinoglio, Luigi; Tommasin, Silvia
2008-03-01
IR spectra will measure the 2 main energy-generating processes by which galactic nuclei shine: black hole accretion and star formation. Both of these play roles in galaxy evolution, and they appear connected. To obtain a complete sample of AGN, covering the range of luminosities and column-densities, we will combine 2 complete all-sky samples with complementary selections, minimally biased by dust obscuration: the 116 IRAS 12um AGN and the 41 Swift/BAT hard Xray AGN. These galaxies have been extensively studied across the entire EM spectrum. Herschel observations have been requested and will be synergistic with the Spitzer database. IRAC and MIPS imaging will allow us to separate the nuclear and galactic continua. We are completing full IR observations of the local AGN population, most of which have already been done. The only remaining observations we request are 10 IRS/HIRES, 57 MIPS-24 and 30 IRAC pointings. These high-quality observations of bright AGN in the bolometric-flux-limited samples should be completed, for the high legacy value of complete uniform datasets. We will measure quantitatively the emission at each wavelength arising from stars and from accretion in each galactic center. Since our complete samples come from flux-limited all-sky surveys in the IR and HX, we will calculate the bi-variate AGN and star formation Luminosity Functions for the local population of active galaxies, for comparison with higher redshifts.Our second aim is to understand the physical differences between AGN classes. This requires statistical comparisons of full multiwavelength observations of complete representative samples. If the difference between Sy1s and Sy2s is caused by orientation, their isotropic properties, including those of the surrounding galactic centers, should be similar. In contrast, if they are different evolutionary stages following a galaxy encounter, then we may find observational evidence that the circumnuclear ISM of Sy2s is relatively younger.
Personality characteristics of hospice volunteers as measured by Myers-Briggs Type Indicator.
Mitchell, C W; Shuff, I M
1995-12-01
A sample of hospice volunteers (n = 99) was administered the Myers-Briggs Type Indicator (Myers & McCaulley, 1985). Frequencies of types observed were compared to population sample (n = 1,105) frequencies. Results indicated that, as a whole, hospice volunteers preferred extraversion over introversion, intuition over sensing, and feeling over thinking. Analysis of four-and two-letter preference combinations also yielded statistically significant differences. Most notably, the sensing-intuitive function appeared pivotal in determining of hospice volunteering. Suggestions are offered as to why the sensing-intuition function appeared central to hospice volunteering. Results appeared consistent with Jungian personality theory.
REPORT FOR COMMERCIAL GRADE NICKEL CHARACTERIZATION AND BENCHMARKING
DOE Office of Scientific and Technical Information (OSTI.GOV)
None
2012-12-20
Oak Ridge Associated Universities (ORAU), under the Oak Ridge Institute for Science and Education (ORISE) contract, has completed the collection, sample analysis, and review of analytical results to benchmark the concentrations of gross alpha-emitting radionuclides, gross beta-emitting radionuclides, and technetium-99 in commercial grade nickel. This report presents methods, change management, observations, and statistical analysis of materials procured from sellers representing nine countries on four continents. The data suggest there is a low probability of detecting alpha- and beta-emitting radionuclides in commercial nickel. Technetium-99 was not detected in any samples, thus suggesting it is not present in commercial nickel.
Spatiotemporal Path-Matching for Comparisons Between Ground- Based and Satellite Lidar Measurements
NASA Technical Reports Server (NTRS)
Berkoff, Timothy A.; Valencia, Sandra; Welton, Ellsworth J.; Spinhirne, James D.
2005-01-01
The spatiotemporal sampling differences between ground-based and satellite lidar data can contribute to significant errors for direct measurement comparisons. Improvement in sample correspondence is examined by the use of radiosonde wind velocity to vary the time average in ground-based lidar data to spatially match coincident satellite lidar measurements. Results are shown for the 26 February 2004 GLAS/ICESat overflight of a ground-based lidar stationed at NASA GSFC. Statistical analysis indicates that improvement in signal correlation is expected under certain conditions, even when a ground-based observation is mismatched in directional orientation to the satellite track.
AgRISTARS: Supporting research. US crop calendars in support of the early warning project
NASA Technical Reports Server (NTRS)
Hodges, T. (Principal Investigator)
1981-01-01
The crop calendars produced for the Large Area Crop Inventory Experiment (LACIE) and crop calendar samples for Colorado, Iowa, Kansas, Minnesota, Montana, Nebraska, North Dakota, South Dakota, and Texas are presented. These calendars are based on weekly crop reporting district level observations of the percentage of various crops at several growth stages. A sample of the statistical treatments of the weekly data is provided. Four to five years of 50-percent dates for stages on a crop reporting district level for Arkansas, Iowa, Kentucky, Louisiana, Michigan, Mississippi, Ohio and Wisconsin are also given.
Driven-dissipative quantum Monte Carlo method for open quantum systems
NASA Astrophysics Data System (ADS)
Nagy, Alexandra; Savona, Vincenzo
2018-05-01
We develop a real-time full configuration-interaction quantum Monte Carlo approach to model driven-dissipative open quantum systems with Markovian system-bath coupling. The method enables stochastic sampling of the Liouville-von Neumann time evolution of the density matrix thanks to a massively parallel algorithm, thus providing estimates of observables on the nonequilibrium steady state. We present the underlying theory and introduce an initiator technique and importance sampling to reduce the statistical error. Finally, we demonstrate the efficiency of our approach by applying it to the driven-dissipative two-dimensional X Y Z spin-1/2 model on a lattice.
VizieR Online Data Catalog: Fundamental parameters of Kepler stars (Silva Aguirre+, 2015)
NASA Astrophysics Data System (ADS)
Silva Aguirre, V.; Davies, G. R.; Basu, S.; Christensen-Dalsgaard, J.; Creevey, O.; Metcalfe, T. S.; Bedding, T. R.; Casagrande, L.; Handberg, R.; Lund, M. N.; Nissen, P. E.; Chaplin, W. J.; Huber, D.; Serenelli, A. M.; Stello, D.; van Eylen, V.; Campante, T. L.; Elsworth, Y.; Gilliland, R. L.; Hekker, S.; Karoff, C.; Kawaler, S. D.; Kjeldsen, H.; Lundkvist, M. S.
2016-02-01
Our sample has been extracted from the 77 exoplanet host stars presented in Huber et al. (2013, Cat. J/ApJ/767/127). We have made use of the full time-base of observations from the Kepler satellite to uniformly determine precise fundamental stellar parameters, including ages, for a sample of exoplanet host stars where high-quality asteroseismic data were available. We devised a Bayesian procedure flexible in its input and applied it to different grids of models to study systematics from input physics and extract statistically robust properties for all stars. (4 data files).
Design of partially supervised classifiers for multispectral image data
NASA Technical Reports Server (NTRS)
Jeon, Byeungwoo; Landgrebe, David
1993-01-01
A partially supervised classification problem is addressed, especially when the class definition and corresponding training samples are provided a priori only for just one particular class. In practical applications of pattern classification techniques, a frequently observed characteristic is the heavy, often nearly impossible requirements on representative prior statistical class characteristics of all classes in a given data set. Considering the effort in both time and man-power required to have a well-defined, exhaustive list of classes with a corresponding representative set of training samples, this 'partially' supervised capability would be very desirable, assuming adequate classifier performance can be obtained. Two different classification algorithms are developed to achieve simplicity in classifier design by reducing the requirement of prior statistical information without sacrificing significant classifying capability. The first one is based on optimal significance testing, where the optimal acceptance probability is estimated directly from the data set. In the second approach, the partially supervised classification is considered as a problem of unsupervised clustering with initially one known cluster or class. A weighted unsupervised clustering procedure is developed to automatically define other classes and estimate their class statistics. The operational simplicity thus realized should make these partially supervised classification schemes very viable tools in pattern classification.
Complexity quantification of dense array EEG using sample entropy analysis.
Ramanand, Pravitha; Nampoori, V P N; Sreenivasan, R
2004-09-01
In this paper, a time series complexity analysis of dense array electroencephalogram signals is carried out using the recently introduced Sample Entropy (SampEn) measure. This statistic quantifies the regularity in signals recorded from systems that can vary from the purely deterministic to purely stochastic realm. The present analysis is conducted with an objective of gaining insight into complexity variations related to changing brain dynamics for EEG recorded from the three cases of passive, eyes closed condition, a mental arithmetic task and the same mental task carried out after a physical exertion task. It is observed that the statistic is a robust quantifier of complexity suited for short physiological signals such as the EEG and it points to the specific brain regions that exhibit lowered complexity during the mental task state as compared to a passive, relaxed state. In the case of mental tasks carried out before and after the performance of a physical exercise, the statistic can detect the variations brought in by the intermediate fatigue inducing exercise period. This enhances its utility in detecting subtle changes in the brain state that can find wider scope for applications in EEG based brain studies.
NASA Astrophysics Data System (ADS)
Roy, P. K.; Pal, S.; Banerjee, G.; Biswas Roy, M.; Ray, D.; Majumder, A.
2014-12-01
River is considered as one of the main sources of freshwater all over the world. Hence analysis and maintenance of this water resource is globally considered a matter of major concern. This paper deals with the assessment of surface water quality of the Ichamati river using multivariate statistical techniques. Eight distinct surface water quality observation stations were located and samples were collected. For the samples collected statistical techniques were applied to the physico-chemical parameters and depth of siltation. In this paper cluster analysis is done to determine the relations between surface water quality and siltation depth of river Ichamati. Multiple regressions and mathematical equation modeling have been done to characterize surface water quality of Ichamati river on the basis of physico-chemical parameters. It was found that surface water quality of the downstream river was different from the water quality of the upstream. The analysis of the water quality parameters of the Ichamati river clearly indicate high pollution load on the river water which can be accounted to agricultural discharge, tidal effect and soil erosion. The results further reveal that with the increase in depth of siltation, water quality degraded.
Patterson, Megan S; Goodson, Patricia
2017-05-01
Compulsive exercise, a form of unhealthy exercise often associated with prioritizing exercise and feeling guilty when exercise is missed, is a common precursor to and symptom of eating disorders. College-aged women are at high risk of exercising compulsively compared with other groups. Social network analysis (SNA) is a theoretical perspective and methodology allowing researchers to observe the effects of relational dynamics on the behaviors of people. SNA was used to assess the relationship between compulsive exercise and body dissatisfaction, physical activity, and network variables. Descriptive statistics were conducted using SPSS, and quadratic assignment procedure (QAP) analyses were conducted using UCINET. QAP regression analysis revealed a statistically significant model (R 2 = .375, P < .0001) predicting compulsive exercise behavior. Physical activity, body dissatisfaction, and network variables were statistically significant predictor variables in the QAP regression model. In our sample, women who are connected to "important" or "powerful" people in their network are likely to have higher compulsive exercise scores. This result provides healthcare practitioners key target points for intervention within similar groups of women. For scholars researching eating disorders and associated behaviors, this study supports looking into group dynamics and network structure in conjunction with body dissatisfaction and exercise frequency.
Talbott, Mariah J; Servid, Sarah A; Cavinato, Anna G; Van Eenennaam, Joel P; Doroshov, Serge I; Struffenegger, Peter; Webb, Molly A H
2014-02-01
Assessing stage of oocyte maturity in female sturgeon by calculating oocyte polarization index (PI) is a necessary tool for both conservation propagation managers and caviar producers to know when to hormonally induce spawning. We tested the assumption that sampling ovarian follicles from one section of one ovary is sufficient for calculating an oocyte PI representative of oocyte maturity for an individual animal. Short-wavelength near-infrared spectroscopy (SW-NIR) scans were performed on three positions per ovary for five fish prior to caviar harvest. Samples of ovarian follicles were subsequently taken from the exact location of the SW-NIR scans for calculation of oocyte PI and follicle diameter. Oocyte PI was statistically different though not biologically relevant within an ovary and between ovaries in four of five fish. Follicle diameter was statistically different but not biologically relevant within an ovary in three of five fish. There were no differences in follicle diameter between ovaries. No statistical differences were observed between SW-NIR spectra collected at different locations within an ovary or between ovaries. These results emphasize the importance of utilizing both oocyte PI measurement and progesterone-induced oocyte maturation assays while deciding when to hormonally induce spawning in sturgeon females.
Semi-empirical seismic relations of A-F stars from COROT and Kepler legacy data
NASA Astrophysics Data System (ADS)
Moya, A.; Suárez, J. C.; García Hernández, A.; Mendoza, M. A.
2017-10-01
Asteroseismology is witnessing a revolution, thanks to high-precise asteroseismic space data (MOST, COROT, Kepler, BRITE) and their large ground-based follow-up programs. Those instruments have provided an unprecedented large amount of information, which allows us to scrutinize its statistical properties in the quest for hidden relations among pulsational and/or physical observables. This approach might be particularly useful for stars whose pulsation content is difficult to interpret. This is the case of intermediate-mass classical pulsating stars (I.e. γ Dor, δ Scuti, hybrids) for which current theories do not properly predict the observed oscillation spectra. Here, we establish a first step in finding such hidden relations from data mining techniques for these stars. We searched for those hidden relations in a sample of δ Scuti and hybrid stars observed by COROT and Kepler (74 and 153, respectively). No significant correlations between pairs of observables were found. However, two statistically significant correlations emerged from multivariable correlations in the observed seismic data, which describe the total number of observed frequencies and the largest one, respectively. Moreover, three different sets of stars were found to cluster according to their frequency density distribution. Such sets are in apparent agreement with the asteroseismic properties commonly accepted for A-F pulsating stars.
Statistical analyses to support guidelines for marine avian sampling. Final report
Kinlan, Brian P.; Zipkin, Elise; O'Connell, Allan F.; Caldow, Chris
2012-01-01
Interest in development of offshore renewable energy facilities has led to a need for high-quality, statistically robust information on marine wildlife distributions. A practical approach is described to estimate the amount of sampling effort required to have sufficient statistical power to identify species-specific “hotspots” and “coldspots” of marine bird abundance and occurrence in an offshore environment divided into discrete spatial units (e.g., lease blocks), where “hotspots” and “coldspots” are defined relative to a reference (e.g., regional) mean abundance and/or occurrence probability for each species of interest. For example, a location with average abundance or occurrence that is three times larger the mean (3x effect size) could be defined as a “hotspot,” and a location that is three times smaller than the mean (1/3x effect size) as a “coldspot.” The choice of the effect size used to define hot and coldspots will generally depend on a combination of ecological and regulatory considerations. A method is also developed for testing the statistical significance of possible hotspots and coldspots. Both methods are illustrated with historical seabird survey data from the USGS Avian Compendium Database. Our approach consists of five main components: 1. A review of the primary scientific literature on statistical modeling of animal group size and avian count data to develop a candidate set of statistical distributions that have been used or may be useful to model seabird counts. 2. Statistical power curves for one-sample, one-tailed Monte Carlo significance tests of differences of observed small-sample means from a specified reference distribution. These curves show the power to detect "hotspots" or "coldspots" of occurrence and abundance at a range of effect sizes, given assumptions which we discuss. 3. A model selection procedure, based on maximum likelihood fits of models in the candidate set, to determine an appropriate statistical distribution to describe counts of a given species in a particular region and season. 4. Using a large database of historical at-sea seabird survey data, we applied this technique to identify appropriate statistical distributions for modeling a variety of species, allowing the distribution to vary by season. For each species and season, we used the selected distribution to calculate and map retrospective statistical power to detect hotspots and coldspots, and map pvalues from Monte Carlo significance tests of hotspots and coldspots, in discrete lease blocks designated by the U.S. Department of Interior, Bureau of Ocean Energy Management (BOEM). 5. Because our definition of hotspots and coldspots does not explicitly include variability over time, we examine the relationship between the temporal scale of sampling and the proportion of variance captured in time series of key environmental correlates of marine bird abundance, as well as available marine bird abundance time series, and use these analyses to develop recommendations for the temporal distribution of sampling to adequately represent both shortterm and long-term variability. We conclude by presenting a schematic “decision tree” showing how this power analysis approach would fit in a general framework for avian survey design, and discuss implications of model assumptions and results. We discuss avenues for future development of this work, and recommendations for practical implementation in the context of siting and wildlife assessment for offshore renewable energy development projects.
Sampling methods to the statistical control of the production of blood components.
Pereira, Paulo; Seghatchian, Jerard; Caldeira, Beatriz; Santos, Paula; Castro, Rosa; Fernandes, Teresa; Xavier, Sandra; de Sousa, Gracinda; de Almeida E Sousa, João Paulo
2017-12-01
The control of blood components specifications is a requirement generalized in Europe by the European Commission Directives and in the US by the AABB standards. The use of a statistical process control methodology is recommended in the related literature, including the EDQM guideline. The control reliability is dependent of the sampling. However, a correct sampling methodology seems not to be systematically applied. Commonly, the sampling is intended to comply uniquely with the 1% specification to the produced blood components. Nevertheless, on a purely statistical viewpoint, this model could be argued not to be related to a consistent sampling technique. This could be a severe limitation to detect abnormal patterns and to assure that the production has a non-significant probability of producing nonconforming components. This article discusses what is happening in blood establishments. Three statistical methodologies are proposed: simple random sampling, sampling based on the proportion of a finite population, and sampling based on the inspection level. The empirical results demonstrate that these models are practicable in blood establishments contributing to the robustness of sampling and related statistical process control decisions for the purpose they are suggested for. Copyright © 2017 Elsevier Ltd. All rights reserved.
Smooth quantile normalization.
Hicks, Stephanie C; Okrah, Kwame; Paulson, Joseph N; Quackenbush, John; Irizarry, Rafael A; Bravo, Héctor Corrada
2018-04-01
Between-sample normalization is a critical step in genomic data analysis to remove systematic bias and unwanted technical variation in high-throughput data. Global normalization methods are based on the assumption that observed variability in global properties is due to technical reasons and are unrelated to the biology of interest. For example, some methods correct for differences in sequencing read counts by scaling features to have similar median values across samples, but these fail to reduce other forms of unwanted technical variation. Methods such as quantile normalization transform the statistical distributions across samples to be the same and assume global differences in the distribution are induced by only technical variation. However, it remains unclear how to proceed with normalization if these assumptions are violated, for example, if there are global differences in the statistical distributions between biological conditions or groups, and external information, such as negative or control features, is not available. Here, we introduce a generalization of quantile normalization, referred to as smooth quantile normalization (qsmooth), which is based on the assumption that the statistical distribution of each sample should be the same (or have the same distributional shape) within biological groups or conditions, but allowing that they may differ between groups. We illustrate the advantages of our method on several high-throughput datasets with global differences in distributions corresponding to different biological conditions. We also perform a Monte Carlo simulation study to illustrate the bias-variance tradeoff and root mean squared error of qsmooth compared to other global normalization methods. A software implementation is available from https://github.com/stephaniehicks/qsmooth.
Expression Profiling of Nonpolar Lipids in Meibum From Patients With Dry Eye: A Pilot Study
Chen, Jianzhong; Keirsey, Jeremy K.; Green, Kari B.; Nichols, Kelly K.
2017-01-01
Purpose The purpose of this investigation was to characterize differentially expressed lipids in meibum samples from patients with dry eye disease (DED) in order to better understand the underlying pathologic mechanisms. Methods Meibum samples were collected from postmenopausal women with DED (PW-DED; n = 5) and a control group of postmenopausal women without DED (n = 4). Lipid profiles were analyzed by direct infusion full-scan electrospray ionization mass spectrometry (ESI-MS). An initial analysis of 145 representative peaks from four classes of lipids in PW-DED samples revealed that additional manual corrections for peak overlap and isotopes only slightly affected the statistical analysis. Therefore, analysis of uncorrected data, which can be applied to a greater number of peaks, was used to compare more than 500 lipid peaks common to PW-DED and control samples. Statistical analysis of peak intensities identified several lipid species that differed significantly between the two groups. Data from contact lens wearers with DED (CL-DED; n = 5) were also analyzed. Results Many species of the two types of diesters (DE) and very long chain wax esters (WE) were decreased by ∼20% in PW-DED, whereas levels of triacylglycerols were increased by an average of 39% ± 3% in meibum from PW-DED compared to that in the control group. Approximately the same reduction (20%) of similar DE and WE was observed for CL-DED. Conclusions Statistical analysis of peak intensities from direct infusion ESI-MS results identified differentially expressed lipids in meibum from dry eye patients. Further studies are warranted to support these findings. PMID:28426869
Statistical transformation and the interpretation of inpatient glucose control data.
Saulnier, George E; Castro, Janna C; Cook, Curtiss B
2014-03-01
To introduce a statistical method of assessing hospital-based non-intensive care unit (non-ICU) inpatient glucose control. Point-of-care blood glucose (POC-BG) data from hospital non-ICUs were extracted for January 1 through December 31, 2011. Glucose data distribution was examined before and after Box-Cox transformations and compared to normality. Different subsets of data were used to establish upper and lower control limits, and exponentially weighted moving average (EWMA) control charts were constructed from June, July, and October data as examples to determine if out-of-control events were identified differently in nontransformed versus transformed data. A total of 36,381 POC-BG values were analyzed. In all 3 monthly test samples, glucose distributions in nontransformed data were skewed but approached a normal distribution once transformed. Interpretation of out-of-control events from EWMA control chart analyses also revealed differences. In the June test data, an out-of-control process was identified at sample 53 with nontransformed data, whereas the transformed data remained in control for the duration of the observed period. Analysis of July data demonstrated an out-of-control process sooner in the transformed (sample 55) than nontransformed (sample 111) data, whereas for October, transformed data remained in control longer than nontransformed data. Statistical transformations increase the normal behavior of inpatient non-ICU glycemic data sets. The decision to transform glucose data could influence the interpretation and conclusions about the status of inpatient glycemic control. Further study is required to determine whether transformed versus nontransformed data influence clinical decisions or evaluation of interventions.
Cantarero, Samuel; Zafra-Gómez, Alberto; Ballesteros, Oscar; Navalón, Alberto; Reis, Marco S; Saraiva, Pedro M; Vílchez, José L
2011-01-01
In this work we present a monitoring study of linear alkylbenzene sulfonates (LAS) and insoluble soap performed on Spanish sewage sludge samples. This work focuses on finding statistical relations between LAS concentrations and insoluble soap in sewage sludge samples and variables related to wastewater treatment plants such as water hardness, population and treatment type. It is worth to mention that 38 samples, collected from different Spanish regions, were studied. The statistical tool we used was Principal Component Analysis (PC), in order to reduce the number of response variables. The analysis of variance (ANOVA) test and a non-parametric test such as the Kruskal-Wallis test were also studied through the estimation of the p-value (probability of obtaining a test statistic at least as extreme as the one that was actually observed, assuming that the null hypothesis is true) in order to study possible relations between the concentration of both analytes and the rest of variables. We also compared LAS and insoluble soap behaviors. In addition, the results obtained for LAS (mean value) were compared with the limit value proposed by the future Directive entitled "Working Document on Sludge". According to the results, the mean obtained for soap and LAS was 26.49 g kg(-1) and 6.15 g kg(-1) respectively. It is worth noting that LAS mean was significantly higher than the limit value (2.6 g kg(-1)). In addition, LAS and soap concentrations depend largely on water hardness. However, only LAS concentration depends on treatment type.
Pocket guide to transportation, 1999
DOT National Transportation Integrated Search
1998-12-01
Statistics published in this Pocket Guide to Transportation come from many different sources. Some statistics are based on samples and are subject to sampling variability. Statistics may also be subject to omissions and errors in reporting, recording...
Pocket guide to transportation, 2009
DOT National Transportation Integrated Search
2009-01-01
Statistics published in this Pocket Guide to Transportation come from many different sources. Some statistics are based on samples and are subject to sampling variability. Statistics may also be subject to omissions and errors in reporting, recording...
Pocket guide to transportation, 2013.
DOT National Transportation Integrated Search
2013-01-01
Abstract Statistics published in this Pocket Guide to Transportation come from many different sources. Some statistics are based on samples and are subject to sampling variability. Statistics may also be subject to omissions and errors in reporting, ...
Pocket guide to transportation, 2010
DOT National Transportation Integrated Search
2010-01-01
Statistics published in this Pocket Guide to Transportation come from many different sources. Some statistics are based on samples and are subject to sampling variability. Statistics may also be subject to omissions and errors in reporting, recording...
Comparative Financial Statistics for Public Two-Year Colleges: FY 1992 National Sample.
ERIC Educational Resources Information Center
Dickmeyer, Nathan; Cirino, Anna Marie
This report, the 15th in an annual series, provides comparative information derived from a national sample of 544 public two-year colleges, highlighting financial statistics for fiscal year 1991-92. The report offers space for colleges to compare their institutional statistics with data provided on national sample medians; quartile data for the…
Comparing Simulated and Theoretical Sampling Distributions of the U3 Person-Fit Statistic.
ERIC Educational Resources Information Center
Emons, Wilco H. M.; Meijer, Rob R.; Sijtsma, Klaas
2002-01-01
Studied whether the theoretical sampling distribution of the U3 person-fit statistic is in agreement with the simulated sampling distribution under different item response theory models and varying item and test characteristics. Simulation results suggest that the use of standard normal deviates for the standardized version of the U3 statistic may…
Westfall, Jacob; Kenny, David A; Judd, Charles M
2014-10-01
Researchers designing experiments in which a sample of participants responds to a sample of stimuli are faced with difficult questions about optimal study design. The conventional procedures of statistical power analysis fail to provide appropriate answers to these questions because they are based on statistical models in which stimuli are not assumed to be a source of random variation in the data, models that are inappropriate for experiments involving crossed random factors of participants and stimuli. In this article, we present new methods of power analysis for designs with crossed random factors, and we give detailed, practical guidance to psychology researchers planning experiments in which a sample of participants responds to a sample of stimuli. We extensively examine 5 commonly used experimental designs, describe how to estimate statistical power in each, and provide power analysis results based on a reasonable set of default parameter values. We then develop general conclusions and formulate rules of thumb concerning the optimal design of experiments in which a sample of participants responds to a sample of stimuli. We show that in crossed designs, statistical power typically does not approach unity as the number of participants goes to infinity but instead approaches a maximum attainable power value that is possibly small, depending on the stimulus sample. We also consider the statistical merits of designs involving multiple stimulus blocks. Finally, we provide a simple and flexible Web-based power application to aid researchers in planning studies with samples of stimuli.
Statistical and Machine Learning forecasting methods: Concerns and ways forward
Makridakis, Spyros; Assimakopoulos, Vassilios
2018-01-01
Machine Learning (ML) methods have been proposed in the academic literature as alternatives to statistical ones for time series forecasting. Yet, scant evidence is available about their relative performance in terms of accuracy and computational requirements. The purpose of this paper is to evaluate such performance across multiple forecasting horizons using a large subset of 1045 monthly time series used in the M3 Competition. After comparing the post-sample accuracy of popular ML methods with that of eight traditional statistical ones, we found that the former are dominated across both accuracy measures used and for all forecasting horizons examined. Moreover, we observed that their computational requirements are considerably greater than those of statistical methods. The paper discusses the results, explains why the accuracy of ML models is below that of statistical ones and proposes some possible ways forward. The empirical results found in our research stress the need for objective and unbiased ways to test the performance of forecasting methods that can be achieved through sizable and open competitions allowing meaningful comparisons and definite conclusions. PMID:29584784
Statistical Symbolic Execution with Informed Sampling
NASA Technical Reports Server (NTRS)
Filieri, Antonio; Pasareanu, Corina S.; Visser, Willem; Geldenhuys, Jaco
2014-01-01
Symbolic execution techniques have been proposed recently for the probabilistic analysis of programs. These techniques seek to quantify the likelihood of reaching program events of interest, e.g., assert violations. They have many promising applications but have scalability issues due to high computational demand. To address this challenge, we propose a statistical symbolic execution technique that performs Monte Carlo sampling of the symbolic program paths and uses the obtained information for Bayesian estimation and hypothesis testing with respect to the probability of reaching the target events. To speed up the convergence of the statistical analysis, we propose Informed Sampling, an iterative symbolic execution that first explores the paths that have high statistical significance, prunes them from the state space and guides the execution towards less likely paths. The technique combines Bayesian estimation with a partial exact analysis for the pruned paths leading to provably improved convergence of the statistical analysis. We have implemented statistical symbolic execution with in- formed sampling in the Symbolic PathFinder tool. We show experimentally that the informed sampling obtains more precise results and converges faster than a purely statistical analysis and may also be more efficient than an exact symbolic analysis. When the latter does not terminate symbolic execution with informed sampling can give meaningful results under the same time and memory limits.
ATLASGAL -- A molecular view of an unbiased sample of massive star forming clumps
NASA Astrophysics Data System (ADS)
Figura, Charles; Urquhart, James; Wyrowski, Friedrich; Giannetti, Andrea; Kim, Wonju
2018-01-01
Massive stars play an important role in many areas of astrophysics, from regulating star formation to driving the evolution of their host galaxy. Study of these stars is made difficult by their short evolutionary timescales, small populations and greater distances, and further complicated because they reach the main sequence while still shrouded in their natal clumps. As a result, many aspects of their formation are still poorly understood.We have assembled a large and statistically representative collection of massive star-forming environments that span all evolutionary stages of development by correlating mid-infrared and dust continnum surveys. We have conducted follow-up single-pointing observations toward a sample of approximately 600 of these clumps with the Mopra telescope using an 8 GHz bandwidth that spans some 27 molecular and mm-radio recombination line transitions. These lines trace a wide range of interstellar conditions with varying thermal, chemical, and kinematic properties. Many of these lines exhibit hyperfine structure allowing more detailed measurements of the clump environment (e.g. rotation temperatures and column densities).From these twenty-seven lines, we have identified thirteen line intensity ratios that strongly trace the evolutionary state of these clumps. We have investigated individual molecular and mm-radio recombination lines, contrasting these with radio and sub-mm continuum observations. We present a summary of the results of the statistical analysis of the sample, and compare them with previous similar studies to test their utility as chemical clocks of the evolutionary processes.
A more powerful test based on ratio distribution for retention noninferiority hypothesis.
Deng, Ling; Chen, Gang
2013-03-11
Rothmann et al. ( 2003 ) proposed a method for the statistical inference of fraction retention noninferiority (NI) hypothesis. A fraction retention hypothesis is defined as a ratio of the new treatment effect verse the control effect in the context of a time to event endpoint. One of the major concerns using this method in the design of an NI trial is that with a limited sample size, the power of the study is usually very low. This makes an NI trial not applicable particularly when using time to event endpoint. To improve power, Wang et al. ( 2006 ) proposed a ratio test based on asymptotic normality theory. Under a strong assumption (equal variance of the NI test statistic under null and alternative hypotheses), the sample size using Wang's test was much smaller than that using Rothmann's test. However, in practice, the assumption of equal variance is generally questionable for an NI trial design. This assumption is removed in the ratio test proposed in this article, which is derived directly from a Cauchy-like ratio distribution. In addition, using this method, the fundamental assumption used in Rothmann's test, that the observed control effect is always positive, that is, the observed hazard ratio for placebo over the control is greater than 1, is no longer necessary. Without assuming equal variance under null and alternative hypotheses, the sample size required for an NI trial can be significantly reduced if using the proposed ratio test for a fraction retention NI hypothesis.
Tupinambá, Rogerio Amaral; Claro, Cristiane Aparecida de Assis; Pereira, Cristiane Aparecida; Nobrega, Celestino José Prudente; Claro, Ana Paula Rosifini Alves
2017-01-01
ABSTRACT Introduction: Plasma-polymerized film deposition was created to modify metallic orthodontic brackets surface properties in order to inhibit bacterial adhesion. Methods: Hexamethyldisiloxane (HMDSO) polymer films were deposited on conventional (n = 10) and self-ligating (n = 10) stainless steel orthodontic brackets using the Plasma-Enhanced Chemical Vapor Deposition (PECVD) radio frequency technique. The samples were divided into two groups according to the kind of bracket and two subgroups after surface treatment. Scanning Electron Microscopy (SEM) analysis was performed to assess the presence of bacterial adhesion over samples surfaces (slot and wings region) and film layer integrity. Surface roughness was assessed by Confocal Interferometry (CI) and surface wettability, by goniometry. For bacterial adhesion analysis, samples were exposed for 72 hours to a Streptococcus mutans solution for biofilm formation. The values obtained for surface roughness were analyzed using the Mann-Whitney test while biofilm adhesion were assessed by Kruskal-Wallis and SNK test. Results: Significant statistical differences (p< 0.05) for surface roughness and bacterial adhesion reduction were observed on conventional brackets after surface treatment and between conventional and self-ligating brackets; no significant statistical differences were observed between self-ligating groups (p> 0.05). Conclusion: Plasma-polymerized film deposition was only effective on reducing surface roughness and bacterial adhesion in conventional brackets. It was also noted that conventional brackets showed lower biofilm adhesion than self-ligating brackets despite the absence of film. PMID:28902253
Dissolved oxygen as an indicator of bioavailable dissolved organic carbon in groundwater
Chapelle, Francis H.; Bradley, Paul M.; McMahon, Peter B.; Kaiser, Karl; Benner, Ron
2012-01-01
Concentrations of dissolved oxygen (DO) plotted vs. dissolved organic carbon (DOC) in groundwater samples taken from a coastal plain aquifer of South Carolina (SC) showed a statistically significant hyperbolic relationship. In contrast, DO-DOC plots of groundwater samples taken from the eastern San Joaquin Valley of California (CA) showed a random scatter. It was hypothesized that differences in the bioavailability of naturally occurring DOC might contribute to these observations. This hypothesis was examined by comparing nine different biochemical indicators of DOC bioavailability in groundwater sampled from these two systems. Concentrations of DOC, total hydrolysable neutral sugars (THNS), total hydrolysable amino acids (THAA), mole% glycine of THAA, initial bacterial cell counts, bacterial growth rates, and carbon dioxide production/consumption were greater in SC samples relative to CA samples. In contrast, the mole% glucose of THNS and the aromaticity (SUVA254) of DOC was greater in CA samples. Each of these indicator parameters were observed to change with depth in the SC system in a manner consistent with active biodegradation. These results are uniformly consistent with the hypothesis that the bioavailability of DOC is greater in SC relative to CA groundwater samples. This, in turn, suggests that the presence/absence of a hyperbolic DO-DOC relationship may be a qualitative indicator of relative DOC bioavailability in groundwater systems.
Comparison of norovirus RNA levels in outbreak-related oysters with background environmental levels.
Lowther, James A; Gustar, Nicole E; Hartnell, Rachel E; Lees, David N
2012-02-01
Norovirus is the principal agent of bivalve shellfish-associated gastroenteric illness worldwide. Numerous studies using PCR have demonstrated norovirus contamination in a significant proportion of both oyster and other bivalve shellfish production areas and ready-to-eat products. By comparison, the number of epidemiologically confirmed shellfish-associated outbreaks is relatively low. This suggests that factors other than the simple presence or absence of virus RNA are important contributors to the amount of illness reported. This study compares norovirus RNA levels in oyster samples strongly linked to norovirus or norovirus-type illness with the levels typically found in commercial production areas (non-outbreak-related samples). A statistically significant difference between norovirus levels in the two sets of samples was observed. The geometric mean of the levels in outbreak samples (1,048 copies per g) was almost one order of magnitude higher than for positive non-outbreak-related samples (121 copies per g). Further, while none of the outbreak-related samples contained fewer than 152 copies per g, the majority of positive results for non-outbreak-related samples was below this level. These observations support the concept of a dose-response for norovirus RNA levels in shellfish and could help inform the establishment of threshold criteria for risk management.
Banff study of pathologic changes in lung allograft biopsy specimens with donor-specific antibodies.
Wallace, William Dean; Li, Ning; Andersen, Claus B; Arrossi, A Valeria; Askar, Medhat; Berry, Gerry J; DeNicola, Matthew M; Neil, Desley A; Pavlisko, Elizabeth N; Reed, Elaine F; Remmelink, Myriam; Weigt, S Sam; Weynand, Birgit; Zhang, Jennifer Q; Budev, Marie M; Farver, Carol F
2016-01-01
The diagnosis of antibody-mediated rejection (AMR) in the lung transplant is still an area under investigation. We performed a blinded multicenter study to determine if any statistically significant histologic findings in transbronchial biopsy specimens from lung transplant patients correlate with the presence of donor-specific antibodies (DSAs). We asked 9 pathologists with experience in lung transplantation to evaluate 161 lung transplant biopsy specimens for various histologic parameters. The findings were correlated with antibody status positive for DSAs, positive for non-DSAs, and no antibodies (NABs) present. The significance of each histologic variable was reviewed. We found no statistically significant association with acute cellular rejection, airway inflammation, or bronchiolitis obliterans and the presence or absence of antibodies. However, biopsy specimens with DSAs had a statistically significant difference vs NABs in the setting of acute lung injury, with or without diffuse alveolar damage (p = 0.0008), in the presence of capillary neutrophilic inflammation (p = 0.0014), and in samples with endotheliitis (p = 0.0155). In samples with complement 4d staining, there was a trend but no statistically significant difference between specimens associated with DSAs and specimens with NABs. Capillary inflammation, acute lung injury, and endotheliitis significantly correlated with DSAs. The infrequently observed diffuse staining for complement 4d limits the usefulness of this stain. Copyright © 2016 International Society for Heart and Lung Transplantation. Published by Elsevier Inc. All rights reserved.
Experiments on Nucleation in Different Flow Regimes
NASA Technical Reports Server (NTRS)
Bayuzick, R. J.; Hofmeister, W. H.; Morton, C. M.; Robinson, M. B.
1999-01-01
The vast majority of metallic engineering materials are solidified from the liquid phase. Understanding the solidification process is essential to control microstructure, which in turn, determines the properties of materials. The genesis of solidification is nucleation, where the first stable solid forms from the liquid phase. Nucleation kinetics determine the degree of undercooling and phase selection. As such, it is important to understand nucleation phenomena in order to control solidification or glass formation in metals and alloys. Early experiments in nucleation kinetics were accomplished by droplet dispersion methods. Dilatometry was used by Turnbull and others, and more recently differential thermal analysis and differential scanning calorimetry have been used for kinetic studies. These techniques have enjoyed success; however, there are difficulties with these experiments. Since materials are dispersed in a medium, the character of the emulsion/metal interface affects the nucleation behavior. Statistics are derived from the large number of particles observed in a single experiment, but dispersions have a finite size distribution which adds to the uncertainty of the kinetic determinations. Even though temperature can be controlled quite well before the onset of nucleation, the release of the latent heat of fusion during nucleation of particles complicates the assumption of isothermality during these experiments. Containerless processing has enabled another approach to the study of nucleation kinetics. With levitation techniques it is possible to undercool one sample to nucleation repeatedly in a controlled manner, such that the statistics of the nucleation process can be derived from multiple experiments on a single sample. The authors have fully developed the analysis of nucleation experiments on single samples following the suggestions of Skripov. The advantage of these experiments is that the samples are directly observable. The nucleation temperature can be measured by noncontact optical pyrometry, the mass of the sample is known, and post processing analysis can be conducted on the sample. The disadvantages are that temperature measurement must have exceptionally high precision, and it is not possible to isolate specific heterogeneous sites as in droplet dispersions.
Broberg, Per
2013-07-19
One major concern with adaptive designs, such as the sample size adjustable designs, has been the fear of inflating the type I error rate. In (Stat Med 23:1023-1038, 2004) it is however proven that when observations follow a normal distribution and the interim result show promise, meaning that the conditional power exceeds 50%, type I error rate is protected. This bound and the distributional assumptions may seem to impose undesirable restrictions on the use of these designs. In (Stat Med 30:3267-3284, 2011) the possibility of going below 50% is explored and a region that permits an increased sample size without inflation is defined in terms of the conditional power at the interim. A criterion which is implicit in (Stat Med 30:3267-3284, 2011) is derived by elementary methods and expressed in terms of the test statistic at the interim to simplify practical use. Mathematical and computational details concerning this criterion are exhibited. Under very general conditions the type I error rate is preserved under sample size adjustable schemes that permit a raise. The main result states that for normally distributed observations raising the sample size when the result looks promising, where the definition of promising depends on the amount of knowledge gathered so far, guarantees the protection of the type I error rate. Also, in the many situations where the test statistic approximately follows a normal law, the deviation from the main result remains negligible. This article provides details regarding the Weibull and binomial distributions and indicates how one may approach these distributions within the current setting. There is thus reason to consider such designs more often, since they offer a means of adjusting an important design feature at little or no cost in terms of error rate.
Snow, Mathew S.; Clark, Sue B.; Morrison, Samuel S.; ...
2015-10-01
Particulate transport represents an important mechanism for actinides and fission products at the Earth's surface; soil samples taken in the early 1970's near the Subsurface Disposal Area (SDA) at Idaho National Laboratory (INL) provide a case study for examining the mechanisms and characteristics of actinide transport under arid conditions. Transuranic waste was disposed via shallow land burial at the SDA until shortly after a flooding event that occurred in 1969. In this study we analyze soils collected in the early 1970's for ¹³⁷Cs, ²⁴¹Am, and Pu using a combination of radiometric and mass spectrometric techniques. Two distinct ²⁴⁰Pu/²³⁹Pu isotopic ratiosmore » are observed for contamination from the SDA, with values ranging from at least 0.059 to 0.069. ²⁴¹Am concentrations are observed to increase only slightly in 0-4 cm soils over the ~40 year period since soil sampling, contrary to Markham's previous hypothesis that ²⁴¹Pu is principally associated with the 0-4 cm soil fractions (Markham 1978). The lack of statistical difference in ²⁴¹Am/²³⁹⁺²⁴⁰Pu ratios with depth suggests mechanical transport and mixing discrete contaminated particles under arid conditions. Occasional samples beyond the northeastern corner are observed to contain anomalously high Pu concentrations with corresponding low ²⁴⁰Pu/²³⁹Pu atoms ratios, suggesting the occurrence of "hot particles;" application of a background Pu subtraction results in calculated Pu atom ratios for the "hot particles" which are statistically similar to those observed in the northeastern corner. Taken together, our data suggests that flooding resulted in mechanical transport of contaminated particles into the area between the SDA and the flood containment dike in the northeastern corner, following which subsequent contamination spreading resulted from wind transport of discrete particles.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Snow, Mathew S.; Clark, Sue B.; Morrison, Samuel S.
Particulate transport represents an important mechanism for actinides and fission products at the Earth's surface; soil samples taken in the early 1970's near the Subsurface Disposal Area (SDA) at Idaho National Laboratory (INL) provide a case study for examining the mechanisms and characteristics of actinide transport under arid conditions. Transuranic waste was disposed via shallow land burial at the SDA until shortly after a flooding event that occurred in 1969. In this study we analyze soils collected in the early 1970's for ¹³⁷Cs, ²⁴¹Am, and Pu using a combination of radiometric and mass spectrometric techniques. Two distinct ²⁴⁰Pu/²³⁹Pu isotopic ratiosmore » are observed for contamination from the SDA, with values ranging from at least 0.059 to 0.069. ²⁴¹Am concentrations are observed to increase only slightly in 0-4 cm soils over the ~40 year period since soil sampling, contrary to Markham's previous hypothesis that ²⁴¹Pu is principally associated with the 0-4 cm soil fractions (Markham 1978). The lack of statistical difference in ²⁴¹Am/²³⁹⁺²⁴⁰Pu ratios with depth suggests mechanical transport and mixing discrete contaminated particles under arid conditions. Occasional samples beyond the northeastern corner are observed to contain anomalously high Pu concentrations with corresponding low ²⁴⁰Pu/²³⁹Pu atoms ratios, suggesting the occurrence of "hot particles;" application of a background Pu subtraction results in calculated Pu atom ratios for the "hot particles" which are statistically similar to those observed in the northeastern corner. Taken together, our data suggests that flooding resulted in mechanical transport of contaminated particles into the area between the SDA and the flood containment dike in the northeastern corner, following which subsequent contamination spreading resulted from wind transport of discrete particles.« less
Lonc, Elzbieta; Plewa, Kinga; Kiewra, Dorota; Szczepańska, Anna; Firling, Conrad E
2013-01-01
The qualitative and quantitative mycological composition of indoor areas of three private residencies and an academic dormitory in Wroclaw, Poland was investigated. Seasonal fungal samples were obtained using a MAS-100 air sampler. The samples were cultured on three different media: Sabouraud Agar (SAB), Dichloran Glycerol Selective Medium (DG18) and Malt Extract Agar (MEA). The number of colony forming unit (CFU) values ranged from 10 CFU/m3 to 490 CFU/m3 depending on the culture medium, season, and sampling site. The identification of the cultured fungi was performed using macro- and microscopic observations and diagnostic keys. Eleven fungal genera were identified. The most common fungi were members of genera Cladosporium, Penicillium, Aspergillus, Alternaria, and Fusarium; the least common fungi were members of genera Geotrichum and Paecilomyces. Seasonal variations in the concentration of fungi were observed with the highest concentration of fungi in the spring and the lowest concentration of fungi in the winter. There were no statistically significant correlations between fungal concentrations and the temperature or the relative humidity of the sample sites.
Variable Circumstellar Disks of “Classical” Be Stars
NASA Astrophysics Data System (ADS)
Gerhartz, Cody; Bjorkman, K. S.; Wisniewski, J. P.
2013-06-01
Circumstellar disks are common among many stars, all spectral types, and at different stages of their lifetimes. Among the near-main sequence “Classical” Be stars, there is growing evidence that these disks can form, dissipate, and reform, on timescales that are differ from case to case. We present data for a subset of cases where observations have been obtained throughout the different phases of the disk cycle. Using data obtained with the SpeX instrument at the NASA IRTF, we examine the IR spectral line variability of these stars to better understand the timescales and the physical mechanisms involved. The primary focus in this study are the V/R variations that are observed in the sample. The second stage of our project is to examine a sample of star clusters known to contain Be stars, with the goal to develop a more statistically significant sample of variable circumstellar disk systems. With a robust multi-epoch study we can determine whether these Be stars exhibit disk-loss or disk-renewal phases. The larger sample will enable a better understanding of the prevalence of these disk events.
The Importance of Introductory Statistics Students Understanding Appropriate Sampling Techniques
ERIC Educational Resources Information Center
Menil, Violeta C.
2005-01-01
In this paper the author discusses the meaning of sampling, the reasons for sampling, the Central Limit Theorem, and the different techniques of sampling. Practical and relevant examples are given to make the appropriate sampling techniques understandable to students of Introductory Statistics courses. With a thorough knowledge of sampling…
Dutta, Joydev; Chetia, Mridul; Misra, A K
2011-10-01
Contamination of drinking water by arsenic and other heavy metals and their related toxicology is a serious concern now-a-days. Millions of individual world-wide are suffering from the arsenic and other heavy metal related diseases due to the consumption of contaminated groundwater. 60 water samples from different sources of 6 small tea gardens of Sonitpur district were collected to study the potability of water for drinking purposes. The water samples collected from sources like tube wells, ring wells and ponds were analyzed for arsenic, heavy metals like iron, manganese and mercury with sodium, potassium, calcium, magnesium, pH, total hardness, chloride, fluoride and sulphate. Some drain water samples of the tea garden areas were also collected to analyze the above mentioned water parameters to see the contamination level. Experiments revealed that 78% samples of total collection had arsenic content above the permissible limit (0.01 ppm) of WHO guideline value for drinking water. The highest arsenic was observed 0.09 ppm at one sample of Gobindra Dahal tea garden of Gohpur sub division of Sonitpur district. 94% samples had contamination due to manganese 39% samples had iron and 44% samples had Hg. The water quality data was subjected to some statistical treatments like NDA, cluster analysis and pearson correlation to observe the distribution pattern of the different water quality parameters. A strong pearson correlation coefficient was observed between parameters-arsenic and manganese (0.865) and arsenic and mercury (0.837) at 0.01 level, indicated the same sources of drinking water contamination.
Statistical Investigation of Supersonic Downflows in the Transition Region above Sunspots
NASA Astrophysics Data System (ADS)
Samanta, Tanmoy; Tian, Hui; Prasad Choudhary, Debi
2018-06-01
Downflows at supersonic speeds have been observed in the transition region (TR) above sunspots for more than three decades. These downflows are often seen in different TR spectral lines above sunspots. We have performed a statistical investigation of these downflows using a large sample that was missing previously. The Interface Region Imaging Spectrograph (IRIS) has provided a wealth of observational data of sunspots at high spatial and spectral resolutions in the past few years. We have identified 60 data sets obtained with IRIS raster scans. Using an automated code, we identified the locations of strong downflows within these sunspots. We found that around 80% of our sample shows supersonic downflows in the Si IV 1403 Å line. These downflows mostly appear in the penumbral regions, though some of them are found in the umbrae. We also found that almost half of these downflows show signatures in chromospheric lines. Furthermore, a detailed spectral analysis was performed by selecting a small spectral window containing the O IV 1400/1401 Å and Si IV 1403 Å lines. Six Gaussian functions were simultaneously fitted to these three spectral lines and their satellite lines associated with the supersonic downflows. We calculated the intensity, Doppler velocity, and line width for these lines. Using the O IV 1400/1401 Å line ratio, we find that the downflow components are around one order of magnitude less dense than the regular components. Results from our statistical analysis suggest that these downflows may originate from the corona and that they are independent of the background TR plasma.
NASA Astrophysics Data System (ADS)
von Stetten, Eric Carl
The electron-positron momentum density has been measured by the two dimensional angular correlation of annihilation radiation (2D ACAR) technique for single crystal and polycrystalline (sintered powder) YBa_2 Cu_3O_{7-x} samples. For sintered superconducting and nonsuperconducting samples, the shape and temperature variation of the momentum density was investigated using the high sensitivity 2D ACAR technique. The possible existence of Fermi surfaces (FS's) in the YBa_2Cu _3O_{7-x} system was investigated in high precision 2D ACAR experiments on an oriented (twinned) single crystal superconducting YBa_2Cu _3O_{7-x} (x ~ 0.1) sample, at temperatures above and below the superconducting transition temperature (~85 K). These experiments were performed in the c-axis projection, in order to observe the theoretically predicted cylindrical FS's (if they exist) in a single experiment, without a full reconstruction of the three dimensional momentum density. Large differences were observed between the room temperature 2D ACAR spectra for superconducting and nonsuperconducting sintered powder samples, and smaller differences were observed between the spectra for similarly prepared superconducting samples. For sintered superconducting samples, complex sample dependent temperature variations of the momentum density were observed, in contrast to the small linear temperature variation observed for a sintered powder nonsuperconducting sample. These results are interpreted as manifestations of the theoretically predicted preferential sampling of the linear Cu-O chain region by the positron in the YBa _2Cu_3O _{7-x} system. High precision experiments on the single crystal superconducting sample revealed a nearly isotropic 2D ACAR spectrum, with only four small (~3% of the height at p_{x} = p _{y} = 0) peaks centered along the (110) symmetry lines. A small narrowing of the 2D ACAR spectrum was observed above T_{c}. The Brillouin-zone-reduced momentum density was formed using the "Lock-Crisp-West folding" technique, in order to identify possible FS signatures; several small features were observed that could possibly be due to FS's. A computer study of statistical noise propagation in 2D ACAR data, however, found that the possible FS signatures in the experimental data are similar in shape and magnitude to noise produced features.
Multi-Sensor Aerosol Products Sampling System
NASA Technical Reports Server (NTRS)
Petrenko, M.; Ichoku, C.; Leptoukh, G.
2011-01-01
Global and local properties of atmospheric aerosols have been extensively observed and measured using both spaceborne and ground-based instruments, especially during the last decade. Unique properties retrieved by the different instruments contribute to an unprecedented availability of the most complete set of complimentary aerosol measurements ever acquired. However, some of these measurements remain underutilized, largely due to the complexities involved in analyzing them synergistically. To characterize the inconsistencies and bridge the gap that exists between the sensors, we have established a Multi-sensor Aerosol Products Sampling System (MAPSS), which consistently samples and generates the spatial statistics (mean, standard deviation, direction and rate of spatial variation, and spatial correlation coefficient) of aerosol products from multiple spacebome sensors, including MODIS (on Terra and Aqua), MISR, OMI, POLDER, CALIOP, and SeaWiFS. Samples of satellite aerosol products are extracted over Aerosol Robotic Network (AERONET) locations as well as over other locations of interest such as those with available ground-based aerosol observations. In this way, MAPSS enables a direct cross-characterization and data integration between Level-2 aerosol observations from multiple sensors. In addition, the available well-characterized co-located ground-based data provides the basis for the integrated validation of these products. This paper explains the sampling methodology and concepts used in MAPSS, and demonstrates specific examples of using MAPSS for an integrated analysis of multiple aerosol products.
The MUSE-Wide survey: detection of a clustering signal from Lyman α emitters in the range 3 < z < 6
NASA Astrophysics Data System (ADS)
Diener, C.; Wisotzki, L.; Schmidt, K. B.; Herenz, E. C.; Urrutia, T.; Garel, T.; Kerutt, J.; Saust, R. L.; Bacon, R.; Cantalupo, S.; Contini, T.; Guiderdoni, B.; Marino, R. A.; Richard, J.; Schaye, J.; Soucail, G.; Weilbacher, P. M.
2017-11-01
We present a clustering analysis of a sample of 238 Ly α emitters at redshift 3 ≲ z ≲ 6 from the MUSE-Wide survey. This survey mosaics extragalactic legacy fields with 1h MUSE pointings to detect statistically relevant samples of emission line galaxies. We analysed the first year observations from MUSE-Wide making use of the clustering signal in the line-of-sight direction. This method relies on comparing pair-counts at close redshifts for a fixed transverse distance and thus exploits the full potential of the redshift range covered by our sample. A clear clustering signal with a correlation length of r0=2.9^{+1.0}_{-1.1} Mpc (comoving) is detected. Whilst this result is based on only about a quarter of the full survey size, it already shows the immense potential of MUSE for efficiently observing and studying the clustering of Ly α emitters.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1981-09-01
Results of a hydrogeochemical and stream sediment reconnaissance of the Philip Smith Mountains NTMS quadrangle, Alaska are presented. In addition to this abbreviated data release, more complete data are available to the public in machine-readable form. In this data release are location data, field analyses, and laboratory analyses of several different sample media. For the sake of brevity, many field site observations have not been included in this volume. These data are, however, available on the magnetic tape. Appendices A and B describe the sample media and summarize the analytical results for each medium. The data were subsetted by onemore » of the Los Alamos National Laboratory (LANL) sorting programs into groups of stream sediment and lake sediment samples. For each group which contains a sufficient number of observations, statistical tables, tables of raw data, and 1:1000000 scale maps of pertinent elements have been included in this report.« less
Structured Matrix Completion with Applications to Genomic Data Integration.
Cai, Tianxi; Cai, T Tony; Zhang, Anru
2016-01-01
Matrix completion has attracted significant recent attention in many fields including statistics, applied mathematics and electrical engineering. Current literature on matrix completion focuses primarily on independent sampling models under which the individual observed entries are sampled independently. Motivated by applications in genomic data integration, we propose a new framework of structured matrix completion (SMC) to treat structured missingness by design. Specifically, our proposed method aims at efficient matrix recovery when a subset of the rows and columns of an approximately low-rank matrix are observed. We provide theoretical justification for the proposed SMC method and derive lower bound for the estimation errors, which together establish the optimal rate of recovery over certain classes of approximately low-rank matrices. Simulation studies show that the method performs well in finite sample under a variety of configurations. The method is applied to integrate several ovarian cancer genomic studies with different extent of genomic measurements, which enables us to construct more accurate prediction rules for ovarian cancer survival.
García-Esteban, Marta; Ansorena, Diana; Astiasarán, Iciar
2004-05-01
Slices of dry-cured hams (Biceps femoris muscle) were stored during 8 weeks under vacuum and modified atmospheres (100% N(2) and a mixture of 20% CO(2) and 80% N(2)) in order to study the modifications on colour, texture and microbial counts during that period. Lightness was found to be more stable when samples were stored with 20% CO(2) and 80% N(2) without statistical differences between vacuum and 100% N(2). A slight whiteness was observed in the vacuum packed samples. Yellowness increased during time in vacuum packed samples, although no differences were found among the three conditions at the end of the study. Redness values were not affected by time or by the packaging system. With regard to texture, values found for all samples were within the normal range for this type of products, although it was observed that modified atmosphere packaging preserved samples better from hardening than vacuum packaging. No safety problems were detected in relation to the microbial quality in any case. In general, no clear differences were found among the three packaging systems for colour, texture and microbial quality in the storage conditions studied.
Variable Circumstellar Disks of “Classical” Be Stars, Part 2
NASA Astrophysics Data System (ADS)
Gerhartz, Cody; Davidson, J. W.; Bjorkman, K. S.; Wisniewski, J. P.
2014-01-01
Circumstellar disks are common among many stars, all spectral types, and at different stages of their lifetimes. Among the near-main sequence “Classical” Be stars, there is growing evidence that these disks can form, dissipate, and reform, on timescales that are differ from case to case. We present data for a subset of cases where observations have been obtained throughout the different phases of the disk cycle. Using data obtained with the SpeX instrument at the NASA IRTF, we examine the IR spectral line variability of these stars to better understand the timescales and the physical mechanisms involved. The primary focus in this study are the V/R variations that are observed in the sample. A complete run of all double-peaked velocity profiles in the sample is now complete. The second stage of our project is to examine a sample of star clusters known to contain Be stars, with the goal to develop a more statistically significant sample of variable circumstellar disk systems. With a robust multi-epoch study we can determine whether these Be stars exhibit disk-loss or disk-renewal phases. The larger sample will enable an understanding of the prevalence of these disk events.
Internal pilots for a class of linear mixed models with Gaussian and compound symmetric data
Gurka, Matthew J.; Coffey, Christopher S.; Muller, Keith E.
2015-01-01
SUMMARY An internal pilot design uses interim sample size analysis, without interim data analysis, to adjust the final number of observations. The approach helps to choose a sample size sufficiently large (to achieve the statistical power desired), but not too large (which would waste money and time). We report on recent research in cerebral vascular tortuosity (curvature in three dimensions) which would benefit greatly from internal pilots due to uncertainty in the parameters of the covariance matrix used for study planning. Unfortunately, observations correlated across the four regions of the brain and small sample sizes preclude using existing methods. However, as in a wide range of medical imaging studies, tortuosity data have no missing or mistimed data, a factorial within-subject design, the same between-subject design for all responses, and a Gaussian distribution with compound symmetry. For such restricted models, we extend exact, small sample univariate methods for internal pilots to linear mixed models with any between-subject design (not just two groups). Planning a new tortuosity study illustrates how the new methods help to avoid sample sizes that are too small or too large while still controlling the type I error rate. PMID:17318914
Determining significant material properties: A discovery approach
NASA Technical Reports Server (NTRS)
Karplus, Alan K.
1992-01-01
The following is a laboratory experiment designed to further understanding of materials science. The experiment itself can be informative for persons of any age past elementary school, and even for some in elementary school. The preparation of the plastic samples is readily accomplished by persons with resonable dexterity in the cutting of paper designs. The completion of the statistical Design of Experiments, which uses Yates' Method, requires basic math (addition and subtraction). Interpretive work requires plotting of data and making observations. Knowledge of statistical methods would be helpful. The purpose of this experiment is to acquaint students with the seven classes of recyclable plastics, and provide hands-on learning about the response of these plastics to mechanical tensile loading.
Quantifying Density Fluctuations in Volumes of All Shapes and Sizes Using Indirect Umbrella Sampling
NASA Astrophysics Data System (ADS)
Patel, Amish J.; Varilly, Patrick; Chandler, David; Garde, Shekhar
2011-10-01
Water density fluctuations are an important statistical mechanical observable and are related to many-body correlations, as well as hydrophobic hydration and interactions. Local water density fluctuations at a solid-water surface have also been proposed as a measure of its hydrophobicity. These fluctuations can be quantified by calculating the probability, P v ( N), of observing N waters in a probe volume of interest v. When v is large, calculating P v ( N) using molecular dynamics simulations is challenging, as the probability of observing very few waters is exponentially small, and the standard procedure for overcoming this problem (umbrella sampling in N) leads to undesirable impulsive forces. Patel et al. (J. Phys. Chem. B 114:1632, 2010) have recently developed an indirect umbrella sampling (INDUS) method, that samples a coarse-grained particle number to obtain P v ( N) in cuboidal volumes. Here, we present and demonstrate an extension of that approach to volumes of other basic shapes, like spheres and cylinders, as well as to collections of such volumes. We further describe the implementation of INDUS in the NPT ensemble and calculate P v ( N) distributions over a broad range of pressures. Our method may be of particular interest in characterizing the hydrophobicity of interfaces of proteins, nanotubes and related systems.
Jansson, Daniel; Lindström, Susanne Wiklund; Norlin, Rikard; Hok, Saphon; Valdez, Carlos A; Williams, Audrey M; Alcaraz, Armando; Nilsson, Calle; Åstot, Crister
2018-08-15
This work is part two of a three-part series in this issue of a Sweden-United States collaborative effort towards the understanding of the chemical attribution signatures of Russian VX (VR) in synthesized samples and complex food matrices. In this study, we describe the sourcing of VR present in food based on chemical analysis of attribution signatures by liquid chromatography-tandem mass spectrometry (LC-MS/MS) combined with multivariate data analysis. Analytical data was acquired from seven different foods spiked with VR batches that were synthesized via six different routes in two separate laboratories. The synthesis products were spiked at a lethal dose into seven food matrices: water, orange juice, apple purée, baby food, pea purée, liquid eggs and hot dog. After acetonitrile sample extraction, the samples were analyzed by LC-MS/MS operated in MRM mode. A multivariate statistical calibration model was built on the chemical attribution profiles from 118 VR spiked food samples. Using the model, an external test-set of the six synthesis routes employed for VR production was correctly identified with no observable major impact of the food matrices to the classification. The overall performance of the statistical models was found to be exceptional (94%) for the test set samples retrospectively classified to their synthesis routes. Copyright © 2018 Elsevier B.V. All rights reserved.
Lu, Qing; Kim, Jaegil; Straub, John E
2013-03-14
The generalized Replica Exchange Method (gREM) is extended into the isobaric-isothermal ensemble, and applied to simulate a vapor-liquid phase transition in Lennard-Jones fluids. Merging an optimally designed generalized ensemble sampling with replica exchange, gREM is particularly well suited for the effective simulation of first-order phase transitions characterized by "backbending" in the statistical temperature. While the metastable and unstable states in the vicinity of the first-order phase transition are masked by the enthalpy gap in temperature replica exchange method simulations, they are transformed into stable states through the parameterized effective sampling weights in gREM simulations, and join vapor and liquid phases with a succession of unimodal enthalpy distributions. The enhanced sampling across metastable and unstable states is achieved without the need to identify a "good" order parameter for biased sampling. We performed gREM simulations at various pressures below and near the critical pressure to examine the change in behavior of the vapor-liquid phase transition at different pressures. We observed a crossover from the first-order phase transition at low pressure, characterized by the backbending in the statistical temperature and the "kink" in the Gibbs free energy, to a continuous second-order phase transition near the critical pressure. The controlling mechanisms of nucleation and continuous phase transition are evident and the coexistence properties and phase diagram are found in agreement with literature results.
Li, Cen; Yang, Hongxia; Xiao, Yuancan; Zhandui; Sanglao; Wang, Zhang; Ladan, Duojie; Bi, Hongtao
2016-01-01
Zuotai (gTso thal) is one of the famous drugs containing mercury in Tibetan medicine. However, little is known about the chemical substance basis of its pharmacodynamics and the intrinsic link of different samples sources so far. Given this, energy dispersive spectrometry of X-ray (EDX), scanning electron microscopy (SEM), atomic force microscopy (AFM), and powder X-ray diffraction (XRD) were used to assay the elements, micromorphology, and phase composition of nine Zuotai samples from different regions, respectively; the XRD fingerprint features of Zuotai were analyzed by multivariate statistical analysis. EDX result shows that Zuotai contains Hg, S, O, Fe, Al, Cu, and other elements. SEM and AFM observations suggest that Zuotai is a kind of ancient nanodrug. Its particles are mainly in the range of 100–800 nm, which commonly further aggregate into 1–30 μm loosely amorphous particles. XRD test shows that β-HgS, S8, and α-HgS are its main phase compositions. XRD fingerprint analysis indicates that the similarity degrees of nine samples are very high, and the results of multivariate statistical analysis are broadly consistent with sample sources. The present research has revealed the physicochemical characteristics of Zuotai, and it would play a positive role in interpreting this mysterious Tibetan drug. PMID:27738409
Li, Cen; Yang, Hongxia; Du, Yuzhi; Xiao, Yuancan; Zhandui; Sanglao; Wang, Zhang; Ladan, Duojie; Bi, Hongtao; Wei, Lixin
2016-01-01
Zuotai ( gTso thal ) is one of the famous drugs containing mercury in Tibetan medicine. However, little is known about the chemical substance basis of its pharmacodynamics and the intrinsic link of different samples sources so far. Given this, energy dispersive spectrometry of X-ray (EDX), scanning electron microscopy (SEM), atomic force microscopy (AFM), and powder X-ray diffraction (XRD) were used to assay the elements, micromorphology, and phase composition of nine Zuotai samples from different regions, respectively; the XRD fingerprint features of Zuotai were analyzed by multivariate statistical analysis. EDX result shows that Zuotai contains Hg, S, O, Fe, Al, Cu, and other elements. SEM and AFM observations suggest that Zuotai is a kind of ancient nanodrug. Its particles are mainly in the range of 100-800 nm, which commonly further aggregate into 1-30 μ m loosely amorphous particles. XRD test shows that β -HgS, S 8 , and α -HgS are its main phase compositions. XRD fingerprint analysis indicates that the similarity degrees of nine samples are very high, and the results of multivariate statistical analysis are broadly consistent with sample sources. The present research has revealed the physicochemical characteristics of Zuotai , and it would play a positive role in interpreting this mysterious Tibetan drug.
Kotani, Akira; Tsutsumi, Risa; Shoji, Asaki; Hayashi, Yuzuru; Kusu, Fumiyo; Yamamoto, Kazuhiro; Hakamata, Hideki
2016-07-08
This paper puts forward a time and material-saving method for evaluating the repeatability of area measurements in gradient HPLC with UV detection (HPLC-UV), based on the function of mutual information (FUMI) theory which can theoretically provide the measurement standard deviation (SD) and detection limits through the stochastic properties of baseline noise with no recourse to repetitive measurements of real samples. The chromatographic determination of terbinafine hydrochloride and enalapril maleate is taken as an example. The best choice of the number of noise data points, inevitable for the theoretical evaluation, is shown to be 512 data points (10.24s at 50 point/s sampling rate of an A/D converter). Coupled with the relative SD (RSD) of sample injection variability in the instrument used, the theoretical evaluation is proved to give identical values of area measurement RSDs to those estimated by the usual repetitive method (n=6) over a wide concentration range of the analytes within the 95% confidence intervals of the latter RSD. The FUMI theory is not a statistical one, but the "statistical" reliability of its SD estimates (n=1) is observed to be as high as that attained by thirty-one measurements of the same samples (n=31). Copyright © 2016 Elsevier B.V. All rights reserved.
STATISTICAL SAMPLING AND DATA ANALYSIS
Research is being conducted to develop approaches to improve soil and sediment sampling techniques, measurement design and geostatistics, and data analysis via chemometric, environmetric, and robust statistical methods. Improvements in sampling contaminated soil and other hetero...
ERIC Educational Resources Information Center
Karadag, Engin; Caliskan, Nihat; Yesil, Rustu
2008-01-01
In this research, it is aimed to develop a scale to observe the body language which is used during an argument. A sample group of 266 teacher candidates study at the departments of Class, Turkish or Social Sciences at the Faculty of Education was used in this study. A logical and statistical approach was pursued during the development of scale. An…
Houts, Carrie R; Edwards, Michael C; Wirth, R J; Deal, Linda S
2016-11-01
There has been a notable increase in the advocacy of using small-sample designs as an initial quantitative assessment of item and scale performance during the scale development process. This is particularly true in the development of clinical outcome assessments (COAs), where Rasch analysis has been advanced as an appropriate statistical tool for evaluating the developing COAs using a small sample. We review the benefits such methods are purported to offer from both a practical and statistical standpoint and detail several problematic areas, including both practical and statistical theory concerns, with respect to the use of quantitative methods, including Rasch-consistent methods, with small samples. The feasibility of obtaining accurate information and the potential negative impacts of misusing large-sample statistical methods with small samples during COA development are discussed.
NASA Astrophysics Data System (ADS)
Niino, Yuu
2018-05-01
We investigate how the statistical properties of dispersion measure (DM) and apparent flux density/fluence of (nonrepeating) fast radio bursts (FRBs) are determined by unknown cosmic rate density history [ρ FRB(z)] and luminosity function (LF) of the transient events. We predict the distributions of DMs, flux densities, and fluences of FRBs taking account of the variation of the receiver efficiency within its beam, using analytical models of ρ FRB(z) and LF. Comparing the predictions with the observations, we show that the cumulative distribution of apparent fluences suggests that FRBs originate at cosmological distances and ρ FRB increases with redshift resembling the cosmic star formation history (CSFH). We also show that an LF model with a bright-end cutoff at log10 L ν (erg s‑1 Hz‑1) ∼ 34 are favored to reproduce the observed DM distribution if ρ FRB(z) ∝ CSFH, although the statistical significance of the constraints obtained with the current size of the observed sample is not high. Finally, we find that the correlation between DM and flux density of FRBs is potentially a powerful tool to distinguish whether FRBs are at cosmological distances or in the local universe more robustly with future observations.
Optical microvariability properties of BALQSOs
NASA Astrophysics Data System (ADS)
Joshi, Ravi; Chand, Hum; Gupta, Alok C.; Wiita, Paul J.
2011-04-01
We present optical light curves of 19 radio-quiet (RQ) broad absorption line (BAL) quasi-stellar objects (QSOs) and study their rapid variability characteristics. Systematic CCD observations, aided by a careful data analysis procedure, have allowed us to clearly detect any such microvariability exceeding 0.01-0.02 mag. Our observations cover a total of 13 nights (˜72 h) with each quasar monitored for about 4 h on a given night. Our sample size is a factor of 3 larger than the number of RQ BALQSOs previously searched for microvariability. We introduce a scaled F-test statistic for evaluating the presence of optical microvariability and demonstrate why it is generally preferable to the statistics usually employed for this purpose. Considering only unambiguous detections of microvariability we find that ˜11 per cent of RQ BALQSOs (two out of 19 sources) show microvariability for an individual observation length of about 4 h. This new duty cycle of 11 per cent is similar to the usual low microvariability fraction of normal radio-quiet QSOs (RQQSOs) with observation lengths similar to those of ours. This result provides support for models where RQ BALQSO do not appear to be a special case of the RQQSOs in terms of their microvariability properties.
NASA Technical Reports Server (NTRS)
Bommier, V.; Leroy, J. L.; Sahal-Brechot, S.
1985-01-01
The Hanle effect method for magnetic field vector diagnostics has now provided results on the magnetic field strength and direction in quiescent prominences, from linear polarization measurements in the He I E sub 3 line, performed at the Pic-du-Midi and at Sacramento Peak. However, there is an inescapable ambiguity in the field vector determination: each polarization measurement provides two field vector solutions symmetrical with respect to the line-of-sight. A statistical analysis capable of solving this ambiguity was applied to the large sample of prominences observed at the Pic-du-Midi (Leroy, et al., 1984); the same method of analysis applied to the prominences observed at Sacramento Peak (Athay, et al., 1983) provides results in agreement on the most probable magnetic structure of prominences; these results are detailed. The statistical results were confirmed on favorable individual cases: for 15 prominences observed at Pic-du-Midi, the two-field vectors are pointing on the same side of the prominence, and the alpha angles are large enough with respect to the measurements and interpretation inaccuracies, so that the field polarity is derived without any ambiguity.
NASA Technical Reports Server (NTRS)
Liu, Junjie; Bowman, Kevin W.; Lee, Memong; Henze, David K.; Bousserez, Nicolas; Brix, Holger; Collatz, G. James; Menemenlis, Dimitris; Ott, Lesley; Pawson, Steven;
2014-01-01
Using an Observing System Simulation Experiment (OSSE), we investigate the impact of JAXA Greenhouse gases Observing SATellite 'IBUKI' (GOSAT) sampling on the estimation of terrestrial biospheric flux with the NASA Carbon Monitoring System Flux (CMS-Flux) estimation and attribution strategy. The simulated observations in the OSSE use the actual column carbon dioxide (X(CO2)) b2.9 retrieval sensitivity and quality control for the year 2010 processed through the Atmospheric CO2 Observations from Space algorithm. CMS-Flux is a variational inversion system that uses the GEOS-Chem forward and adjoint model forced by a suite of observationally constrained fluxes from ocean, land and anthropogenic models. We investigate the impact of GOSAT sampling on flux estimation in two aspects: 1) random error uncertainty reduction and 2) the global and regional bias in posterior flux resulted from the spatiotemporally biased GOSAT sampling. Based on Monte Carlo calculations, we find that global average flux uncertainty reduction ranges from 25% in September to 60% in July. When aggregated to the 11 land regions designated by the phase 3 of the Atmospheric Tracer Transport Model Intercomparison Project, the annual mean uncertainty reduction ranges from 10% over North American boreal to 38% over South American temperate, which is driven by observational coverage and the magnitude of prior flux uncertainty. The uncertainty reduction over the South American tropical region is 30%, even with sparse observation coverage. We show that this reduction results from the large prior flux uncertainty and the impact of non-local observations. Given the assumed prior error statistics, the degree of freedom for signal is approx.1132 for 1-yr of the 74 055 GOSAT X(CO2) observations, which indicates that GOSAT provides approx.1132 independent pieces of information about surface fluxes. We quantify the impact of GOSAT's spatiotemporally sampling on the posterior flux, and find that a 0.7 gigatons of carbon bias in the global annual posterior flux resulted from the seasonally and diurnally biased sampling when using a diagonal prior flux error covariance.
TV Viewing and BMI by Race/Ethnicity and Socio-Economic Status
Shuval, Kerem; Gabriel, Kelley Pettee; Leonard, Tammy
2013-01-01
Objective To assess the association between TV viewing and obesity by race/ethnicity and socio-economic status. Design Cross-sectional analysis of 5,087 respondents to the Health Information National Trends Survey (HINTS), a nationally representative sample of US adults. Multivariate regression models were computed to assess the association between quartiles of TV viewing and BMI, stratified by race/ethnicity, educational attainment, employment and health insurance status. Results Findings indicate that increased TV viewing was associated with higher odds for being overweight/obese in the entire sample, while adjusting for physical activity and other confounders. After stratification by race/ethnicity, increased odds for overweight/obesity in the 3rd and 4th quartiles of TV viewing (e.g., 3rd quartile- cumulative OR = 1.43, 95%CI 1.07–1.92) was observed in non-Hispanic whites, with statistical significance. In non-Hispanic blacks and Hispanics, the odds were similar to whites, but did not reach statistical significance. Significant relations between greater TV viewing and increased BMI were observed in college graduates and non-graduates, those with health insurance and the employed. Conclusions This study extends previous research by examining potential inconsistencies in this association between various racial/ethnic groups and some socio-economic variables, which primarily were not found. PMID:23691070
The impact of obesity on specific airway resistance and conductance among schoolchildren.
Parraguez Arévalo, Andrea; Rojas Navarro, Francisco; Ruz Céspedes, Macarena; Medina González, Paul; Escobar Cabello, Máximo; Muñoz Cofré, Rodrigo
2018-04-01
Child and adolescent obesity is an epidemiological problem in developing countries. Its prevalence among preschoolers and schoolchildren is over 30%. It has been associated with a wide range of health complications, including rapid loss of lung function leading to changes in physiology and ventilatory mechanics. The objective of this study was to analyze the association between obesity and the increase in specific airway resistance (sRaw) in a sample of obese children and adolescents from the district of Talca. In a sample of 36 subjects with an average age of 9.38 ± 1.99 years, divided into 2 groups (normal weight and obese), the tricipital, subscapular, and abdominal skinfolds and lung volumes were measured. For the statistical analysis, data normality was determined and then the Student's t test or the Mann-Whitney U test and Pearson's or Spearman's correlations were used, as applicable. A value of p < 0.05 was considered statistically significant. When comparing normal weight and obese subjects, a significant increase in sRaw and a significant reduction in specific airway conductance (sGaw) were observed in obese subjects. In addition, an adequate and significant correlation was observed between sRaw and fat percentage. Obese subjects showed an increased sRaw and a reduced sGaw. Sociedad Argentina de Pediatría.
Single-cell forensic short tandem repeat typing within microfluidic droplets.
Geng, Tao; Novak, Richard; Mathies, Richard A
2014-01-07
A short tandem repeat (STR) typing method is developed for forensic identification of individual cells. In our strategy, monodisperse 1.5 nL agarose-in-oil droplets are produced with a high frequency using a microfluidic droplet generator. Statistically dilute single cells, along with primer-functionalized microbeads, are randomly compartmentalized in the droplets. Massively parallel single-cell droplet polymerase chain reaction (PCR) is performed to transfer replicas of desired STR targets from the single-cell genomic DNA onto the coencapsulated microbeads. These DNA-conjugated beads are subsequently harvested and reamplified under statistically dilute conditions for conventional capillary electrophoresis (CE) STR fragment size analysis. The 9-plex STR profiles of single cells from both pure and mixed populations of GM09947 and GM09948 human lymphoid cells show that all alleles are correctly called and allelic drop-in/drop-out is not observed. The cell mixture study exhibits a good linear relationship between the observed and input cell ratios in the range of 1:1 to 10:1. Additionally, the STR profile of GM09947 cells could be deduced even in the presence of a high concentration of cell-free contaminating 9948 genomic DNA. Our method will be valuable for the STR analysis of samples containing mixtures of cells/DNA from multiple contributors and for low-concentration samples.
Saulnier, George E; Castro, Janna C; Cook, Curtiss B
2014-05-01
Glucose control can be problematic in critically ill patients. We evaluated the impact of statistical transformation on interpretation of intensive care unit inpatient glucose control data. Point-of-care blood glucose (POC-BG) data derived from patients in the intensive care unit for 2011 was obtained. Box-Cox transformation of POC-BG measurements was performed, and distribution of data was determined before and after transformation. Different data subsets were used to establish statistical upper and lower control limits. Exponentially weighted moving average (EWMA) control charts constructed from April, October, and November data determined whether out-of-control events could be identified differently in transformed versus nontransformed data. A total of 8679 POC-BG values were analyzed. POC-BG distributions in nontransformed data were skewed but approached normality after transformation. EWMA control charts revealed differences in projected detection of out-of-control events. In April, an out-of-control process resulting in the lower control limit being exceeded was identified at sample 116 in nontransformed data but not in transformed data. October transformed data detected an out-of-control process exceeding the upper control limit at sample 27 that was not detected in nontransformed data. Nontransformed November results remained in control, but transformation identified an out-of-control event less than 10 samples into the observation period. Using statistical methods to assess population-based glucose control in the intensive care unit could alter conclusions about the effectiveness of care processes for managing hyperglycemia. Further study is required to determine whether transformed versus nontransformed data change clinical decisions about the interpretation of care or intervention results. © 2014 Diabetes Technology Society.
Saulnier, George E.; Castro, Janna C.
2014-01-01
Glucose control can be problematic in critically ill patients. We evaluated the impact of statistical transformation on interpretation of intensive care unit inpatient glucose control data. Point-of-care blood glucose (POC-BG) data derived from patients in the intensive care unit for 2011 was obtained. Box–Cox transformation of POC-BG measurements was performed, and distribution of data was determined before and after transformation. Different data subsets were used to establish statistical upper and lower control limits. Exponentially weighted moving average (EWMA) control charts constructed from April, October, and November data determined whether out-of-control events could be identified differently in transformed versus nontransformed data. A total of 8679 POC-BG values were analyzed. POC-BG distributions in nontransformed data were skewed but approached normality after transformation. EWMA control charts revealed differences in projected detection of out-of-control events. In April, an out-of-control process resulting in the lower control limit being exceeded was identified at sample 116 in nontransformed data but not in transformed data. October transformed data detected an out-of-control process exceeding the upper control limit at sample 27 that was not detected in nontransformed data. Nontransformed November results remained in control, but transformation identified an out-of-control event less than 10 samples into the observation period. Using statistical methods to assess population-based glucose control in the intensive care unit could alter conclusions about the effectiveness of care processes for managing hyperglycemia. Further study is required to determine whether transformed versus nontransformed data change clinical decisions about the interpretation of care or intervention results. PMID:24876620
Foundational Principles for Large-Scale Inference: Illustrations Through Correlation Mining.
Hero, Alfred O; Rajaratnam, Bala
2016-01-01
When can reliable inference be drawn in fue "Big Data" context? This paper presents a framework for answering this fundamental question in the context of correlation mining, wifu implications for general large scale inference. In large scale data applications like genomics, connectomics, and eco-informatics fue dataset is often variable-rich but sample-starved: a regime where the number n of acquired samples (statistical replicates) is far fewer than fue number p of observed variables (genes, neurons, voxels, or chemical constituents). Much of recent work has focused on understanding the computational complexity of proposed methods for "Big Data". Sample complexity however has received relatively less attention, especially in the setting when the sample size n is fixed, and the dimension p grows without bound. To address fuis gap, we develop a unified statistical framework that explicitly quantifies the sample complexity of various inferential tasks. Sampling regimes can be divided into several categories: 1) the classical asymptotic regime where fue variable dimension is fixed and fue sample size goes to infinity; 2) the mixed asymptotic regime where both variable dimension and sample size go to infinity at comparable rates; 3) the purely high dimensional asymptotic regime where the variable dimension goes to infinity and the sample size is fixed. Each regime has its niche but only the latter regime applies to exa cale data dimension. We illustrate this high dimensional framework for the problem of correlation mining, where it is the matrix of pairwise and partial correlations among the variables fua t are of interest. Correlation mining arises in numerous applications and subsumes the regression context as a special case. we demonstrate various regimes of correlation mining based on the unifying perspective of high dimensional learning rates and sample complexity for different structured covariance models and different inference tasks.
Measurement of Muon Neutrino Quasielastic Scattering on Carbon
NASA Astrophysics Data System (ADS)
Aguilar-Arevalo, A. A.; Bazarko, A. O.; Brice, S. J.; Brown, B. C.; Bugel, L.; Cao, J.; Coney, L.; Conrad, J. M.; Cox, D. C.; Curioni, A.; Djurcic, Z.; Finley, D. A.; Fleming, B. T.; Ford, R.; Garcia, F. G.; Garvey, G. T.; Green, C.; Green, J. A.; Hart, T. L.; Hawker, E.; Imlay, R.; Johnson, R. A.; Kasper, P.; Katori, T.; Kobilarcik, T.; Kourbanis, I.; Koutsoliotas, S.; Laird, E. M.; Link, J. M.; Liu, Y.; Liu, Y.; Louis, W. C.; Mahn, K. B. M.; Marsh, W.; Martin, P. S.; McGregor, G.; Metcalf, W.; Meyers, P. D.; Mills, F.; Mills, G. B.; Monroe, J.; Moore, C. D.; Nelson, R. H.; Nienaber, P.; Ouedraogo, S.; Patterson, R. B.; Perevalov, D.; Polly, C. C.; Prebys, E.; Raaf, J. L.; Ray, H.; Roe, B. P.; Russell, A. D.; Sandberg, V.; Schirato, R.; Schmitz, D.; Shaevitz, M. H.; Shoemaker, F. C.; Smith, D.; Sorel, M.; Spentzouris, P.; Stancu, I.; Stefanski, R. J.; Sung, M.; Tanaka, H. A.; Tayloe, R.; Tzanov, M.; van de Water, R.; Wascko, M. O.; White, D. H.; Wilking, M. J.; Yang, H. J.; Zeller, G. P.; Zimmerman, E. D.
2008-01-01
The observation of neutrino oscillations is clear evidence for physics beyond the standard model. To make precise measurements of this phenomenon, neutrino oscillation experiments, including MiniBooNE, require an accurate description of neutrino charged current quasielastic (CCQE) cross sections to predict signal samples. Using a high-statistics sample of νμ CCQE events, MiniBooNE finds that a simple Fermi gas model, with appropriate adjustments, accurately characterizes the CCQE events observed in a carbon-based detector. The extracted parameters include an effective axial mass, MAeff=1.23±0.20GeV, that describes the four-momentum dependence of the axial-vector form factor of the nucleon, and a Pauli-suppression parameter, κ=1.019±0.011. Such a modified Fermi gas model may also be used by future accelerator-based experiments measuring neutrino oscillations on nuclear targets.
Revealing physical interaction networks from statistics of collective dynamics
Nitzan, Mor; Casadiego, Jose; Timme, Marc
2017-01-01
Revealing physical interactions in complex systems from observed collective dynamics constitutes a fundamental inverse problem in science. Current reconstruction methods require access to a system’s model or dynamical data at a level of detail often not available. We exploit changes in invariant measures, in particular distributions of sampled states of the system in response to driving signals, and use compressed sensing to reveal physical interaction networks. Dynamical observations following driving suffice to infer physical connectivity even if they are temporally disordered, are acquired at large sampling intervals, and stem from different experiments. Testing various nonlinear dynamic processes emerging on artificial and real network topologies indicates high reconstruction quality for existence as well as type of interactions. These results advance our ability to reveal physical interaction networks in complex synthetic and natural systems. PMID:28246630
Pearce, J.M.; Talbot, S.L.; Petersen, M.R.; Rearick, J.R.
2005-01-01
Due to declines in the Alaska breeding population, the Steller's eider (Polysticta stelleri) was listed as threatened in North America in 1997. Periodic non-breeding in Russia and Alaska has hampered field-based assessments of behavioral patterns critical to recovery plans, such as levels of breeding site fidelity and movements among three regional populations: Atlantic-Russia, Pacific-Russia and Alaska. Therefore, we analyzed samples from across the species range with seven nuclear microsatellite DNA loci and cytochrome b mitochondrial (mt)DNA sequence data to infer levels of interchange among sampling areas and patterns of site fidelity. Results demonstrated low levels of population differentiation within Atlantic and Pacific nesting areas, with higher levels observed between these regions, but only for mtDNA. Bayesian analysis of microsatellite data from wintering and molting birds showed no signs of sub-population structure, even though band-recovery data suggests multiple breeding areas are present. We observed higher estimates of F-statistics for female mtDNA data versus male data, suggesting female-biased natal site fidelity. Summary statistics for mtDNA were consistent with models of historic population expansion. Lack of spatial structure in Steller's eiders may result largely from insufficient time since historic population expansions for behaviors, such as natal site fidelity, to isolate breeding areas genetically. However, other behaviors such as the periodic non-breeding observed in Steller's eiders may also play a more contemporary role in genetic homogeneity, especially for microsatellite loci.
A population study of hot Jupiter atmospheres
NASA Astrophysics Data System (ADS)
Tsiaras, A.; Waldmann, I. P.; Zingales, T.; Rocchetto, M.; Damiano, M.; Karpouzas, K.; Tinetti, G.; McKemmish, L. K.; Tennyson, J.; Yrchenko, S. N.
2017-09-01
In the past two decades, we have learnt that every star hosts more than one planet. While the hunt for new exoplanets is on-going, the current sample of more than 3500 confirmed planets reveals a wide spectrum of planetary characteristics. While small planets appear to be the most common, the big and gaseous planets play a key role in the process of planetary formation. We present here the analysis of 30 gaseous extra-solar planets, with temperatures between 600 and 2400 K and radii between 0.35 and 1.9 Jupiter radii. These planets were spectroscopically observed with the Wide Field Camera 3 on-board the Hubble Space Telescope, which is currently one of the most successful instruments for observing exoplanetary atmospheres. The quality of the HST/WFC3 spatially-scanned data combined with our specialised analysis tools, allows us to create the largest and most self-consistent sample of exoplanetary transmission spectra to date and study the collective behaviour of warm and hot gaseous planets rather than isolated case-studies. We define a new metric, the Atmospheric Detectability Index (ADI) to evaluate the statistical significance of an atmospheric detection and find statistically significant atmospheres around 16 planets. For most of the Jupiters in our sample we find the detectability of their atmospheres to be dependent on the planetary radius but not on the planetary mass. This indicates that planetary gravity is a secondary factor in the evolution of planetary atmospheres. We detect the presence of water vapour in all the statistically detectable atmospheres and we cannot rule out its presence in the atmospheres of the others. In addition, TiO and/or VO signatures are detected with 4σ confidence in WASP-76 b, and they are most likely present on WASP-121 b. We find no correlation between expected signal-to-noise and atmospheric detectability for most targets. This has important implications for future large-scale surveys.
Statistical Measures of Large-Scale Structure
NASA Astrophysics Data System (ADS)
Vogeley, Michael; Geller, Margaret; Huchra, John; Park, Changbom; Gott, J. Richard
1993-12-01
\\inv Mpc} To quantify clustering in the large-scale distribution of galaxies and to test theories for the formation of structure in the universe, we apply statistical measures to the CfA Redshift Survey. This survey is complete to m_{B(0)}=15.5 over two contiguous regions which cover one-quarter of the sky and include ~ 11,000 galaxies. The salient features of these data are voids with diameter 30-50\\hmpc and coherent dense structures with a scale ~ 100\\hmpc. Comparison with N-body simulations rules out the ``standard" CDM model (Omega =1, b=1.5, sigma_8 =1) at the 99% confidence level because this model has insufficient power on scales lambda >30\\hmpc. An unbiased open universe CDM model (Omega h =0.2) and a biased CDM model with non-zero cosmological constant (Omega h =0.24, lambda_0 =0.6) match the observed power spectrum. The amplitude of the power spectrum depends on the luminosity of galaxies in the sample; bright (L>L(*) ) galaxies are more strongly clustered than faint galaxies. The paucity of bright galaxies in low-density regions may explain this dependence. To measure the topology of large-scale structure, we compute the genus of isodensity surfaces of the smoothed density field. On scales in the ``non-linear" regime, <= 10\\hmpc, the high- and low-density regions are multiply-connected over a broad range of density threshold, as in a filamentary net. On smoothing scales >10\\hmpc, the topology is consistent with statistics of a Gaussian random field. Simulations of CDM models fail to produce the observed coherence of structure on non-linear scales (>95% confidence level). The underdensity probability (the frequency of regions with density contrast delta rho //lineρ=-0.8) depends strongly on the luminosity of galaxies; underdense regions are significantly more common (>2sigma ) in bright (L>L(*) ) galaxy samples than in samples which include fainter galaxies.
Prevalence of Dental Anomalies in Permanent Dentition of Brazilian Individuals with Down Syndrome
Cuoghi, Osmar Aparecido; Topolski, Francielle; Perciliano de Faria, Lorraine; Occhiena, Carla Machado; Ferreira, Nancy dos Santos Pinto; Ferlin, Camila Ribeiro; Rogério de Mendonça, Marcos
2016-01-01
Objective: The aim of this study was to evaluate the incidence of dental anomalies in the permanent dentition of individuals with Down Syndrome (DS) to increase the knowledge on the dental issues in this syndrome. Method: One hundred and five panoramic X-rays of patients with DS (61 males and 44 females), aged 7 to 42 years were used. The data were statistically analyzed using bivariate analyses test (p <0.05). Results: Dental anomalies were observed in 50.47% of the sample. More than one anomaly was observed in 9.52% of the individuals. The most frequent dental anomalies were hypodontia and microdontia (16.19%), followed by retained tooth (10.47%), taurodontism (9.52%), supernumerary teeth (5.71%), macrodontia (2.85%) and root dilaceration (0.95%). There was no statistically significant difference between genders for any of the anomalies. Conclusion: A high prevalence of dental anomalies was observed in individuals with DS. The results of the present study reinforce the importance of good dental care, offering a greater basis for professionals who provide dental service to these patients. PMID:27733874
Prevalence of Dental Anomalies in Permanent Dentition of Brazilian Individuals with Down Syndrome.
Cuoghi, Osmar Aparecido; Topolski, Francielle; Perciliano de Faria, Lorraine; Occhiena, Carla Machado; Ferreira, Nancy Dos Santos Pinto; Ferlin, Camila Ribeiro; Rogério de Mendonça, Marcos
2016-01-01
The aim of this study was to evaluate the incidence of dental anomalies in the permanent dentition of individuals with Down Syndrome (DS) to increase the knowledge on the dental issues in this syndrome. One hundred and five panoramic X-rays of patients with DS (61 males and 44 females), aged 7 to 42 years were used. The data were statistically analyzed using bivariate analyses test ( p <0.05). Dental anomalies were observed in 50.47% of the sample. More than one anomaly was observed in 9.52% of the individuals. The most frequent dental anomalies were hypodontia and microdontia (16.19%), followed by retained tooth (10.47%), taurodontism (9.52%), supernumerary teeth (5.71%), macrodontia (2.85%) and root dilaceration (0.95%). There was no statistically significant difference between genders for any of the anomalies. A high prevalence of dental anomalies was observed in individuals with DS. The results of the present study reinforce the importance of good dental care, offering a greater basis for professionals who provide dental service to these patients.
Fluctuating observation time ensembles in the thermodynamics of trajectories
NASA Astrophysics Data System (ADS)
Budini, Adrián A.; Turner, Robert M.; Garrahan, Juan P.
2014-03-01
The dynamics of stochastic systems, both classical and quantum, can be studied by analysing the statistical properties of dynamical trajectories. The properties of ensembles of such trajectories for long, but fixed, times are described by large-deviation (LD) rate functions. These LD functions play the role of dynamical free energies: they are cumulant generating functions for time-integrated observables, and their analytic structure encodes dynamical phase behaviour. This ‘thermodynamics of trajectories’ approach is to trajectories and dynamics what the equilibrium ensemble method of statistical mechanics is to configurations and statics. Here we show that, just like in the static case, there are a variety of alternative ensembles of trajectories, each defined by their global constraints, with that of trajectories of fixed total time being just one of these. We show how the LD functions that describe an ensemble of trajectories where some time-extensive quantity is constant (and large) but where total observation time fluctuates can be mapped to those of the fixed-time ensemble. We discuss how the correspondence between generalized ensembles can be exploited in path sampling schemes for generating rare dynamical trajectories.
NASA Astrophysics Data System (ADS)
Vigan, A.; Chauvin, G.; Bonavita, M.; Desidera, S.; Bonnefoy, M.; Mesa, D.; Beuzit, J.-L.; Augereau, J.-C.; Biller, B.; Boccaletti, A.; Brugaletta, E.; Buenzli, E.; Carson, J.; Covino, E.; Delorme, P.; Eggenberger, A.; Feldt, M.; Hagelberg, J.; Henning, T.; Lagrange, A.-M.; Lanzafame, A.; Ménard, F.; Messina, S.; Meyer, M.; Montagnier, G.; Mordasini, C.; Mouillet, D.; Moutou, C.; Mugnier, L.; Quanz, S. P.; Reggiani, M.; Ségransan, D.; Thalmann, C.; Waters, R.; Zurlo, A.
2014-01-01
Over the past decade, a growing number of deep imaging surveys have started to provide meaningful constraints on the population of extrasolar giant planets at large orbital separation. Primary targets for these surveys have been carefully selected based on their age, distance and spectral type, and often on their membership to young nearby associations where all stars share common kinematics, photometric and spectroscopic properties. The next step is a wider statistical analysis of the frequency and properties of low mass companions as a function of stellar mass and orbital separation. In late 2009, we initiated a coordinated European Large Program using angular differential imaging in the H band (1.66 μm) with NaCo at the VLT. Our aim is to provide a comprehensive and statistically significant study of the occurrence of extrasolar giant planets and brown dwarfs at large (5-500 AU) orbital separation around ~150 young, nearby stars, a large fraction of which have never been observed at very deep contrast. The survey has now been completed and we present the data analysis and detection limits for the observed sample, for which we reach the planetary-mass domain at separations of >~50 AU on average. We also present the results of the statistical analysis that has been performed over the 75 targets newly observed at high-contrast. We discuss the details of the statistical analysis and the physical constraints that our survey provides for the frequency and formation scenario of planetary mass companions at large separation.
Kalra, Pinky; Rao, Arathi; Suman, Ethel; Shenoy, Ramya; Suprabha, Baranya-Shrikrishna
2017-02-01
Endodontic instrumentation carries the risk of over extrusion of debris and bacteria. The technique used and the type of instrumentation influences this risk. The purpose of this study was to evaluate and compare the K-file, ProTaper hand and ProTaper rotary instrumentation systems for the amount of apically extruded debris, irrigant solution and intracanal bacteria. Experimental single blinded randomized type of in vitro study with sample of 30 single rooted teeth. Endodontic access cavities were prepared and the root canals were filled with the suspension of E. faecalis . Myers and Montogomery Model was used to collect apically extruded debris and irrigant. Canals were prepared using K files, Hand protapers and Protaper rotary files. Non Parametric test like Kruskal-Wallis and Mann-Whitney U test were applied to determine the significant differences among the group. Tests revealed statistically significant difference between the amount of debris and number of bacteria extruded by the ProTaper hand and the K-files. No statistically significant difference was observed between the amounts of irrigant extruded by the ProTaper hand and the K-file system. Statistically significant differences were observed between the amounts of bacteria and irrigant extruded by the ProTaper rotary and the Protaper hand. No statistically significant difference was observed between the amounts of debris extruded by the ProTaper hand and the K-file system. Amount of apical extrusion of irrigant solution, bacteria and debris are significantly greater with K File instruments and least with Protaper rotary instruments. Key words: Protaper, rotary, periapical extrusion.
Statistics of the geomagnetic secular variation for the past 5Ma
NASA Technical Reports Server (NTRS)
Constable, C. G.; Parker, R. L.
1986-01-01
A new statistical model is proposed for the geomagnetic secular variation over the past 5Ma. Unlike previous models, the model makes use of statistical characteristics of the present day geomagnetic field. The spatial power spectrum of the non-dipole field is consistent with a white source near the core-mantle boundary with Gaussian distribution. After a suitable scaling, the spherical harmonic coefficients may be regarded as statistical samples from a single giant Gaussian process; this is the model of the non-dipole field. The model can be combined with an arbitrary statistical description of the dipole and probability density functions and cumulative distribution functions can be computed for declination and inclination that would be observed at any site on Earth's surface. Global paleomagnetic data spanning the past 5Ma are used to constrain the statistics of the dipole part of the field. A simple model is found to be consistent with the available data. An advantage of specifying the model in terms of the spherical harmonic coefficients is that it is a complete statistical description of the geomagnetic field, enabling us to test specific properties for a general description. Both intensity and directional data distributions may be tested to see if they satisfy the expected model distributions.
Statistics of the geomagnetic secular variation for the past 5 m.y
NASA Technical Reports Server (NTRS)
Constable, C. G.; Parker, R. L.
1988-01-01
A new statistical model is proposed for the geomagnetic secular variation over the past 5Ma. Unlike previous models, the model makes use of statistical characteristics of the present day geomagnetic field. The spatial power spectrum of the non-dipole field is consistent with a white source near the core-mantle boundary with Gaussian distribution. After a suitable scaling, the spherical harmonic coefficients may be regarded as statistical samples from a single giant Gaussian process; this is the model of the non-dipole field. The model can be combined with an arbitrary statistical description of the dipole and probability density functions and cumulative distribution functions can be computed for declination and inclination that would be observed at any site on Earth's surface. Global paleomagnetic data spanning the past 5Ma are used to constrain the statistics of the dipole part of the field. A simple model is found to be consistent with the available data. An advantage of specifying the model in terms of the spherical harmonic coefficients is that it is a complete statistical description of the geomagnetic field, enabling us to test specific properties for a general description. Both intensity and directional data distributions may be tested to see if they satisfy the expected model distributions.
NASA Astrophysics Data System (ADS)
ten Veldhuis, Marie-Claire; Schleiss, Marc
2017-04-01
In this study, we introduced an alternative approach for analysis of hydrological flow time series, using an adaptive sampling framework based on inter-amount times (IATs). The main difference with conventional flow time series is the rate at which low and high flows are sampled: the unit of analysis for IATs is a fixed flow amount, instead of a fixed time window. We analysed statistical distributions of flows and IATs across a wide range of sampling scales to investigate sensitivity of statistical properties such as quantiles, variance, skewness, scaling parameters and flashiness indicators to the sampling scale. We did this based on streamflow time series for 17 (semi)urbanised basins in North Carolina, US, ranging from 13 km2 to 238 km2 in size. Results showed that adaptive sampling of flow time series based on inter-amounts leads to a more balanced representation of low flow and peak flow values in the statistical distribution. While conventional sampling gives a lot of weight to low flows, as these are most ubiquitous in flow time series, IAT sampling gives relatively more weight to high flow values, when given flow amounts are accumulated in shorter time. As a consequence, IAT sampling gives more information about the tail of the distribution associated with high flows, while conventional sampling gives relatively more information about low flow periods. We will present results of statistical analyses across a range of subdaily to seasonal scales and will highlight some interesting insights that can be derived from IAT statistics with respect to basin flashiness and impact urbanisation on hydrological response.
Is there gender bias in nursing research?
Polit, Denise F; Beck, Cheryl Tatano
2008-10-01
Using data from a consecutive sample of 259 studies published in four leading nursing research journals in 2005-2006, we examined whether nurse researchers favor females as study participants. On average, 75.3% of study participants were female, and 38% of studies had all-female samples. The bias favoring female participants was statistically significant and persistent. The bias was observed regardless of funding source, methodological features, and other participant and researcher characteristics, with one exception: studies that had male investigators had more sex-balanced samples. When designing studies, nurse researchers need to pay close attention to who will benefit from their research and to whether they are leaving out a specific group about which there is a gap in knowledge. (c) 2008 Wiley Periodicals, Inc.
Potential, velocity, and density fields from sparse and noisy redshift-distance samples - Method
NASA Technical Reports Server (NTRS)
Dekel, Avishai; Bertschinger, Edmund; Faber, Sandra M.
1990-01-01
A method for recovering the three-dimensional potential, velocity, and density fields from large-scale redshift-distance samples is described. Galaxies are taken as tracers of the velocity field, not of the mass. The density field and the initial conditions are calculated using an iterative procedure that applies the no-vorticity assumption at an initial time and uses the Zel'dovich approximation to relate initial and final positions of particles on a grid. The method is tested using a cosmological N-body simulation 'observed' at the positions of real galaxies in a redshift-distance sample, taking into account their distance measurement errors. Malmquist bias and other systematic and statistical errors are extensively explored using both analytical techniques and Monte Carlo simulations.
SPA- STATISTICAL PACKAGE FOR TIME AND FREQUENCY DOMAIN ANALYSIS
NASA Technical Reports Server (NTRS)
Brownlow, J. D.
1994-01-01
The need for statistical analysis often arises when data is in the form of a time series. This type of data is usually a collection of numerical observations made at specified time intervals. Two kinds of analysis may be performed on the data. First, the time series may be treated as a set of independent observations using a time domain analysis to derive the usual statistical properties including the mean, variance, and distribution form. Secondly, the order and time intervals of the observations may be used in a frequency domain analysis to examine the time series for periodicities. In almost all practical applications, the collected data is actually a mixture of the desired signal and a noise signal which is collected over a finite time period with a finite precision. Therefore, any statistical calculations and analyses are actually estimates. The Spectrum Analysis (SPA) program was developed to perform a wide range of statistical estimation functions. SPA can provide the data analyst with a rigorous tool for performing time and frequency domain studies. In a time domain statistical analysis the SPA program will compute the mean variance, standard deviation, mean square, and root mean square. It also lists the data maximum, data minimum, and the number of observations included in the sample. In addition, a histogram of the time domain data is generated, a normal curve is fit to the histogram, and a goodness-of-fit test is performed. These time domain calculations may be performed on both raw and filtered data. For a frequency domain statistical analysis the SPA program computes the power spectrum, cross spectrum, coherence, phase angle, amplitude ratio, and transfer function. The estimates of the frequency domain parameters may be smoothed with the use of Hann-Tukey, Hamming, Barlett, or moving average windows. Various digital filters are available to isolate data frequency components. Frequency components with periods longer than the data collection interval are removed by least-squares detrending. As many as ten channels of data may be analyzed at one time. Both tabular and plotted output may be generated by the SPA program. This program is written in FORTRAN IV and has been implemented on a CDC 6000 series computer with a central memory requirement of approximately 142K (octal) of 60 bit words. This core requirement can be reduced by segmentation of the program. The SPA program was developed in 1978.
Nash, J. Thomas; Frishman, David
1983-01-01
Analytical results for 61 elements in 370 samples from the Ranger Mine area are reported. Most of the rocks come from drill core in the Ranger No. 1 and Ranger No. 3 deposits, but 20 samples are from unmineralized drill core more than 1 km from ore. Statistical tests show that the elements Mg, Fe, F, Be, Co, Li, Ni, Pb, Sc, Th, Ti, V, CI, As, Br, Au, Ce, Dy, La Sc, Eu, Tb, Yb, and Tb have positive association with uranium, and Si, Ca, Na, K, Sr, Ba, Ce, and Cs have negative association. For most lithologic subsets Mg, Fe, Li, Cr, Ni, Pb, V, Y, Sm, Sc, Eu, and Yb are significantly enriched in ore-bearing rocks, whereas Ca, Na, K, Sr, Ba, Mn, Ce, and Cs are significantly depleted. These results are consistent with petrographic observations on altered rocks. Lithogeochemistry can aid exploration, but for these rocks requires methods that are expensive and not amenable to routine use.
Williams, Donald R; Carlsson, Rickard; Bürkner, Paul-Christian
2017-10-01
Developmental studies of hormones and behavior often include littermates-rodent siblings that share early-life experiences and genes. Due to between-litter variation (i.e., litter effects), the statistical assumption of independent observations is untenable. In two literatures-natural variation in maternal care and prenatal stress-entire litters are categorized based on maternal behavior or experimental condition. Here, we (1) review both literatures; (2) simulate false positive rates for commonly used statistical methods in each literature; and (3) characterize small sample performance of multilevel models (MLM) and generalized estimating equations (GEE). We found that the assumption of independence was routinely violated (>85%), false positives (α=0.05) exceeded nominal levels (up to 0.70), and power (1-β) rarely surpassed 0.80 (even for optimistic sample and effect sizes). Additionally, we show that MLMs and GEEs have adequate performance for common research designs. We discuss implications for the extant literature, the field of behavioral neuroendocrinology, and provide recommendations. Copyright © 2017 Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
Usry, J. W.
1983-01-01
Wind shear statistics were calculated for a simulated set of wind profiles based on a proposed standard wind field data base. Wind shears were grouped in altitude in altitude bands of 100 ft between 100 and 1400 ft and in wind shear increments of 0.025 knot/ft. Frequency distributions, means, and standard deviations for each altitude band were derived for the total sample were derived for both sets. It was found that frequency distributions in each altitude band for the simulated data set were more dispersed below 800 ft and less dispersed above 900 ft than those for the measured data set. Total sample frequency of occurrence for the two data sets was about equal for wind shear values between +0.075 knot/ft, but the simulated data set had significantly larger values for all wind shears outside these boundaries. It is shown that normal distribution in both data sets neither data set was normally distributed; similar results are observed from the cumulative frequency distributions.