Sample records for statistically well-defined sample

  1. Just the right age: well-clustered exposure ages from a global glacial 10Be compilation

    NASA Astrophysics Data System (ADS)

    Heyman, Jakob; Margold, Martin

    2017-04-01

    Cosmogenic exposure dating has been used extensively for defining glacial chronologies, both in ice sheet and alpine settings, and the global set of published ages today reaches well beyond 10,000 samples. Over the last few years, a number of important developments have improved the measurements (with well-defined AMS standards) and exposure age calculations (with updated data and methods for calculating production rates), in the best case enabling high precision dating of past glacial events. A remaining problem, however, is the fact that a large portion of all dated samples have been affected by prior and/or incomplete exposure, yielding erroneous exposure ages under the standard assumptions. One way to address this issue is to only use exposure ages that can be confidently considered as unaffected by prior/incomplete exposure, such as groups of samples with statistically identical ages. Here we use objective statistical criteria to identify groups of well-clustered exposure ages from the global glacial "expage" 10Be compilation. Out of ˜1700 groups with at least 3 individual samples ˜30% are well-clustered, increasing to ˜45% if allowing outlier rejection of a maximum of 1/3 of the samples (still requiring a minimum of 3 well-clustered ages). The dataset of well-clustered ages is heavily dominated by ages <30 ka, showing that well-defined cosmogenic chronologies primarily exist for the last glaciation. We observe a large-scale global synchronicity in the timing of the last deglaciation from ˜20 to 10 ka. There is also a general correlation between the timing of deglaciation and latitude (or size of the individual ice mass), with earlier deglaciation in lower latitudes and later deglaciation towards the poles. Grouping the data into regions and comparing with available paleoclimate data we can start to untangle regional differences in the last deglaciation and the climate events controlling the ice mass loss. The extensive dataset and the statistical analysis enables an unprecedented global view on the last deglaciation.

  2. New Approaches to Robust Confidence Intervals for Location: A Simulation Study.

    DTIC Science & Technology

    1984-06-01

    obtain a denominator for the test statistic. Those statistics based on location estimates derived from Hampel’s redescending influence function or v...defined an influence function for a test in terms of the behavior of its P-values when the data are sampled from a model distribution modified by point...proposal could be used for interval estimation as well as hypothesis testing, the extension is immediate. Once an influence function has been defined

  3. Survey of rural, private wells. Statistical design

    USGS Publications Warehouse

    Mehnert, Edward; Schock, Susan C.; ,

    1991-01-01

    Half of Illinois' 38 million acres were planted in corn and soybeans in 1988. On the 19 million acres planted in corn and soybeans, approximately 1 million tons of nitrogen fertilizer and 50 million pounds of pesticides were applied. Because groundwater is the water supply for over 90 percent of rural Illinois, the occurrence of agricultural chemicals in groundwater in Illinois is of interest to the agricultural community, the public, and regulatory agencies. The occurrence of agricultural chemicals in groundwater is well documented. However, the extent of this contamination still needs to be defined. This can be done by randomly sampling wells across a geographic area. Key elements of a random, water-well sampling program for regional groundwater quality include the overall statistical design of the program, definition of the sample population, selection of wells to be sampled, and analysis of survey results. These elements must be consistent with the purpose for conducting the program; otherwise, the program will not provide the desired information. The need to carefully design and conduct a sampling program becomes readily apparent when one considers the high cost of collecting and analyzing a sample. For a random sampling program conducted in Illinois, the key elements, as well as the limitations imposed by available information, are described.

  4. Concentrations of tritium and strontium-90 in water from selected wells at the Idaho National Engineering Laboratory after purging one, two, and three borehole volumes

    USGS Publications Warehouse

    Bartholomay, R.C.

    1993-01-01

    Water from 11 wells completed in the Snake River Plain aquifer at the Idaho National Engineering Laboratory was sampled as part of the U.S. Geological Survey's quality assurance program to determine the effect of purging different borehole volumes on tritium and strontium-90 concentrations. Wells were selected for sampling on the basis of the length of time it took to purge a borehole volume of water. Samples were collected after purging one, two, and three borehole volumes. The U.S. Department of Energy's Radiological and Environmental Sciences Laboratory provided analytical services. Statistics were used to determine the reproducibility of analytical results. The comparison between tritium and strontium-90 concentrations after purging one and three borehole volumes and two and three borehole volumes showed that all but two sample pairs with defined numbers were in statistical agreement. Results indicate that concentrations of tritium and strontium-90 are not affected measurably by the number of borehole volumes purged.

  5. Understanding the Sampling Distribution and the Central Limit Theorem.

    ERIC Educational Resources Information Center

    Lewis, Charla P.

    The sampling distribution is a common source of misuse and misunderstanding in the study of statistics. The sampling distribution, underlying distribution, and the Central Limit Theorem are all interconnected in defining and explaining the proper use of the sampling distribution of various statistics. The sampling distribution of a statistic is…

  6. Toward Improving Research in Social Studies Education. SSEC Monograph Series.

    ERIC Educational Resources Information Center

    Fraenkel, Jack R.; Wallen, Norman E.

    Social studies research has been criticized for sampling bias, inappropriate methodologies, incorrect or inappropriate use of statistics, weak or ill-defined treatments, and lack of replication and/or longitudinal follow-up. In an effort to ascertain whether past criticisms were true of current research as well, a review was conducted of 118…

  7. Concentrations of tritium and strontium-90 in water from selected wells at the Idaho National Engineering Laboratory after purging one, two, and three borehole volumes

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bartholomay, R.C.

    1993-12-31

    Water from 11 wells completed in the Snake River Plain aquifer at the Idaho National Engineering Laboratory was sampled as Part of the US. Geological Survey`s quality assurance program to determine the effect of Purging different borehole volumes on tritium and strontium-90 concentrations. Wells were selected for sampling on the basis of the length of time it took to purge a borehole volume of water. Samples were collected after purging one, two, and three borehole volumes. The US Department of Energy`s Radiological and Environmental Sciences Laboratory provided analytical services. Statistics were used to determine the reproducibility of analytical results. Themore » comparison between tritium and strontium-90 concentrations after purging one and three borehole volumes and two and three borehole volumes showed that all but two sample pairs with defined numbers were in statistical agreement. Results indicate that concentrations of tritium and strontium-90 are not affected measurably by the number of borehole volumes purged.« less

  8. Statistical Inference for Data Adaptive Target Parameters.

    PubMed

    Hubbard, Alan E; Kherad-Pajouh, Sara; van der Laan, Mark J

    2016-05-01

    Consider one observes n i.i.d. copies of a random variable with a probability distribution that is known to be an element of a particular statistical model. In order to define our statistical target we partition the sample in V equal size sub-samples, and use this partitioning to define V splits in an estimation sample (one of the V subsamples) and corresponding complementary parameter-generating sample. For each of the V parameter-generating samples, we apply an algorithm that maps the sample to a statistical target parameter. We define our sample-split data adaptive statistical target parameter as the average of these V-sample specific target parameters. We present an estimator (and corresponding central limit theorem) of this type of data adaptive target parameter. This general methodology for generating data adaptive target parameters is demonstrated with a number of practical examples that highlight new opportunities for statistical learning from data. This new framework provides a rigorous statistical methodology for both exploratory and confirmatory analysis within the same data. Given that more research is becoming "data-driven", the theory developed within this paper provides a new impetus for a greater involvement of statistical inference into problems that are being increasingly addressed by clever, yet ad hoc pattern finding methods. To suggest such potential, and to verify the predictions of the theory, extensive simulation studies, along with a data analysis based on adaptively determined intervention rules are shown and give insight into how to structure such an approach. The results show that the data adaptive target parameter approach provides a general framework and resulting methodology for data-driven science.

  9. Quantum probabilistic logic programming

    NASA Astrophysics Data System (ADS)

    Balu, Radhakrishnan

    2015-05-01

    We describe a quantum mechanics based logic programming language that supports Horn clauses, random variables, and covariance matrices to express and solve problems in probabilistic logic. The Horn clauses of the language wrap random variables, including infinite valued, to express probability distributions and statistical correlations, a powerful feature to capture relationship between distributions that are not independent. The expressive power of the language is based on a mechanism to implement statistical ensembles and to solve the underlying SAT instances using quantum mechanical machinery. We exploit the fact that classical random variables have quantum decompositions to build the Horn clauses. We establish the semantics of the language in a rigorous fashion by considering an existing probabilistic logic language called PRISM with classical probability measures defined on the Herbrand base and extending it to the quantum context. In the classical case H-interpretations form the sample space and probability measures defined on them lead to consistent definition of probabilities for well formed formulae. In the quantum counterpart, we define probability amplitudes on Hinterpretations facilitating the model generations and verifications via quantum mechanical superpositions and entanglements. We cast the well formed formulae of the language as quantum mechanical observables thus providing an elegant interpretation for their probabilities. We discuss several examples to combine statistical ensembles and predicates of first order logic to reason with situations involving uncertainty.

  10. Relationships among Classical Test Theory and Item Response Theory Frameworks via Factor Analytic Models

    ERIC Educational Resources Information Center

    Kohli, Nidhi; Koran, Jennifer; Henn, Lisa

    2015-01-01

    There are well-defined theoretical differences between the classical test theory (CTT) and item response theory (IRT) frameworks. It is understood that in the CTT framework, person and item statistics are test- and sample-dependent. This is not the perception with IRT. For this reason, the IRT framework is considered to be theoretically superior…

  11. 16(th) IHIW: analysis of HLA population data, with updated results for 1996 to 2012 workshop data (AHPD project report).

    PubMed

    Riccio, M E; Buhler, S; Nunes, J M; Vangenot, C; Cuénod, M; Currat, M; Di, D; Andreani, M; Boldyreva, M; Chambers, G; Chernova, M; Chiaroni, J; Darke, C; Di Cristofaro, J; Dubois, V; Dunn, P; Edinur, H A; Elamin, N; Eliaou, J-F; Grubic, Z; Jaatinen, T; Kanga, U; Kervaire, B; Kolesar, L; Kunachiwa, W; Lokki, M L; Mehra, N; Nicoloso, G; Paakkanen, R; Voniatis, D Papaioannou; Papasteriades, C; Poli, F; Richard, L; Romón Alonso, I; Slavčev, A; Sulcebe, G; Suslova, T; Testi, M; Tiercy, J-M; Varnavidou, A; Vidan-Jeras, B; Wennerström, A; Sanchez-Mazas, A

    2013-02-01

    We present here the results of the Analysis of HLA Population Data (AHPD) project of the 16th International HLA and Immunogenetics Workshop (16IHIW) held in Liverpool in May-June 2012. Thanks to the collaboration of 25 laboratories from 18 different countries, HLA genotypic data for 59 new population samples (either well-defined populations or donor registry samples) were gathered and 55 were analysed statistically following HLA-NET recommendations. The new data included, among others, large sets of well-defined populations from north-east Europe and West Asia, as well as many donor registry data from European countries. The Gene[rate] computer tools were combined to create a Gene[rate] computer pipeline to automatically (i) estimate allele frequencies by an expectation-maximization algorithm accommodating ambiguities, (ii) estimate heterozygosity, (iii) test for Hardy-Weinberg equilibrium (HWE), (iv) test for selective neutrality, (v) generate frequency graphs and summary statistics for each sample at each locus and (vi) plot multidimensional scaling (MDS) analyses comparing the new samples with previous IHIW data. Intrapopulation analyses show that HWE is rarely rejected, while neutrality tests often indicate a significant excess of heterozygotes compared with neutral expectations. The comparison of the 16IHIW AHPD data with data collected during previous workshops (12th-15th) shows that geography is an excellent predictor of HLA genetic differentiations for HLA-A, -B and -DRB1 loci but not for HLA-DQ, whose patterns are probably more influenced by natural selection. In Europe, HLA genetic variation clearly follows a north to south-east axis despite a low level of differentiation between European, North African and West Asian populations. Pacific populations are genetically close to Austronesian-speaking South-East Asian and Taiwanese populations, in agreement with current theories on the peopling of Oceania. Thanks to this project, HLA genetic variation is more clearly defined worldwide and better interpreted in relation to human peopling history and HLA molecular evolution. © 2012 Blackwell Publishing Ltd.

  12. Design of partially supervised classifiers for multispectral image data

    NASA Technical Reports Server (NTRS)

    Jeon, Byeungwoo; Landgrebe, David

    1993-01-01

    A partially supervised classification problem is addressed, especially when the class definition and corresponding training samples are provided a priori only for just one particular class. In practical applications of pattern classification techniques, a frequently observed characteristic is the heavy, often nearly impossible requirements on representative prior statistical class characteristics of all classes in a given data set. Considering the effort in both time and man-power required to have a well-defined, exhaustive list of classes with a corresponding representative set of training samples, this 'partially' supervised capability would be very desirable, assuming adequate classifier performance can be obtained. Two different classification algorithms are developed to achieve simplicity in classifier design by reducing the requirement of prior statistical information without sacrificing significant classifying capability. The first one is based on optimal significance testing, where the optimal acceptance probability is estimated directly from the data set. In the second approach, the partially supervised classification is considered as a problem of unsupervised clustering with initially one known cluster or class. A weighted unsupervised clustering procedure is developed to automatically define other classes and estimate their class statistics. The operational simplicity thus realized should make these partially supervised classification schemes very viable tools in pattern classification.

  13. Non-Immunogenic Structurally and Biologically Intact Tissue Matrix Grafts for the Immediate Repair of Ballistic-Induced Vascular and Nerve Tissue Injury in Combat Casualty Care

    DTIC Science & Technology

    2005-07-01

    as an access graft is addressed using statistical methods below. Graft consistency can be defined statistically as the variance associated with the...addressed using statistical methods below. Graft consistency can be defined statistically as the variance associated with the sample of grafts tested in...measured using a refractometer (Brix % method). The equilibration data are shown in Graph 1. The results suggest the following equilibration scheme: 40% v/v

  14. Wide binaries in the direction of Andromeda

    NASA Technical Reports Server (NTRS)

    Bahcall, J. N.; Ratnatunga, K. U.; Jones, B. F.

    1986-01-01

    A statistically well-defined sample of candidate binary stars with separations that are expected to be mostly in the range 0.01-0.1 pc is presented. The 36 candidate pairs are all brighter than apparent visual magnitude 12; about half of the projected pairs are expected to be physically associated. After the candidates are studied spectroscopically and photometrically to establish which pairs are real binaries and to measure their physical characteristics, the sample can be used to help determine the dependence of number density on semimajor axis for wide binaries, a function that is of considerable theoretical interest.

  15. Counting at low concentrations: the statistical challenges of verifying ballast water discharge standards

    USGS Publications Warehouse

    Frazier, Melanie; Miller, A. Whitman; Lee, Henry; Reusser, Deborah A.

    2013-01-01

    Discharge from the ballast tanks of ships is one of the primary vectors of nonindigenous species in marine environments. To mitigate this environmental and economic threat, international, national, and state entities are establishing regulations to limit the concentration of living organisms that may be discharged from the ballast tanks of ships. The proposed discharge standards have ranged from zero detectable organisms to 3. If standard sampling methods are used, verifying whether ballast discharge complies with these stringent standards will be challenging due to the inherent stochasticity of sampling. Furthermore, at low concentrations, very large volumes of water must be sampled to find enough organisms to accurately estimate concentration. Despite these challenges, adequate sampling protocols comprise a critical aspect of establishing standards because they help define the actual risk level associated with a standard. A standard that appears very stringent may be effectively lax if it is paired with an inadequate sampling protocol. We describe some of the statistical issues associated with sampling at low concentrations to help regulators understand the uncertainties of sampling as well as to inform the development of sampling protocols that ensure discharge standards are adequately implemented.

  16. Characterizations of linear sufficient statistics

    NASA Technical Reports Server (NTRS)

    Peters, B. C., Jr.; Reoner, R.; Decell, H. P., Jr.

    1977-01-01

    A surjective bounded linear operator T from a Banach space X to a Banach space Y must be a sufficient statistic for a dominated family of probability measures defined on the Borel sets of X. These results were applied, so that they characterize linear sufficient statistics for families of the exponential type, including as special cases the Wishart and multivariate normal distributions. The latter result was used to establish precisely which procedures for sampling from a normal population had the property that the sample mean was a sufficient statistic.

  17. Effects of nutrient management on nitrate levels in ground water near Ephrata Pennsylvania

    USGS Publications Warehouse

    Hall, David W.

    1992-01-01

    Effects of the implementation of nutrient management practices on ground-water quality were studied at a 55-acre farm in Lancaster County, Pennsylvania, from 1985-90. After nutrient management practices were implemented at the site in October 1986, statistically significant decreases (Wilcoxon Mann-Whitney test) in median nitrate concentrations in ground-water samples occurred at four of the five wells monitored. The largest decreases in nitrate concentration occurred in samples collected at the wells that had the largest nitrate concentrations prior to nutrient management. The decreases in median nitrate concentrations in ground-water samples ranged from 8 to 32 percent of the median concentrations prior to nutrient management and corresponded to nitrogen application decreases of 39 to 67 percent in contributing areas that were defined upgradient of these wells. Changes in nitrogen applications to the contributing areas of five water wells were correlated (Spearman rank-sum test) with nitrate concentrations of the well water. Changes in ground-water nitrate concentrations lagged behind the changes in applied-nitrogen fertilizers (primarily manure) by approximately 4 to 19 months.

  18. Why am I not disabled? Making state subjects, making statistics in post--Mao China.

    PubMed

    Kohrman, Matthew

    2003-03-01

    In this article I examine how and why disability was defined and statistically quantified by China's party-state in the late 1980s. I describe the unfolding of a particular epidemiological undertaking--China's 1987 National Sample Survey of Disabled Persons--as well as the ways the survey was an extension of what Ian Hacking has called modernity's "avalanche of numbers." I argue that, to a large degree, what fueled and shaped the 1987 survey's codification and quantification of disability was how Chinese officials were incited to shape their own identities as they negotiated an array of social, political, and ethical forces, which were at once national and transnational in orientation.

  19. Orientation of Hittite Monuments

    NASA Astrophysics Data System (ADS)

    González-García, A. César; Belmonte, Juan Antonio

    The possible astronomical or topographical orientations of the Hittite monuments of the Bronze Age has remained unexplored until recently. This would provide an important insight into how temporality was imprinted by this culture in sacred spaces and in the landscape. The authors' analysis of a statistically significant sample of Hittite temples - and a few monumental gates - has demonstrated that ancient Hittite monuments were not randomly orientated as previously thought. On the contrary, there were well-defined patterns of orientation that can be interpreted within the context of Hittite culture and religion.

  20. Groundwater-quality data for the Sierra Nevada study unit, 2008: Results from the California GAMA program

    USGS Publications Warehouse

    Shelton, Jennifer L.; Fram, Miranda S.; Munday, Cathy M.; Belitz, Kenneth

    2010-01-01

    Groundwater quality in the approximately 25,500-square-mile Sierra Nevada study unit was investigated in June through October 2008, as part of the Priority Basin Project of the Groundwater Ambient Monitoring and Assessment (GAMA) Program. The GAMA Priority Basin Project is being conducted by the U.S. Geological Survey (USGS) in cooperation with the California State Water Resources Control Board (SWRCB). The Sierra Nevada study was designed to provide statistically robust assessments of untreated groundwater quality within the primary aquifer systems in the study unit, and to facilitate statistically consistent comparisons of groundwater quality throughout California. The primary aquifer systems (hereinafter, primary aquifers) are defined by the depth of the screened or open intervals of the wells listed in the California Department of Public Health (CDPH) database of wells used for public and community drinking-water supplies. The quality of groundwater in shallower or deeper water-bearing zones may differ from that in the primary aquifers; shallow groundwater may be more vulnerable to contamination from the surface. In the Sierra Nevada study unit, groundwater samples were collected from 84 wells (and springs) in Lassen, Plumas, Butte, Sierra, Yuba, Nevada, Placer, El Dorado, Amador, Alpine, Calaveras, Tuolumne, Madera, Mariposa, Fresno, Inyo, Tulare, and Kern Counties. The wells were selected on two overlapping networks by using a spatially-distributed, randomized, grid-based approach. The primary grid-well network consisted of 30 wells, one well per grid cell in the study unit, and was designed to provide statistical representation of groundwater quality throughout the entire study unit. The lithologic grid-well network is a secondary grid that consisted of the wells in the primary grid-well network plus 53 additional wells and was designed to provide statistical representation of groundwater quality in each of the four major lithologic units in the Sierra Nevada study unit: granitic, metamorphic, sedimentary, and volcanic rocks. One natural spring that is not used for drinking water was sampled for comparison with a nearby primary grid well in the same cell. Groundwater samples were analyzed for organic constituents (volatile organic compounds [VOC], pesticides and pesticide degradates, and pharmaceutical compounds), constituents of special interest (N-nitrosodimethylamine [NDMA] and perchlorate), naturally occurring inorganic constituents (nutrients, major ions, total dissolved solids, and trace elements), and radioactive constituents (radium isotopes, radon-222, gross alpha and gross beta particle activities, and uranium isotopes). Naturally occurring isotopes and geochemical tracers (stable isotopes of hydrogen and oxygen in water, stable isotopes of carbon, carbon-14, strontium isotopes, and tritium), and dissolved noble gases also were measured to help identify the sources and ages of the sampled groundwater. Three types of quality-control samples (blanks, replicates, and samples for matrix spikes) each were collected at approximately 10 percent of the wells sampled for each analysis, and the results for these samples were used to evaluate the quality of the data for the groundwater samples. Field blanks rarely contained detectable concentrations of any constituent, suggesting that contamination from sample collection, handling, and analytical procedures was not a significant source of bias in the data for the groundwater samples. Differences between replicate samples were within acceptable ranges, with few exceptions. Matrix-spike recoveries were within acceptable ranges for most compounds. This study did not attempt to evaluate the quality of water delivered to consumers; after withdrawal from the ground, groundwater typically is treated, disinfected, or blended with other waters to maintain water quality. Regulatory benchmarks apply to finished drinking water that is served to the consumer, not to untre

  1. Defining Multiple Characteristic Raman Bands of α-Amino Acids as Biomarkers for Planetary Missions Using a Statistical Method

    NASA Astrophysics Data System (ADS)

    Rolfe, S. M.; Patel, M. R.; Gilmour, I.; Olsson-Francis, K.; Ringrose, T. J.

    2016-06-01

    Biomarker molecules, such as amino acids, are key to discovering whether life exists elsewhere in the Solar System. Raman spectroscopy, a technique capable of detecting biomarkers, will be on board future planetary missions including the ExoMars rover. Generally, the position of the strongest band in the spectra of amino acids is reported as the identifying band. However, for an unknown sample, it is desirable to define multiple characteristic bands for molecules to avoid any ambiguous identification. To date, there has been no definition of multiple characteristic bands for amino acids of interest to astrobiology. This study examined l-alanine, l-aspartic acid, l-cysteine, l-glutamine and glycine and defined several Raman bands per molecule for reference as characteristic identifiers. Per amino acid, 240 spectra were recorded and compared using established statistical tests including ANOVA. The number of characteristic bands defined were 10, 12, 12, 14 and 19 for l-alanine (strongest intensity band: 832 cm-1), l-aspartic acid (938 cm-1), l-cysteine (679 cm-1), l-glutamine (1090 cm-1) and glycine (875 cm-1), respectively. The intensity of bands differed by up to six times when several points on the crystal sample were rotated through 360 °; to reduce this effect when defining characteristic bands for other molecules, we find that spectra should be recorded at a statistically significant number of points per sample to remove the effect of sample rotation. It is crucial that sets of characteristic Raman bands are defined for biomarkers that are targets for future planetary missions to ensure a positive identification can be made.

  2. Defining Multiple Characteristic Raman Bands of α-Amino Acids as Biomarkers for Planetary Missions Using a Statistical Method.

    PubMed

    Rolfe, S M; Patel, M R; Gilmour, I; Olsson-Francis, K; Ringrose, T J

    2016-06-01

    Biomarker molecules, such as amino acids, are key to discovering whether life exists elsewhere in the Solar System. Raman spectroscopy, a technique capable of detecting biomarkers, will be on board future planetary missions including the ExoMars rover. Generally, the position of the strongest band in the spectra of amino acids is reported as the identifying band. However, for an unknown sample, it is desirable to define multiple characteristic bands for molecules to avoid any ambiguous identification. To date, there has been no definition of multiple characteristic bands for amino acids of interest to astrobiology. This study examined L-alanine, L-aspartic acid, L-cysteine, L-glutamine and glycine and defined several Raman bands per molecule for reference as characteristic identifiers. Per amino acid, 240 spectra were recorded and compared using established statistical tests including ANOVA. The number of characteristic bands defined were 10, 12, 12, 14 and 19 for L-alanine (strongest intensity band: 832 cm(-1)), L-aspartic acid (938 cm(-1)), L-cysteine (679 cm(-1)), L-glutamine (1090 cm(-1)) and glycine (875 cm(-1)), respectively. The intensity of bands differed by up to six times when several points on the crystal sample were rotated through 360 °; to reduce this effect when defining characteristic bands for other molecules, we find that spectra should be recorded at a statistically significant number of points per sample to remove the effect of sample rotation. It is crucial that sets of characteristic Raman bands are defined for biomarkers that are targets for future planetary missions to ensure a positive identification can be made.

  3. Statistical analyses to support guidelines for marine avian sampling. Final report

    USGS Publications Warehouse

    Kinlan, Brian P.; Zipkin, Elise; O'Connell, Allan F.; Caldow, Chris

    2012-01-01

    Interest in development of offshore renewable energy facilities has led to a need for high-quality, statistically robust information on marine wildlife distributions. A practical approach is described to estimate the amount of sampling effort required to have sufficient statistical power to identify species-specific “hotspots” and “coldspots” of marine bird abundance and occurrence in an offshore environment divided into discrete spatial units (e.g., lease blocks), where “hotspots” and “coldspots” are defined relative to a reference (e.g., regional) mean abundance and/or occurrence probability for each species of interest. For example, a location with average abundance or occurrence that is three times larger the mean (3x effect size) could be defined as a “hotspot,” and a location that is three times smaller than the mean (1/3x effect size) as a “coldspot.” The choice of the effect size used to define hot and coldspots will generally depend on a combination of ecological and regulatory considerations. A method is also developed for testing the statistical significance of possible hotspots and coldspots. Both methods are illustrated with historical seabird survey data from the USGS Avian Compendium Database. Our approach consists of five main components: 1. A review of the primary scientific literature on statistical modeling of animal group size and avian count data to develop a candidate set of statistical distributions that have been used or may be useful to model seabird counts. 2. Statistical power curves for one-sample, one-tailed Monte Carlo significance tests of differences of observed small-sample means from a specified reference distribution. These curves show the power to detect "hotspots" or "coldspots" of occurrence and abundance at a range of effect sizes, given assumptions which we discuss. 3. A model selection procedure, based on maximum likelihood fits of models in the candidate set, to determine an appropriate statistical distribution to describe counts of a given species in a particular region and season. 4. Using a large database of historical at-sea seabird survey data, we applied this technique to identify appropriate statistical distributions for modeling a variety of species, allowing the distribution to vary by season. For each species and season, we used the selected distribution to calculate and map retrospective statistical power to detect hotspots and coldspots, and map pvalues from Monte Carlo significance tests of hotspots and coldspots, in discrete lease blocks designated by the U.S. Department of Interior, Bureau of Ocean Energy Management (BOEM). 5. Because our definition of hotspots and coldspots does not explicitly include variability over time, we examine the relationship between the temporal scale of sampling and the proportion of variance captured in time series of key environmental correlates of marine bird abundance, as well as available marine bird abundance time series, and use these analyses to develop recommendations for the temporal distribution of sampling to adequately represent both shortterm and long-term variability. We conclude by presenting a schematic “decision tree” showing how this power analysis approach would fit in a general framework for avian survey design, and discuss implications of model assumptions and results. We discuss avenues for future development of this work, and recommendations for practical implementation in the context of siting and wildlife assessment for offshore renewable energy development projects.

  4. Introduction to Sample Size Choice for Confidence Intervals Based on "t" Statistics

    ERIC Educational Resources Information Center

    Liu, Xiaofeng Steven; Loudermilk, Brandon; Simpson, Thomas

    2014-01-01

    Sample size can be chosen to achieve a specified width in a confidence interval. The probability of obtaining a narrow width given that the confidence interval includes the population parameter is defined as the power of the confidence interval, a concept unfamiliar to many practitioners. This article shows how to utilize the Statistical Analysis…

  5. A New Approach to Galaxy Morphology. I. Analysis of the Sloan Digital Sky Survey Early Data Release

    NASA Astrophysics Data System (ADS)

    Abraham, Roberto G.; van den Bergh, Sidney; Nair, Preethi

    2003-05-01

    In this paper we present a new statistic for quantifying galaxy morphology based on measurements of the Gini coefficient of galaxy light distributions. This statistic is easy to measure and is commonly used in econometrics to measure how wealth is distributed in human populations. When applied to galaxy images, the Gini coefficient provides a quantitative measure of the inequality with which a galaxy's light is distributed among its constituent pixels. We measure the Gini coefficient of local galaxies in the Early Data Release of the Sloan Digital Sky Survey and demonstrate that this quantity is closely correlated with measurements of central concentration, but with significant scatter. This scatter is almost entirely due to variations in the mean surface brightness of galaxies. By exploring the distribution of galaxies in the three-dimensional parameter space defined by the Gini coefficient, central concentration, and mean surface brightness, we show that all nearby galaxies lie on a well-defined two-dimensional surface (a slightly warped plane) embedded within a three-dimensional parameter space. By associating each galaxy sample with the equation of this plane, we can encode the morphological composition of the entire SDSS g*-band sample using the following three numbers: {22.451, 5.366, 7.010}. The i*-band sample is encoded as {22.149, 5.373, and 7.627}.

  6. Prospects for AGN Science using the ART-XC on the SRG Mission

    NASA Technical Reports Server (NTRS)

    Swartz, Douglas A.; Elsner, Ronald F.; Gubarev, Mikhail V.; O'Dell, Stephen L.; Ramsey, Brian D.; Bonamente, Massimiliano

    2012-01-01

    The enhanced hard X-ray sensitivity provided by the Astronomical Roentgen Telescope to the Spectrum Roentgen Gamma mission facilitates the detection of heavily obscured and other hard-spectrum cosmic X-ray sources. The SRG all-sky survey will obtain large, statistically-well-defined samples of active galactic nuclei (AGN) including a significant population of local heavily-obscured AGN. In anticipation of the SRG all-sky survey, we investigate the prospects for refining the bright end of the AGN luminosity function and determination of the local black hole mass function and comparing the spatial distribution of AGN with large-scale structure defined by galaxy clusters and groups. Particular emphasis is placed on studies of the deep survey Ecliptic Pole regions.

  7. Progress in tropical isotope dendroclimatology

    NASA Astrophysics Data System (ADS)

    Evans, M. N.; Schrag, D. P.; Poussart, P. F.; Anchukaitis, K. J.

    2005-12-01

    The terrestrial tropics remain an important gap in the growing high resolution proxy network used to characterize the mean state and variability of the hydrological cycle. Here we review early efforts to develop a new class of proxy paleorainfall/humidity indicators using intraseasonal to interannual-resolution stable isotope data from tropical trees. The approach invokes a recently published model of oxygen isotopic composition of alpha-cellulose, rapid methods for cellulose extraction from raw wood, and continuous flow isotope ratio mass spectrometry to develop proxy chronological, rainfall and growth rate estimates from tropical trees, even those lacking annual rings. Isotopically-derived age models may be confirmed for modern intervals using trees of known age, radiocarbon measurements, direct measurements of tree diameter, and time series replication. Studies are now underway at a number of laboratories on samples from Costa Rica, northwestern coastal Peru, Indonesia, Thailand, New Guinea, Paraguay, Brazil, India, and the South American Altiplano. Improved sample extraction chemistry and online pyrolysis techniques should increase sample throughput, precision, and time series replication. Statistical calibration together with simple forward modeling based on the well-observed modern period can provide for objective interpretation of the data. Ultimately, replicated data series with well-defined uncertainties can be entered into multiproxy efforts to define aspects of tropical hydrological variability associated with ENSO, the meridional overturning circulation, and the monsoon systems.

  8. Identifying natural flow regimes using fish communities

    NASA Astrophysics Data System (ADS)

    Chang, Fi-John; Tsai, Wen-Ping; Wu, Tzu-Ching; Chen, Hung-kwai; Herricks, Edwin E.

    2011-10-01

    SummaryModern water resources management has adopted natural flow regimes as reasonable targets for river restoration and conservation. The characterization of a natural flow regime begins with the development of hydrologic statistics from flow records. However, little guidance exists for defining the period of record needed for regime determination. In Taiwan, the Taiwan Eco-hydrological Indicator System (TEIS), a group of hydrologic statistics selected for fisheries relevance, is being used to evaluate ecological flows. The TEIS consists of a group of hydrologic statistics selected to characterize the relationships between flow and the life history of indigenous species. Using the TEIS and biosurvey data for Taiwan, this paper identifies the length of hydrologic record sufficient for natural flow regime characterization. To define the ecological hydrology of fish communities, this study connected hydrologic statistics to fish communities by using methods to define antecedent conditions that influence existing community composition. A moving average method was applied to TEIS statistics to reflect the effects of antecedent flow condition and a point-biserial correlation method was used to relate fisheries collections with TEIS statistics. The resulting fish species-TEIS (FISH-TEIS) hydrologic statistics matrix takes full advantage of historical flows and fisheries data. The analysis indicates that, in the watersheds analyzed, averaging TEIS statistics for the present year and 3 years prior to the sampling date, termed MA(4), is sufficient to develop a natural flow regime. This result suggests that flow regimes based on hydrologic statistics for the period of record can be replaced by regimes developed for sampled fish communities.

  9. Seismic sample areas defined from incomplete catalogues: an application to the Italian territory

    NASA Astrophysics Data System (ADS)

    Mulargia, F.; Tinti, S.

    1985-11-01

    The comprehensive understanding of earthquake source-physics under real conditions requires the study not of single faults as separate entities but rather of a seismically active region as a whole, accounting for the interaction among different structures. We define "seismic sample area" the most convenient region to be used as a natural laboratory for the study of seismic source physics. This coincides with the region where the average large magnitude seismicity is the highest. To this end, time and space future distributions of large earthquakes are to be estimated. Using catalog seismicity as an input, the rate of occurrence is not constant but appears generally biased by incompleteness in some parts of the catalog and possible nonstationarities in seismic activity. We present a statistical procedure which is capable, under a few mild assumptions, of both detecting nonstationarities in seismicity and finding the incomplete parts of a seismic catalog. The procedure is based on Kolmogorov-Smirnov nonparametric statistics, and can be applied without a priori assuming the parent distribution of the events. The efficiency of this procedure allows the analysis of small data sets. An application to the Italian territory is presented, using the most recent version of the ENEL seismic catalog. Seismic activity takes place in six well defined areas but only five of them have a number of events sufficient for analysis. Barring a few exceptions, seismicity is found stationary throughout the whole catalog span 1000-1980. The eastern Alps region stands out as the best "sample area", with the highest average probability of event occurrence per time and area unit. Final objective of this characterization is to stimulate a program of intensified research.

  10. Statistical Analyses of Scatterplots to Identify Important Factors in Large-Scale Simulations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kleijnen, J.P.C.; Helton, J.C.

    1999-04-01

    The robustness of procedures for identifying patterns in scatterplots generated in Monte Carlo sensitivity analyses is investigated. These procedures are based on attempts to detect increasingly complex patterns in the scatterplots under consideration and involve the identification of (1) linear relationships with correlation coefficients, (2) monotonic relationships with rank correlation coefficients, (3) trends in central tendency as defined by means, medians and the Kruskal-Wallis statistic, (4) trends in variability as defined by variances and interquartile ranges, and (5) deviations from randomness as defined by the chi-square statistic. The following two topics related to the robustness of these procedures are consideredmore » for a sequence of example analyses with a large model for two-phase fluid flow: the presence of Type I and Type II errors, and the stability of results obtained with independent Latin hypercube samples. Observations from analysis include: (1) Type I errors are unavoidable, (2) Type II errors can occur when inappropriate analysis procedures are used, (3) physical explanations should always be sought for why statistical procedures identify variables as being important, and (4) the identification of important variables tends to be stable for independent Latin hypercube samples.« less

  11. EVALUATION OF A NEW MEAN SCALED AND MOMENT ADJUSTED TEST STATISTIC FOR SEM.

    PubMed

    Tong, Xiaoxiao; Bentler, Peter M

    2013-01-01

    Recently a new mean scaled and skewness adjusted test statistic was developed for evaluating structural equation models in small samples and with potentially nonnormal data, but this statistic has received only limited evaluation. The performance of this statistic is compared to normal theory maximum likelihood and two well-known robust test statistics. A modification to the Satorra-Bentler scaled statistic is developed for the condition that sample size is smaller than degrees of freedom. The behavior of the four test statistics is evaluated with a Monte Carlo confirmatory factor analysis study that varies seven sample sizes and three distributional conditions obtained using Headrick's fifth-order transformation to nonnormality. The new statistic performs badly in most conditions except under the normal distribution. The goodness-of-fit χ(2) test based on maximum-likelihood estimation performed well under normal distributions as well as under a condition of asymptotic robustness. The Satorra-Bentler scaled test statistic performed best overall, while the mean scaled and variance adjusted test statistic outperformed the others at small and moderate sample sizes under certain distributional conditions.

  12. Reflexion on linear regression trip production modelling method for ensuring good model quality

    NASA Astrophysics Data System (ADS)

    Suprayitno, Hitapriya; Ratnasari, Vita

    2017-11-01

    Transport Modelling is important. For certain cases, the conventional model still has to be used, in which having a good trip production model is capital. A good model can only be obtained from a good sample. Two of the basic principles of a good sampling is having a sample capable to represent the population characteristics and capable to produce an acceptable error at a certain confidence level. It seems that this principle is not yet quite understood and used in trip production modeling. Therefore, investigating the Trip Production Modelling practice in Indonesia and try to formulate a better modeling method for ensuring the Model Quality is necessary. This research result is presented as follows. Statistics knows a method to calculate span of prediction value at a certain confidence level for linear regression, which is called Confidence Interval of Predicted Value. The common modeling practice uses R2 as the principal quality measure, the sampling practice varies and not always conform to the sampling principles. An experiment indicates that small sample is already capable to give excellent R2 value and sample composition can significantly change the model. Hence, good R2 value, in fact, does not always mean good model quality. These lead to three basic ideas for ensuring good model quality, i.e. reformulating quality measure, calculation procedure, and sampling method. A quality measure is defined as having a good R2 value and a good Confidence Interval of Predicted Value. Calculation procedure must incorporate statistical calculation method and appropriate statistical tests needed. A good sampling method must incorporate random well distributed stratified sampling with a certain minimum number of samples. These three ideas need to be more developed and tested.

  13. Eye-gaze determination of user intent at the computer interface

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Goldberg, J.H.; Schryver, J.C.

    1993-12-31

    Determination of user intent at the computer interface through eye-gaze monitoring can significantly aid applications for the disabled, as well as telerobotics and process control interfaces. Whereas current eye-gaze control applications are limited to object selection and x/y gazepoint tracking, a methodology was developed here to discriminate a more abstract interface operation: zooming-in or out. This methodology first collects samples of eve-gaze location looking at controlled stimuli, at 30 Hz, just prior to a user`s decision to zoom. The sample is broken into data frames, or temporal snapshots. Within a data frame, all spatial samples are connected into a minimummore » spanning tree, then clustered, according to user defined parameters. Each cluster is mapped to one in the prior data frame, and statistics are computed from each cluster. These characteristics include cluster size, position, and pupil size. A multiple discriminant analysis uses these statistics both within and between data frames to formulate optimal rules for assigning the observations into zooming, zoom-out, or no zoom conditions. The statistical procedure effectively generates heuristics for future assignments, based upon these variables. Future work will enhance the accuracy and precision of the modeling technique, and will empirically test users in controlled experiments.« less

  14. Using the bootstrap to establish statistical significance for relative validity comparisons among patient-reported outcome measures

    PubMed Central

    2013-01-01

    Background Relative validity (RV), a ratio of ANOVA F-statistics, is often used to compare the validity of patient-reported outcome (PRO) measures. We used the bootstrap to establish the statistical significance of the RV and to identify key factors affecting its significance. Methods Based on responses from 453 chronic kidney disease (CKD) patients to 16 CKD-specific and generic PRO measures, RVs were computed to determine how well each measure discriminated across clinically-defined groups of patients compared to the most discriminating (reference) measure. Statistical significance of RV was quantified by the 95% bootstrap confidence interval. Simulations examined the effects of sample size, denominator F-statistic, correlation between comparator and reference measures, and number of bootstrap replicates. Results The statistical significance of the RV increased as the magnitude of denominator F-statistic increased or as the correlation between comparator and reference measures increased. A denominator F-statistic of 57 conveyed sufficient power (80%) to detect an RV of 0.6 for two measures correlated at r = 0.7. Larger denominator F-statistics or higher correlations provided greater power. Larger sample size with a fixed denominator F-statistic or more bootstrap replicates (beyond 500) had minimal impact. Conclusions The bootstrap is valuable for establishing the statistical significance of RV estimates. A reasonably large denominator F-statistic (F > 57) is required for adequate power when using the RV to compare the validity of measures with small or moderate correlations (r < 0.7). Substantially greater power can be achieved when comparing measures of a very high correlation (r > 0.9). PMID:23721463

  15. Status and understanding of groundwater quality in the northern San Joaquin Basin, 2005

    USGS Publications Warehouse

    Bennett, George L.; Fram, Miranda S.; Belitz, Kenneth; Jurgens, Bryant C.

    2010-01-01

    Groundwater quality in the 2,079 square mile Northern San Joaquin Basin (Northern San Joaquin) study unit was investigated from December 2004 through February 2005 as part of the Priority Basin Project of the Groundwater Ambient Monitoring and Assessment (GAMA) Program. The GAMA Priority Basin Project was developed in response to the Groundwater Quality Monitoring Act of 2001 that was passed by the State of California and is being conducted by the California State Water Resources Control Board in collaboration with the U.S. Geological Survey and the Lawrence Livermore National Laboratory. The Northern San Joaquin study unit was the third study unit to be designed and sampled as part of the Priority Basin Project. Results of the study provide a spatially unbiased assessment of the quality of raw (untreated) groundwater, as well as a statistically consistent basis for comparing water quality throughout California. Samples were collected from 61 wells in parts of Alameda, Amador, Calaveras, Contra Costa, San Joaquin, and Stanislaus Counties; 51 of the wells were selected using a spatially distributed, randomized grid-based approach to provide statistical representation of the study area (grid wells), and 10 of the wells were sampled to increase spatial density and provide additional information for the evaluation of water chemistry in the study unit (understanding/flowpath wells). The primary aquifer systems (hereinafter, primary aquifers) assessed in this study are defined by the depth intervals of the wells in the California Department of Public Health database for each study unit. The quality of groundwater in shallow or deep water-bearing zones may differ from quality of groundwater in the primary aquifers; shallow groundwater may be more vulnerable to contamination from the surface. Two types of assessments were made: (1) status, assessment of the current quality of the groundwater resource; and (2) understanding, identification of the natural and human factors affecting groundwater quality. Relative-concentrations (sample concentrations divided by benchmark concentrations) were used for evaluating groundwater quality for those constituents that have Federal or California regulatory or non-regulatory benchmarks for drinking-water quality. Benchmarks used in this study were either health-based (regulatory and non-regulatory) or aesthetic based (non-regulatory). For inorganic constituents, relative-concentrations were classified as high (equal to or greater than 1.0), indicating relative-concentrations greater than benchmarks; moderate (equal to or greater than 0.5, and less than 1.0); or, low (less than 0.5). For organic and special- interest constituents [1,2,3-trichloropropane (1,2,3-TCP), N-nitrosodimethylamine (NDMA), and perchlorate], relative- concentrations were classified as high (equal to or greater than 1.0); moderate (equal to or greater than 0.1 and less than 1.0); or, low (less than 0.1). Aquifer-scale proportion was used as the primary metric in the status assessment for groundwater quality. High aquifer- scale proportion is defined as the percentage of the primary aquifer with relative-concentrations greater than 1.0; moderate and low aquifer-scale proportions are defined as the percentage of the primary aquifer with moderate and low relative- concentrations, respectively. The methods used to calculate aquifer-scale proportions are based on an equal-area grid; thus, the proportions are areal rather than volumetric. Two statistical approaches - grid-based, which used one value per grid cell, and spatially weighted, which used the full dataset - were used to calculate aquifer-scale proportions for individual constituents and classes of constituents. The spatially weighted estimates of high aquifer-scale proportions were within the 90-percent confidence intervals of the grid-based estimates in all cases. The understanding assessment used statistical correlations between constituent relative-concentrations and

  16. The grapevine expression atlas reveals a deep transcriptome shift driving the entire plant into a maturation program.

    PubMed

    Fasoli, Marianna; Dal Santo, Silvia; Zenoni, Sara; Tornielli, Giovanni Battista; Farina, Lorenzo; Zamboni, Anita; Porceddu, Andrea; Venturini, Luca; Bicego, Manuele; Murino, Vittorio; Ferrarini, Alberto; Delledonne, Massimo; Pezzotti, Mario

    2012-09-01

    We developed a genome-wide transcriptomic atlas of grapevine (Vitis vinifera) based on 54 samples representing green and woody tissues and organs at different developmental stages as well as specialized tissues such as pollen and senescent leaves. Together, these samples expressed ∼91% of the predicted grapevine genes. Pollen and senescent leaves had unique transcriptomes reflecting their specialized functions and physiological status. However, microarray and RNA-seq analysis grouped all the other samples into two major classes based on maturity rather than organ identity, namely, the vegetative/green and mature/woody categories. This division represents a fundamental transcriptomic reprogramming during the maturation process and was highlighted by three statistical approaches identifying the transcriptional relationships among samples (correlation analysis), putative biomarkers (O2PLS-DA approach), and sets of strongly and consistently expressed genes that define groups (topics) of similar samples (biclustering analysis). Gene coexpression analysis indicated that the mature/woody developmental program results from the reiterative coactivation of pathways that are largely inactive in vegetative/green tissues, often involving the coregulation of clusters of neighboring genes and global regulation based on codon preference. This global transcriptomic reprogramming during maturation has not been observed in herbaceous annual species and may be a defining characteristic of perennial woody plants.

  17. Cosmological velocity correlations - Observations and model predictions

    NASA Technical Reports Server (NTRS)

    Gorski, Krzysztof M.; Davis, Marc; Strauss, Michael A.; White, Simon D. M.; Yahil, Amos

    1989-01-01

    By applying the present simple statistics for two-point cosmological peculiar velocity-correlation measurements to the actual data sets of the Local Supercluster spiral galaxy of Aaronson et al. (1982) and the elliptical galaxy sample of Burstein et al. (1987), as well as to the velocity field predicted by the distribution of IRAS galaxies, a coherence length of 1100-1600 km/sec is obtained. Coherence length is defined as that separation at which the correlations drop to half their zero-lag value. These results are compared with predictions from two models of large-scale structure formation: that of cold dark matter and that of baryon isocurvature proposed by Peebles (1980). N-body simulations of these models are performed to check the linear theory predictions and measure sampling fluctuations.

  18. Estimating statistical uncertainty of Monte Carlo efficiency-gain in the context of a correlated sampling Monte Carlo code for brachytherapy treatment planning with non-normal dose distribution.

    PubMed

    Mukhopadhyay, Nitai D; Sampson, Andrew J; Deniz, Daniel; Alm Carlsson, Gudrun; Williamson, Jeffrey; Malusek, Alexandr

    2012-01-01

    Correlated sampling Monte Carlo methods can shorten computing times in brachytherapy treatment planning. Monte Carlo efficiency is typically estimated via efficiency gain, defined as the reduction in computing time by correlated sampling relative to conventional Monte Carlo methods when equal statistical uncertainties have been achieved. The determination of the efficiency gain uncertainty arising from random effects, however, is not a straightforward task specially when the error distribution is non-normal. The purpose of this study is to evaluate the applicability of the F distribution and standardized uncertainty propagation methods (widely used in metrology to estimate uncertainty of physical measurements) for predicting confidence intervals about efficiency gain estimates derived from single Monte Carlo runs using fixed-collision correlated sampling in a simplified brachytherapy geometry. A bootstrap based algorithm was used to simulate the probability distribution of the efficiency gain estimates and the shortest 95% confidence interval was estimated from this distribution. It was found that the corresponding relative uncertainty was as large as 37% for this particular problem. The uncertainty propagation framework predicted confidence intervals reasonably well; however its main disadvantage was that uncertainties of input quantities had to be calculated in a separate run via a Monte Carlo method. The F distribution noticeably underestimated the confidence interval. These discrepancies were influenced by several photons with large statistical weights which made extremely large contributions to the scored absorbed dose difference. The mechanism of acquiring high statistical weights in the fixed-collision correlated sampling method was explained and a mitigation strategy was proposed. Copyright © 2011 Elsevier Ltd. All rights reserved.

  19. Developing Students' Reasoning about Samples and Sampling Variability as a Path to Expert Statistical Thinking

    ERIC Educational Resources Information Center

    Garfield, Joan; Le, Laura; Zieffler, Andrew; Ben-Zvi, Dani

    2015-01-01

    This paper describes the importance of developing students' reasoning about samples and sampling variability as a foundation for statistical thinking. Research on expert-novice thinking as well as statistical thinking is reviewed and compared. A case is made that statistical thinking is a type of expert thinking, and as such, research…

  20. Feature-based and statistical methods for analyzing the Deepwater Horizon oil spill with AVIRIS imagery

    USGS Publications Warehouse

    Rand, R.S.; Clark, R.N.; Livo, K.E.

    2011-01-01

    The Deepwater Horizon oil spill covered a very large geographical area in the Gulf of Mexico creating potentially serious environmental impacts on both marine life and the coastal shorelines. Knowing the oil's areal extent and thickness as well as denoting different categories of the oil's physical state is important for assessing these impacts. High spectral resolution data in hyperspectral imagery (HSI) sensors such as Airborne Visible and Infrared Imaging Spectrometer (AVIRIS) provide a valuable source of information that can be used for analysis by semi-automatic methods for tracking an oil spill's areal extent, oil thickness, and oil categories. However, the spectral behavior of oil in water is inherently a highly non-linear and variable phenomenon that changes depending on oil thickness and oil/water ratios. For certain oil thicknesses there are well-defined absorption features, whereas for very thin films sometimes there are almost no observable features. Feature-based imaging spectroscopy methods are particularly effective at classifying materials that exhibit specific well-defined spectral absorption features. Statistical methods are effective at classifying materials with spectra that exhibit a considerable amount of variability and that do not necessarily exhibit well-defined spectral absorption features. This study investigates feature-based and statistical methods for analyzing oil spills using hyperspectral imagery. The appropriate use of each approach is investigated and a combined feature-based and statistical method is proposed.

  1. Min and Max Exponential Extreme Interval Values and Statistics

    ERIC Educational Resources Information Center

    Jance, Marsha; Thomopoulos, Nick

    2009-01-01

    The extreme interval values and statistics (expected value, median, mode, standard deviation, and coefficient of variation) for the smallest (min) and largest (max) values of exponentially distributed variables with parameter ? = 1 are examined for different observation (sample) sizes. An extreme interval value g[subscript a] is defined as a…

  2. A quantitative approach to evolution of music and philosophy

    NASA Astrophysics Data System (ADS)

    Vieira, Vilson; Fabbri, Renato; Travieso, Gonzalo; Oliveira, Osvaldo N., Jr.; da Fontoura Costa, Luciano

    2012-08-01

    The development of new statistical and computational methods is increasingly making it possible to bridge the gap between hard sciences and humanities. In this study, we propose an approach based on a quantitative evaluation of attributes of objects in fields of humanities, from which concepts such as dialectics and opposition are formally defined mathematically. As case studies, we analyzed the temporal evolution of classical music and philosophy by obtaining data for 8 features characterizing the corresponding fields for 7 well-known composers and philosophers, which were treated with multivariate statistics and pattern recognition methods. A bootstrap method was applied to avoid statistical bias caused by the small sample data set, with which hundreds of artificial composers and philosophers were generated, influenced by the 7 names originally chosen. Upon defining indices for opposition, skewness and counter-dialectics, we confirmed the intuitive analysis of historians in that classical music evolved according to a master-apprentice tradition, while in philosophy changes were driven by opposition. Though these case studies were meant only to show the possibility of treating phenomena in humanities quantitatively, including a quantitative measure of concepts such as dialectics and opposition, the results are encouraging for further application of the approach presented here to many other areas, since it is entirely generic.

  3. Statistical Hypothesis Testing in Intraspecific Phylogeography: NCPA versus ABC

    PubMed Central

    Templeton, Alan R.

    2009-01-01

    Nested clade phylogeographic analysis (NCPA) and approximate Bayesian computation (ABC) have been used to test phylogeographic hypotheses. Multilocus NCPA tests null hypotheses, whereas ABC discriminates among a finite set of alternatives. The interpretive criteria of NCPA are explicit and allow complex models to be built from simple components. The interpretive criteria of ABC are ad hoc and require the specification of a complete phylogeographic model. The conclusions from ABC are often influenced by implicit assumptions arising from the many parameters needed to specify a complex model. These complex models confound many assumptions so that biological interpretations are difficult. Sampling error is accounted for in NCPA, but ABC ignores important sources of sampling error that creates pseudo-statistical power. NCPA generates the full sampling distribution of its statistics, but ABC only yields local probabilities, which in turn make it impossible to distinguish between a good fitting model, a non-informative model, and an over-determined model. Both NCPA and ABC use approximations, but convergences of the approximations used in NCPA are well defined whereas those in ABC are not. NCPA can analyze a large number of locations, but ABC cannot. Finally, the dimensionality of tested hypothesis is known in NCPA, but not for ABC. As a consequence, the “probabilities” generated by ABC are not true probabilities and are statistically non-interpretable. Accordingly, ABC should not be used for hypothesis testing, but simulation approaches are valuable when used in conjunction with NCPA or other methods that do not rely on highly parameterized models. PMID:19192182

  4. [Evaluation of using statistical methods in selected national medical journals].

    PubMed

    Sych, Z

    1996-01-01

    The paper covers the performed evaluation of frequency with which the statistical methods were applied in analyzed works having been published in six selected, national medical journals in the years 1988-1992. For analysis the following journals were chosen, namely: Klinika Oczna, Medycyna Pracy, Pediatria Polska, Polski Tygodnik Lekarski, Roczniki Państwowego Zakładu Higieny, Zdrowie Publiczne. Appropriate number of works up to the average in the remaining medical journals was randomly selected from respective volumes of Pol. Tyg. Lek. The studies did not include works wherein the statistical analysis was not implemented, which referred both to national and international publications. That exemption was also extended to review papers, casuistic ones, reviews of books, handbooks, monographies, reports from scientific congresses, as well as papers on historical topics. The number of works was defined in each volume. Next, analysis was performed to establish the mode of finding out a suitable sample in respective studies, differentiating two categories: random and target selections. Attention was also paid to the presence of control sample in the individual works. In the analysis attention was also focussed on the existence of sample characteristics, setting up three categories: complete, partial and lacking. In evaluating the analyzed works an effort was made to present the results of studies in tables and figures (Tab. 1, 3). Analysis was accomplished with regard to the rate of employing statistical methods in analyzed works in relevant volumes of six selected, national medical journals for the years 1988-1992, simultaneously determining the number of works, in which no statistical methods were used. Concurrently the frequency of applying the individual statistical methods was analyzed in the scrutinized works. Prominence was given to fundamental statistical methods in the field of descriptive statistics (measures of position, measures of dispersion) as well as most important methods of mathematical statistics such as parametric tests of significance, analysis of variance (in single and dual classifications). non-parametric tests of significance, correlation and regression. The works, in which use was made of either multiple correlation or multiple regression or else more complex methods of studying the relationship for two or more numbers of variables, were incorporated into the works whose statistical methods were constituted by correlation and regression as well as other methods, e.g. statistical methods being used in epidemiology (coefficients of incidence and morbidity, standardization of coefficients, survival tables) factor analysis conducted by Jacobi-Hotellng's method, taxonomic methods and others. On the basis of the performed studies it has been established that the frequency of employing statistical methods in the six selected national, medical journals in the years 1988-1992 was 61.1-66.0% of the analyzed works (Tab. 3), and they generally were almost similar to the frequency provided in English language medical journals. On a whole, no significant differences were disclosed in the frequency of applied statistical methods (Tab. 4) as well as in frequency of random tests (Tab. 3) in the analyzed works, appearing in the medical journals in respective years 1988-1992. The most frequently used statistical methods in analyzed works for 1988-1992 were the measures of position 44.2-55.6% and measures of dispersion 32.5-38.5% as well as parametric tests of significance 26.3-33.1% of the works analyzed (Tab. 4). For the purpose of increasing the frequency and reliability of the used statistical methods, the didactics should be widened in the field of biostatistics at medical studies and postgraduation training designed for physicians and scientific-didactic workers.

  5. Radio-Optical Alignments in a Low Radio Luminosity Sample

    NASA Technical Reports Server (NTRS)

    Lacy, Mark; Ridgway, Susan E.; Wold, Margrethe; Lilje, Per B.; Rawlings, Steve

    1999-01-01

    We present an optically-based study of the alignment between the radio axes and the optical major axes of eight z approximately 0.7 radio galaxies in a 7C sample. The radio galaxies in this sample are approximately 20-times less radio luminous than 3C galaxies at the same redshift, and are significantly less radio-luminous than any other well-defined samples studied to date. Using Nordic Optical Telescope images taken in good seeing conditions at rest-frame wavelengths just longward of the 4000A break, we find a statistically significant alignment effect in the 7C sample. Furthermore, in two cases where the aligned components are well separated from the host we have been able to confirm spectroscopically that they are indeed at the same redshift as the radio galaxy. However, a quantitative analysis of the alignment in this sample and in a corresponding 3C sample from HST (Hubble Space Telescope) archival data indicates that the percentage of aligned flux may be lower and of smaller spatial scale in the 7C sample. Our study suggests that alignments on the 50-kpc scale are probably closely related to the radio luminosity, whereas those on the 15 kpc scale are not. We discuss these results in the context of popular models for the alignment effect.

  6. AGES: THE AGN AND GALAXY EVOLUTION SURVEY

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kochanek, C. S.; Eisenstein, D. J.; Caldwell, N.

    2012-05-01

    The AGN and Galaxy Evolution Survey (AGES) is a redshift survey covering, in its standard fields, 7.7 deg{sup 2} of the Booetes field of the NOAO Deep Wide-Field Survey. The final sample consists of 23,745 redshifts. There are well-defined galaxy samples in 10 bands (the B{sub W} , R, I, J, K, IRAC 3.6, 4.5, 5.8, and 8.0 {mu}m, and MIPS 24 {mu}m bands) to a limiting magnitude of I < 20 mag for spectroscopy. For these galaxies, we obtained 18,163 redshifts from a sample of 35,200 galaxies, where random sparse sampling was used to define statistically complete sub-samples inmore » all 10 photometric bands. The median galaxy redshift is 0.31, and 90% of the redshifts are in the range 0.085 < z < 0.66. Active galactic nuclei (AGNs) were selected as radio, X-ray, IRAC mid-IR, and MIPS 24 {mu}m sources to fainter limiting magnitudes (I < 22.5 mag for point sources). Redshifts were obtained for 4764 quasars and galaxies with AGN signatures, with 2926, 1718, 605, 119, and 13 above redshifts of 0.5, 1, 2, 3, and 4, respectively. We detail all the AGES selection procedures and present the complete spectroscopic redshift catalogs and spectral energy distribution decompositions. Photometric redshift estimates are provided for all sources in the AGES samples.« less

  7. Experimental control in software reliability certification

    NASA Technical Reports Server (NTRS)

    Trammell, Carmen J.; Poore, Jesse H.

    1994-01-01

    There is growing interest in software 'certification', i.e., confirmation that software has performed satisfactorily under a defined certification protocol. Regulatory agencies, customers, and prospective reusers all want assurance that a defined product standard has been met. In other industries, products are typically certified under protocols in which random samples of the product are drawn, tests characteristic of operational use are applied, analytical or statistical inferences are made, and products meeting a standard are 'certified' as fit for use. A warranty statement is often issued upon satisfactory completion of a certification protocol. This paper outlines specific engineering practices that must be used to preserve the validity of the statistical certification testing protocol. The assumptions associated with a statistical experiment are given, and their implications for statistical testing of software are described.

  8. DNA analysis in Disaster Victim Identification.

    PubMed

    Montelius, Kerstin; Lindblom, Bertil

    2012-06-01

    DNA profiling and matching is one of the primary methods to identify missing persons in a disaster, as defined by the Interpol Disaster Victim Identification Guide. The process to identify a victim by DNA includes: the collection of the best possible ante-mortem (AM) samples, the choice of post-mortem (PM) samples, DNA-analysis, matching and statistical weighting of the genetic relationship or match. Each disaster has its own scenario, and each scenario defines its own methods for identification of the deceased.

  9. [Practical aspects regarding sample size in clinical research].

    PubMed

    Vega Ramos, B; Peraza Yanes, O; Herrera Correa, G; Saldívar Toraya, S

    1996-01-01

    The knowledge of the right sample size let us to be sure if the published results in medical papers had a suitable design and a proper conclusion according to the statistics analysis. To estimate the sample size we must consider the type I error, type II error, variance, the size of the effect, significance and power of the test. To decide what kind of mathematics formula will be used, we must define what kind of study we have, it means if its a prevalence study, a means values one or a comparative one. In this paper we explain some basic topics of statistics and we describe four simple samples of estimation of sample size.

  10. Computerized EEG analysis for studying the effect of drugs on the central nervous system.

    PubMed

    Rosadini, G; Cavazza, B; Rodriguez, G; Sannita, W G; Siccardi, A

    1977-11-01

    Samples of our experience in quantitative pharmaco-EEG are reviewed to discuss and define its applicability and limits. Simple processing systems, such as the computation of Hjorth's descriptors, are useful for on-line monitoring of drug-induced EEG modifications which are evident also at the visual visual analysis. Power spectral analysis is suitable to identify and quantify EEG effects not evident at the visual inspection. It demonstrated how the EEG effects of compounds in a long-acting formulation vary according to the sampling time and the explored cerebral area. EEG modifications not detected by power spectral analysis can be defined by comparing statistically (F test) the spectral values of the EEG from a single lead at the different samples (longitudinal comparison), or the spectral values from different leads at any sample (intrahemispheric comparison). The presently available procedures of quantitative pharmaco-EEG are effective when applied to the study of mutltilead EEG recordings in a statistically significant sample of population. They do not seem reliable in the monitoring of directing of neuropyschiatric therapies in single patients, due to individual variability of drug effects.

  11. Delaunay-based derivative-free optimization for efficient minimization of time-averaged statistics of turbulent flows

    NASA Astrophysics Data System (ADS)

    Beyhaghi, Pooriya

    2016-11-01

    This work considers the problem of the efficient minimization of the infinite time average of a stationary ergodic process in the space of a handful of independent parameters which affect it. Problems of this class, derived from physical or numerical experiments which are sometimes expensive to perform, are ubiquitous in turbulence research. In such problems, any given function evaluation, determined with finite sampling, is associated with a quantifiable amount of uncertainty, which may be reduced via additional sampling. This work proposes the first algorithm of this type. Our algorithm remarkably reduces the overall cost of the optimization process for problems of this class. Further, under certain well-defined conditions, rigorous proof of convergence is established to the global minimum of the problem considered.

  12. Planetary mass function and planetary systems

    NASA Astrophysics Data System (ADS)

    Dominik, M.

    2011-02-01

    With planets orbiting stars, a planetary mass function should not be seen as a low-mass extension of the stellar mass function, but a proper formalism needs to take care of the fact that the statistical properties of planet populations are linked to the properties of their respective host stars. This can be accounted for by describing planet populations by means of a differential planetary mass-radius-orbit function, which together with the fraction of stars with given properties that are orbited by planets and the stellar mass function allows the derivation of all statistics for any considered sample. These fundamental functions provide a framework for comparing statistics that result from different observing techniques and campaigns which all have their very specific selection procedures and detection efficiencies. Moreover, recent results both from gravitational microlensing campaigns and radial-velocity surveys of stars indicate that planets tend to cluster in systems rather than being the lonely child of their respective parent star. While planetary multiplicity in an observed system becomes obvious with the detection of several planets, its quantitative assessment however comes with the challenge to exclude the presence of further planets. Current exoplanet samples begin to give us first hints at the population statistics, whereas pictures of planet parameter space in its full complexity call for samples that are 2-4 orders of magnitude larger. In order to derive meaningful statistics, however, planet detection campaigns need to be designed in such a way that well-defined fully deterministic target selection, monitoring and detection criteria are applied. The probabilistic nature of gravitational microlensing makes this technique an illustrative example of all the encountered challenges and uncertainties.

  13. The Probability of Obtaining Two Statistically Different Test Scores as a Test Index

    ERIC Educational Resources Information Center

    Muller, Jorg M.

    2006-01-01

    A new test index is defined as the probability of obtaining two randomly selected test scores (PDTS) as statistically different. After giving a concept definition of the test index, two simulation studies are presented. The first analyzes the influence of the distribution of test scores, test reliability, and sample size on PDTS within classical…

  14. A complete sample of double-lobed radio quasars for VLBI tests of source models - Definition and statistics

    NASA Technical Reports Server (NTRS)

    Hough, D. H.; Readhead, A. C. S.

    1989-01-01

    A complete, flux-density-limited sample of double-lobed radio quasars is defined, with nuclei bright enough to be mapped with the Mark III VLBI system. It is shown that the statistics of linear size, nuclear strength, and curvature are consistent with the assumption of random source orientations and simple relativistic beaming in the nuclei. However, these statistics are also consistent with the effects of interaction between the beams and the surrounding medium. The distribution of jet velocities in the nuclei, as measured with VLBI, will provide a powerful test of physical theories of extragalactic radio sources.

  15. Spatial Autocorrelation Approaches to Testing Residuals from Least Squares Regression.

    PubMed

    Chen, Yanguang

    2016-01-01

    In geo-statistics, the Durbin-Watson test is frequently employed to detect the presence of residual serial correlation from least squares regression analyses. However, the Durbin-Watson statistic is only suitable for ordered time or spatial series. If the variables comprise cross-sectional data coming from spatial random sampling, the test will be ineffectual because the value of Durbin-Watson's statistic depends on the sequence of data points. This paper develops two new statistics for testing serial correlation of residuals from least squares regression based on spatial samples. By analogy with the new form of Moran's index, an autocorrelation coefficient is defined with a standardized residual vector and a normalized spatial weight matrix. Then by analogy with the Durbin-Watson statistic, two types of new serial correlation indices are constructed. As a case study, the two newly presented statistics are applied to a spatial sample of 29 China's regions. These results show that the new spatial autocorrelation models can be used to test the serial correlation of residuals from regression analysis. In practice, the new statistics can make up for the deficiencies of the Durbin-Watson test.

  16. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhang, R; Bai, W

    Purpose: Because of statistical noise in Monte Carlo dose calculations, effective point doses may not be accurate. Volume spheres are useful for evaluating dose in Monte Carlo plans, which have an inherent statistical uncertainty.We use a user-defined sphere volume instead of a point, take sphere sampling around effective point make the dose statistics to decrease the stochastic errors. Methods: Direct dose measurements were made using a 0.125cc Semiflex ion chamber (IC) 31010 isocentrically placed in the center of a homogeneous Cylindric sliced RW3 phantom (PTW, Germany).In the scanned CT phantom series the sensitive volume length of the IC (6.5mm) weremore » delineated and defined the isocenter as the simulation effective points. All beams were simulated in Monaco in accordance to the measured model. In our simulation using 2mm voxels calculation grid spacing and choose calculate dose to medium and request the relative standard deviation ≤0.5%. Taking three different assigned IC over densities (air electron density(ED) as 0.01g/cm3 default CT scanned ED and Esophageal lumen ED 0.21g/cm3) were tested at different sampling sphere radius (2.5, 2, 1.5 and 1 mm) statistics dose were compared with the measured does. Results: The results show that in the Monaco TPS for the IC using Esophageal lumen ED 0.21g/cm3 and sampling sphere radius 1.5mm the statistical value is the best accordance with the measured value, the absolute average percentage deviation is 0.49%. And when the IC using air electron density(ED) as 0.01g/cm3 and default CT scanned EDthe recommented statistical sampling sphere radius is 2.5mm, the percentage deviation are 0.61% and 0.70%, respectivly. Conclusion: In Monaco treatment planning system for the ionization chamber 31010 recommend air cavity using ED 0.21g/cm3 and sampling 1.5mm sphere volume instead of a point dose to decrease the stochastic errors. Funding Support No.C201505006.« less

  17. Development of parallel line analysis criteria for recombinant adenovirus potency assay and definition of a unit of potency.

    PubMed

    Ogawa, Yasushi; Fawaz, Farah; Reyes, Candice; Lai, Julie; Pungor, Erno

    2007-01-01

    Parameter settings of a parallel line analysis procedure were defined by applying statistical analysis procedures to the absorbance data from a cell-based potency bioassay for a recombinant adenovirus, Adenovirus 5 Fibroblast Growth Factor-4 (Ad5FGF-4). The parallel line analysis was performed with a commercially available software, PLA 1.2. The software performs Dixon outlier test on replicates of the absorbance data, performs linear regression analysis to define linear region of the absorbance data, and tests parallelism between the linear regions of standard and sample. Width of Fiducial limit, expressed as a percent of the measured potency, was developed as a criterion for rejection of the assay data and to significantly improve the reliability of the assay results. With the linear range-finding criteria of the software set to a minimum of 5 consecutive dilutions and best statistical outcome, and in combination with the Fiducial limit width acceptance criterion of <135%, 13% of the assay results were rejected. With these criteria applied, the assay was found to be linear over the range of 0.25 to 4 relative potency units, defined as the potency of the sample normalized to the potency of Ad5FGF-4 standard containing 6 x 10(6) adenovirus particles/mL. The overall precision of the assay was estimated to be 52%. Without the application of Fiducial limit width criterion, the assay results were not linear over the range, and an overall precision of 76% was calculated from the data. An absolute unit of potency for the assay was defined by using the parallel line analysis procedure as the amount of Ad5FGF-4 that results in an absorbance value that is 121% of the average absorbance readings of the wells containing cells not infected with the adenovirus.

  18. A main sequence for quasars

    NASA Astrophysics Data System (ADS)

    Marziani, Paola; Dultzin, Deborah; Sulentic, Jack W.; Del Olmo, Ascensión; Negrete, C. A.; Martínez-Aldama, Mary L.; D'Onofrio, Mauro; Bon, Edi; Bon, Natasa; Stirpe, Giovanna M.

    2018-03-01

    The last 25 years saw a major step forward in the analysis of optical and UV spectroscopic data of large quasar samples. Multivariate statistical approaches have led to the definition of systematic trends in observational properties that are the basis of physical and dynamical modeling of quasar structure. We discuss the empirical correlates of the so-called “main sequence” associated with the quasar Eigenvector 1, its governing physical parameters and several implications on our view of the quasar structure, as well as some luminosity effects associated with the virialized component of the line emitting regions. We also briefly discuss quasars in a segment of the main sequence that includes the strongest FeII emitters. These sources show a small dispersion around a well-defined Eddington ratio value, a property which makes them potential Eddington standard candles.

  19. Numerical consideration on trapping and guiding of nanoparticles in a flow using scattering field of laser light

    NASA Astrophysics Data System (ADS)

    Yokoi, Naomichi; Aizu, Yoshihisa

    2017-04-01

    Optical manipulation techniques proposed so far almost depend on carefully fabricated setups and samples. Similar conditions can be fixed in laboratories, however, it is still a challenging work to manipulate nanoparticles when the environment is not well controlled and is unknown in advance. Nonetheless, coherent light scattered by rough object generates speckles which are random interference patterns with well-defined statistical properties. In the present study, we numerically investigate the motion of a particle in a flow under the illumination of a speckle pattern that is at rest or in motion. Trajectory of the particle is simulated in relation to a flow velocity and a speckle contrast to confirm the feasibility of the present method for performing optical manipulation tasks such as trapping and guiding.

  20. Numerical considerations on control of motion of nanoparticles using scattering field of laser light

    NASA Astrophysics Data System (ADS)

    Yokoi, Naomichi; Aizu, Yoshihisa

    2017-05-01

    Most of optical manipulation techniques proposed so far depend on carefully fabricated setups and samples. Similar conditions can be fixed in laboratories; however, it is still challenging to manipulate nanoparticles when the environment is not well controlled and is unknown in advance. Nonetheless, coherent light scattered by rough object generates a speckle pattern which consists of random interference speckle grains with well-defined statistical properties. In the present study, we numerically investigate the motion of a Brownian particle suspended in water under the illumination of a speckle pattern. Particle-captured time and size of particle-captured area are quantitatively estimated in relation to an optical force and a speckle diameter to confirm the feasibility of the present method for performing optical manipulation tasks such as trapping and guiding.

  1. Landauer-Büttiker and Thouless Conductance

    NASA Astrophysics Data System (ADS)

    Bruneau, L.; Jakšić, V.; Last, Y.; Pillet, C.-A.

    2015-08-01

    In the independent electron approximation, the average (energy/charge/entropy) current flowing through a finite sample connected to two electronic reservoirs can be computed by scattering theoretic arguments which lead to the famous Landauer-Büttiker formula. Another well known formula has been proposed by Thouless on the basis of a scaling argument. The Thouless formula relates the conductance of the sample to the width of the spectral bands of the infinite crystal obtained by periodic juxtaposition of . In this spirit, we define Landauer-Büttiker crystalline currents by extending the Landauer-Büttiker formula to a setup where the sample is replaced by a periodic structure whose unit cell is . We argue that these crystalline currents are closely related to the Thouless currents. For example, the crystalline heat current is bounded above by the Thouless heat current, and this bound saturates iff the coupling between the reservoirs and the sample is reflectionless. Our analysis leads to a rigorous derivation of the Thouless formula from the first principles of quantum statistical mechanics.

  2. Ground-water quality and effects of poultry confined animal feeding operations on shallow ground water, upper Shoal Creek basin, Southwest Missouri, 2000

    USGS Publications Warehouse

    Mugel, Douglas N.

    2002-01-01

    Forty-seven wells and 8 springs were sampled in May, October, and November 2000 in the upper Shoal Creek Basin, southwest Missouri, to determine if nutrient concentrations and fecal bacteria densities are increasing in the shallow aquifer as a result of poultry confined animal feeding operations (CAFOs). Most of the land use in the basin is agricultural, with cattle and hay production dominating; the number of poultry CAFOs has increased in recent years. Poultry waste (litter) is used as a source of nutrients on pasture land as much as several miles away from poultry barns.Most wells in the sample network were classified as ?P? wells, which were open only or mostly to the Springfield Plateau aquifer and where poultry litter was applied to a substantial acreage within 0.5 mile of the well both in spring 2000 and in several previous years; and ?Ag? wells, which were open only or mostly to the Springfield Plateau aquifer and which had limited or no association with poultry CAFOs. Water-quality data from wells and springs were grouped for statistical purposes as P1, Ag1, and Sp1 (May 2000 samples) and P2, Ag2, and Sp2 (October or November 2000 samples). The results of this study do not indicate that poultry CAFOs are affecting the shallow ground water in the upper Shoal Creek Basin with respect to nutrient concentrations and fecal bacteria densities. Statistical tests do not indicate that P wells sampled in spring 2000 have statistically larger concentrations of nitrite plus nitrate or fecal indicator bacteria densities than Ag wells sampled during the same time, at a 95-percent confidence level. Instead, the Ag wells had statistically larger concentrations of nitrite plus nitrate and fecal coliform bacteria densities than the P wells.The results of this study do not indicate seasonal variations from spring 2000 to fall 2000 in the concentrations of nutrients or fecal indicator bacteria densities from well samples. Statistical tests do not indicate statistically significant differences at a 95-percent confidence level for nitrite plus nitrate concentrations or fecal indicator bacteria densities between either P wells sampled in spring and fall 2000, or Ag wells sampled in spring and fall 2000. However, analysis of samples from springs shows that fecal streptococcus bacteria densities were statistically smaller in fall 2000 than in spring 2000 at a 95-percent confidence level.Nitrite plus nitrate concentrations in spring 2000 samples ranged from less than the detection level [0.02 mg/L (milligram per liter) as nitrogen] to 18 mg/L as nitrogen. Seven samples from three wells had nitrite plus nitrate concentrations at or larger than the maximum contaminant level (MCL) of 10 mg/L as nitrogen. The median nitrite plus nitrate concentrations were 0.28 mg/L as nitrogen for P1 samples, 4.6 mg/L as nitrogen for Ag1 samples, and 3.9 mg/L as nitrogen for Sp1 samples.Fecal coliform bacteria were detected in 1 of 25 P1 samples and 5 of 15 Ag1 samples. Escherichia coli (E. coli) bacteria were detected in 3 of 24 P1 samples and 1 of 13 Ag1 samples. Fecal streptococcus bacteria were detected in 8 of 25 P1 samples and 6 of 15 Ag1 samples. Bacteria densities in samples from wells ranged from less than 1 to 81 col/100 mL (colonies per 100 milliliters) of fecal coliform, less than 1 to 140 col/100 mL of E. coli, and less than 1 to 130 col/100 mL of fecal streptococcus. Fecal indicator bacteria densities in samples from springs were substantially larger than in samples from wells. In Sp1 samples, bacteria densities ranged from 12 to 3,300 col/100 mL of fecal coliform, 40 to 2,700 col/100 mL of E. coli, and 42 to 3,100 col/100 mL of fecal streptococcus.

  3. Sample Size and Statistical Conclusions from Tests of Fit to the Rasch Model According to the Rasch Unidimensional Measurement Model (Rumm) Program in Health Outcome Measurement.

    PubMed

    Hagell, Peter; Westergren, Albert

    Sample size is a major factor in statistical null hypothesis testing, which is the basis for many approaches to testing Rasch model fit. Few sample size recommendations for testing fit to the Rasch model concern the Rasch Unidimensional Measurement Models (RUMM) software, which features chi-square and ANOVA/F-ratio based fit statistics, including Bonferroni and algebraic sample size adjustments. This paper explores the occurrence of Type I errors with RUMM fit statistics, and the effects of algebraic sample size adjustments. Data with simulated Rasch model fitting 25-item dichotomous scales and sample sizes ranging from N = 50 to N = 2500 were analysed with and without algebraically adjusted sample sizes. Results suggest the occurrence of Type I errors with N less then or equal to 500, and that Bonferroni correction as well as downward algebraic sample size adjustment are useful to avoid such errors, whereas upward adjustment of smaller samples falsely signal misfit. Our observations suggest that sample sizes around N = 250 to N = 500 may provide a good balance for the statistical interpretation of the RUMM fit statistics studied here with respect to Type I errors and under the assumption of Rasch model fit within the examined frame of reference (i.e., about 25 item parameters well targeted to the sample).

  4. Approaches for estimating minimal clinically important differences in systemic lupus erythematosus.

    PubMed

    Rai, Sharan K; Yazdany, Jinoos; Fortin, Paul R; Aviña-Zubieta, J Antonio

    2015-06-03

    A minimal clinically important difference (MCID) is an important concept used to determine whether a medical intervention improves perceived outcomes in patients. Prior to the introduction of the concept in 1989, studies focused primarily on statistical significance. As most recent clinical trials in systemic lupus erythematosus (SLE) have failed to show significant effects, determining a clinically relevant threshold for outcome scores (that is, the MCID) of existing instruments may be critical for conducting and interpreting meaningful clinical trials as well as for facilitating the establishment of treatment recommendations for patients. To that effect, methods to determine the MCID can be divided into two well-defined categories: distribution-based and anchor-based approaches. Distribution-based approaches are based on statistical characteristics of the obtained samples. There are various methods within the distribution-based approach, including the standard error of measurement, the standard deviation, the effect size, the minimal detectable change, the reliable change index, and the standardized response mean. Anchor-based approaches compare the change in a patient-reported outcome to a second, external measure of change (that is, one that is more clearly understood, such as a global assessment), which serves as the anchor. Finally, the Delphi technique can be applied as an adjunct to defining a clinically important difference. Despite an abundance of methods reported in the literature, little work in MCID estimation has been done in the context of SLE. As the MCID can help determine the effect of a given therapy on a patient and add meaning to statistical inferences made in clinical research, we believe there ought to be renewed focus on this area. Here, we provide an update on the use of MCIDs in clinical research, review some of the work done in this area in SLE, and propose an agenda for future research.

  5. On the spectrum of inhomogeneous turbulence

    NASA Technical Reports Server (NTRS)

    Trevino, G.

    1979-01-01

    Inhomogeneous turbulence is defined as turbulence whose statistics are functions of spatial position. The turbulence spectrum, and particularly how the shape of the spectrum varies, from point to point in space, as a consequence of well defined spatial variations in the turbulence intensity and/or integral scale is investigated.

  6. Social Indicators and Social Forecasting.

    ERIC Educational Resources Information Center

    Johnston, Denis F.

    The paper identifies major types of social indicators and explains how they can be used in social forecasting. Social indicators are defined as statistical measures relating to major areas of social concern and/or individual well being. Examples of social indicators are projections, forecasts, outlook statements, time-series statistics, and…

  7. Thermal heterogeneity within aqueous materials quantified by 1H NMR spectroscopy: Multiparametric validation in silico and in vitro

    NASA Astrophysics Data System (ADS)

    Lutz, Norbert W.; Bernard, Monique

    2018-02-01

    We recently suggested a new paradigm for statistical analysis of thermal heterogeneity in (semi-)aqueous materials by 1H NMR spectroscopy, using water as a temperature probe. Here, we present a comprehensive in silico and in vitro validation that demonstrates the ability of this new technique to provide accurate quantitative parameters characterizing the statistical distribution of temperature values in a volume of (semi-)aqueous matter. First, line shape parameters of numerically simulated water 1H NMR spectra are systematically varied to study a range of mathematically well-defined temperature distributions. Then, corresponding models based on measured 1H NMR spectra of agarose gel are analyzed. In addition, dedicated samples based on hydrogels or biological tissue are designed to produce temperature gradients changing over time, and dynamic NMR spectroscopy is employed to analyze the resulting temperature profiles at sub-second temporal resolution. Accuracy and consistency of the previously introduced statistical descriptors of temperature heterogeneity are determined: weighted median and mean temperature, standard deviation, temperature range, temperature mode(s), kurtosis, skewness, entropy, and relative areas under temperature curves. Potential and limitations of this method for quantitative analysis of thermal heterogeneity in (semi-)aqueous materials are discussed in view of prospective applications in materials science as well as biology and medicine.

  8. Validity of strong lensing statistics for constraints on the galaxy evolution model

    NASA Astrophysics Data System (ADS)

    Matsumoto, Akiko; Futamase, Toshifumi

    2008-02-01

    We examine the usefulness of the strong lensing statistics to constrain the evolution of the number density of lensing galaxies by adopting the values of the cosmological parameters determined by recent Wilkinson Microwave Anisotropy Probe observation. For this purpose, we employ the lens-redshift test proposed by Kochanek and constrain the parameters in two evolution models, simple power-law model characterized by the power-law indexes νn and νv, and the evolution model by Mitchell et al. based on cold dark matter structure formation scenario. We use the well-defined lens sample from the Sloan Digital Sky Survey (SDSS) and this is similarly sized samples used in the previous studies. Furthermore, we adopt the velocity dispersion function of early-type galaxies based on SDSS DR1 and DR5. It turns out that the indexes of power-law model are consistent with the previous studies, thus our results indicate the mild evolution in the number and velocity dispersion of early-type galaxies out to z = 1. However, we found that the values for p and q used by Mitchell et al. are inconsistent with the presently available observational data. More complete sample is necessary to withdraw more realistic determination on these parameters.

  9. Photometric Properties of Face-on Isolated Spiral Galaxies

    NASA Astrophysics Data System (ADS)

    Bahr, Alexander; Epstein, P.; Durbala, A.

    2011-05-01

    We want to quantify the relative role of nature versus nurture in defining the observed properties of galaxies. In simpler terms we would like to disentangle the ``genetic'’ and the environmental influences in shaping the morphology of galaxies. In order to do that one needs to firstly define a zero-order baseline, i.e., a sample of galaxies that have been minimally perturbed by neighbors in the last few billion years of their existence. Such a sample has been produced and refined in different stages in the context of the AMIGA international project (www.iaa.es/AMIGA.html). The recent catalogue ``The All-Sky Catalog of Isolated Galaxies Selected from 2MASS'’ (Karachentseva, V. E. et al. 2010) allows us to complete and enrich the initial sample constructed within AMIGA with new objects, thus enhancing the statistical relevance of our study. Our focus is to define a subset of isolated disk spiral galaxies. We constrain the sample selection by: 1) orientation, restricting to almost face-on galaxies and 2) availability of good photometric images in SDSS. The goal is to ``dissect'’ (decompose) these galaxies in major components (disk, bulge, bars, etc.) and to study the properties of the components in a statistical context. Having a reasonable representation of all morphological types, we aim to test the bimodality of bulges and bars. We present a progress report of our work.

  10. Stokes-correlometry of polarization-inhomogeneous objects

    NASA Astrophysics Data System (ADS)

    Ushenko, O. G.; Dubolazov, A.; Bodnar, G. B.; Bachynskiy, V. T.; Vanchulyak, O.

    2018-01-01

    The paper consists of two parts. The first part presents short theoretical basics of the method of Stokes-correlometry description of optical anisotropy of biological tissues. It was provided experimentally measured coordinate distributions of modulus (MSV) and phase (PhSV) of complex Stokes vector of skeletal muscle tissue. It was defined the values and ranges of changes of statistic moments of the 1st-4th orders, which characterize the distributions of values of MSV and PhSV. The second part presents the data of statistic analysis of the distributions of modulus MSV and PhSV. It was defined the objective criteria of differentiation of samples with urinary incontinence.

  11. Sensation seeking and impulsive traits as personality endophenotypes for antisocial behavior: Evidence from two independent samples

    PubMed Central

    Mann, Frank D.; Engelhardt, Laura; Briley, Daniel A.; Grotzinger, Andrew D.; Patterson, Megan W.; Tackett, Jennifer L.; Strathan, Dixie B.; Heath, Andrew; Lynskey, Michael; Slutske, Wendy; Martin, Nicholas G.; Tucker-Drob, Elliot M.; Harden, K. Paige

    2017-01-01

    Sensation seeking and impulsivity are personality traits that are correlated with risk for antisocial behavior (ASB). This paper uses two independent samples of twins to (a) test the extent to which sensation seeking and impulsivity statistically mediate genetic influence on ASB, and (b) compare this to genetic influences accounted for by other personality traits. In Sample 1, delinquent behavior, as well as impulsivity, sensation seeking and Big Five personality traits, were measured in adolescent twins from the Texas Twin Project. In Sample 2, adult twins from the Australian Twin Registry responded to questionnaires that assessed individual differences in Eysenck's and Cloninger's personality dimensions, and a structured telephone interview that asked participants to retrospectively report DSM-defined symptoms of conduct disorder. Bivariate quantitative genetic models were used to identify genetic overlap between personality traits and ASB. Across both samples, novelty/sensation seeking and impulsive traits accounted for larger portions of genetic variance in ASB than other personality traits. We discuss whether sensation seeking and impulsive personality are causal endophenotypes for ASB, or merely index genetic liability for ASB. PMID:28824215

  12. Experimental Evaluation of Preservation Techniques for Benzene, Toluene, Ethylbenzene, and Total Xylenes in Water Samples.

    PubMed

    Arnold, Ray; Kong, Deyuan; Douglas, Gregory; Hardenstine, Jeffery; Rouhani, Shahrokh; Gala, William

    2018-01-01

    An experiment was designed to address the validity of the prescribed maximum allowable holding-time limit of 14 days when acidified at < 2 pH and maintained at 4°C to prevent significant loss of benzene, toluene, ethyl benzene, and xylenes (BTEX) in preserved water samples. Preservation methods prescribed by the United State Environmental Protection Agency were used as well as adaptions of that procedure to determine stability between 3 and 21 days. Water samples preserved at 4°C and pH of < 2 with hydrochloric acid did not result in unacceptable (> 15%) BTEX losses during the study as defined by procedures and statistical methods described by the American Society for Testing and Materials International. In addition, water samples preserved only with acid (pH < 2) at ambient temperatures (20-27°C) also provided acceptable results during the 21-day study. These results have demonstrated the acceptability of BTEX data derived from water samples exceeding the standard holding-time and/or temperature limits.

  13. Evaluation of a segment-based LANDSAT full-frame approach to corp area estimation

    NASA Technical Reports Server (NTRS)

    Bauer, M. E. (Principal Investigator); Hixson, M. M.; Davis, S. M.

    1981-01-01

    As the registration of LANDSAT full frames enters the realm of current technology, sampling methods should be examined which utilize other than the segment data used for LACIE. The effect of separating the functions of sampling for training and sampling for area estimation. The frame selected for analysis was acquired over north central Iowa on August 9, 1978. A stratification of he full-frame was defined. Training data came from segments within the frame. Two classification and estimation procedures were compared: statistics developed on one segment were used to classify that segment, and pooled statistics from the segments were used to classify a systematic sample of pixels. Comparisons to USDA/ESCS estimates illustrate that the full-frame sampling approach can provide accurate and precise area estimates.

  14. Experimental Design in Clinical 'Omics Biomarker Discovery.

    PubMed

    Forshed, Jenny

    2017-11-03

    This tutorial highlights some issues in the experimental design of clinical 'omics biomarker discovery, how to avoid bias and get as true quantities as possible from biochemical analyses, and how to select samples to improve the chance of answering the clinical question at issue. This includes the importance of defining clinical aim and end point, knowing the variability in the results, randomization of samples, sample size, statistical power, and how to avoid confounding factors by including clinical data in the sample selection, that is, how to avoid unpleasant surprises at the point of statistical analysis. The aim of this Tutorial is to help translational clinical and preclinical biomarker candidate research and to improve the validity and potential of future biomarker candidate findings.

  15. Spatial Autocorrelation Approaches to Testing Residuals from Least Squares Regression

    PubMed Central

    Chen, Yanguang

    2016-01-01

    In geo-statistics, the Durbin-Watson test is frequently employed to detect the presence of residual serial correlation from least squares regression analyses. However, the Durbin-Watson statistic is only suitable for ordered time or spatial series. If the variables comprise cross-sectional data coming from spatial random sampling, the test will be ineffectual because the value of Durbin-Watson’s statistic depends on the sequence of data points. This paper develops two new statistics for testing serial correlation of residuals from least squares regression based on spatial samples. By analogy with the new form of Moran’s index, an autocorrelation coefficient is defined with a standardized residual vector and a normalized spatial weight matrix. Then by analogy with the Durbin-Watson statistic, two types of new serial correlation indices are constructed. As a case study, the two newly presented statistics are applied to a spatial sample of 29 China’s regions. These results show that the new spatial autocorrelation models can be used to test the serial correlation of residuals from regression analysis. In practice, the new statistics can make up for the deficiencies of the Durbin-Watson test. PMID:26800271

  16. VizieR Online Data Catalog: The CLASS BL Lac sample (Marcha+, 2013)

    NASA Astrophysics Data System (ADS)

    Marcha, M. J. M.; Caccianiga, A.

    2014-04-01

    This paper presents a new sample of BL Lac objects selected from a deep (30mJy) radio survey of flat spectrum radio sources (the CLASS blazar survey). The sample is one of the largest well-defined samples in the low-power regime with a total of 130 sources of which 55 satisfy the 'classical' optical BL Lac selection criteria, and the rest have indistinguishable radio properties. The primary goal of this study is to establish the radio luminosity function (RLF) on firm statistical ground at low radio luminosities where previous samples have not been able to investigate. The gain of taking a peek at lower powers is the possibility to search for the flattening of the luminosity function which is a feature predicted by the beaming model but which has remained elusive to observational confirmation. In this study, we extend for the first time the BL Lac RLF down to very low radio powers ~1022W/Hz, i.e. two orders of magnitude below the RLF currently available in the literature. In the process, we confirm the importance of adopting a broader, and more physically meaningful set of classification criteria to avoid the systematic missing of low-luminosity BL Lacs. Thanks to the good statistics we confirm the existence of weak but significant positive cosmological evolution for the BL Lac population, and we detect, for the first time the flattening of the RLF at L~1025W/Hz in agreement with the predictions of the beaming model. (1 data file).

  17. An investigative comparison of purging and non-purging groundwater sampling methods in Karoo aquifer monitoring wells

    NASA Astrophysics Data System (ADS)

    Gomo, M.; Vermeulen, D.

    2015-03-01

    An investigation was conducted to statistically compare the influence of non-purging and purging groundwater sampling methods on analysed inorganic chemistry parameters and calculated saturation indices. Groundwater samples were collected from 15 monitoring wells drilled in Karoo aquifers before and after purging for the comparative study. For the non-purging method, samples were collected from groundwater flow zones located in the wells using electrical conductivity (EC) profiling. The two data sets of non-purged and purged groundwater samples were analysed for inorganic chemistry parameters at the Institute of Groundwater Studies (IGS) laboratory of the Free University in South Africa. Saturation indices for mineral phases that were found in the data base of PHREEQC hydrogeochemical model were calculated for each data set. Four one-way ANOVA tests were conducted using Microsoft excel 2007 to investigate if there is any statistically significant difference between: (1) all inorganic chemistry parameters measured in the non-purged and purged groundwater samples per each specific well, (2) all mineral saturation indices calculated for the non-purged and purged groundwater samples per each specific well, (3) individual inorganic chemistry parameters measured in the non-purged and purged groundwater samples across all wells and (4) Individual mineral saturation indices calculated for non-purged and purged groundwater samples across all wells. For all the ANOVA tests conducted, the calculated alpha values (p) are greater than 0.05 (significance level) and test statistic (F) is less than the critical value (Fcrit) (F < Fcrit). The results imply that there was no statistically significant difference between the two data sets. With a 95% confidence, it was therefore concluded that the variance between groups was rather due to random chance and not to the influence of the sampling methods (tested factor). It is therefore be possible that in some hydrogeologic conditions, non-purged groundwater samples might be just as representative as the purged ones. The findings of this study can provide an important platform for future evidence oriented research investigations to establish the necessity of purging prior to groundwater sampling in different aquifer systems.

  18. Statistical analysis of arsenic contamination in drinking water in a city of Iran and its modeling using GIS.

    PubMed

    Sadeghi, Fatemeh; Nasseri, Simin; Mosaferi, Mohammad; Nabizadeh, Ramin; Yunesian, Masud; Mesdaghinia, Alireza

    2017-05-01

    In this research, probable arsenic contamination in drinking water in the city of Ardabil was studied in 163 samples during four seasons. In each season, sampling was carried out randomly in the study area. Results were analyzed statistically applying SPSS 19 software, and the data was also modeled by Arc GIS 10.1 software. The maximum permissible arsenic concentration in drinking water defined by the World Health Organization and Iranian national standard is 10 μg/L. Statistical analysis showed 75, 88, 47, and 69% of samples in autumn, winter, spring, and summer, respectively, had concentrations higher than the national standard. The mean concentrations of arsenic in autumn, winter, spring, and summer were 19.89, 15.9, 10.87, and 14.6 μg/L, respectively, and the overall average in all samples through the year was 15.32 μg/L. Although GIS outputs indicated that the concentration distribution profiles changed in four consecutive seasons, variance analysis of the results showed that statistically there is no significant difference in arsenic levels in four seasons.

  19. Environmental assessment of Al-Hammar Marsh, Southern Iraq.

    PubMed

    Al-Gburi, Hind Fadhil Abdullah; Al-Tawash, Balsam Salim; Al-Lafta, Hadi Salim

    2017-02-01

    (a) To determine the spatial distributions and levels of major and minor elements, as well as heavy metals, in water, sediment, and biota (plant and fish) in Al-Hammar Marsh, southern Iraq, and ultimately to supply more comprehensive information for policy-makers to manage the contaminants input into the marsh so that their concentrations do not reach toxic levels. (b) to characterize the seasonal changes in the marsh surface water quality. (c) to address the potential environmental risk of these elements by comparison with the historical levels and global quality guidelines (i.e., World Health Organization (WHO) standard limits). (d) to define the sources of these elements (i.e., natural and/or anthropogenic) using combined multivariate statistical techniques such as Principal Component Analysis (PCA) and Agglomerative Hierarchical Cluster Analysis (AHCA) along with pollution analysis (i.e., enrichment factor analysis). Water, sediment, plant, and fish samples were collected from the marsh, and analyzed for major and minor ions, as well as heavy metals, and then compared to historical levels and global quality guidelines (WHO guidelines). Then, multivariate statistical techniques, such as PCA and AHCA, were used to determine the element sourcing. Water analyses revealed unacceptable values for almost all physio-chemical and biological properties, according to WHO standard limits for drinking water. Almost all major ions and heavy metal concentrations in water showed a distinct decreasing trend at the marsh outlet station compared to other stations. In general, major and minor ions, as well as heavy metals exhibit higher concentrations in winter than in summer. Sediment analyses using multivariate statistical techniques revealed that Mg, Fe, S, P, V, Zn, As, Se, Mo, Co, Ni, Cu, Sr, Br, Cd, Ca, N, Mn, Cr, and Pb were derived from anthropogenic sources, while Al, Si, Ti, K, and Zr were primarily derived from natural sources. Enrichment factor analysis gave results compatible with multivariate statistical techniques findings. Analysis of heavy metals in plant samples revealed that there is no pollution in plants in Al-Hammar Marsh. However, the concentrations of heavy metals in fish samples showed that all samples were contaminated by Pb, Mn, and Ni, while some samples were contaminated by Pb, Mn, and Ni. Decreasing of Tigris and Euphrates discharges during the past decades due to drought conditions and upstream damming, as well as the increasing stress of wastewater effluents from anthropogenic activities, led to degradation of the downstream Al-Hammar Marsh water quality in terms of physical, chemical, and biological properties. As such properties were found to consistently exceed the historical and global quality objectives. However, element concentration decreasing trend at the marsh outlet station compared to other stations indicate that the marsh plays an important role as a natural filtration and bioremediation system. Higher element concentrations in winter were due to runoff from the washing of the surrounding Sabkha during flooding by winter rainstorms. Finally, the high concentrations of heavy metals in fish samples can be attributed to bioaccumulation and biomagnification processes.

  20. Visual Sample Plan Version 7.0 User's Guide

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Matzke, Brett D.; Newburn, Lisa LN; Hathaway, John E.

    2014-03-01

    User's guide for VSP 7.0 This user's guide describes Visual Sample Plan (VSP) Version 7.0 and provides instructions for using the software. VSP selects the appropriate number and location of environmental samples to ensure that the results of statistical tests performed to provide input to risk decisions have the required confidence and performance. VSP Version 7.0 provides sample-size equations or algorithms needed by specific statistical tests appropriate for specific environmental sampling objectives. It also provides data quality assessment and statistical analysis functions to support evaluation of the data and determine whether the data support decisions regarding sites suspected of contamination.more » The easy-to-use program is highly visual and graphic. VSP runs on personal computers with Microsoft Windows operating systems (XP, Vista, Windows 7, and Windows 8). Designed primarily for project managers and users without expertise in statistics, VSP is applicable to two- and three-dimensional populations to be sampled (e.g., rooms and buildings, surface soil, a defined layer of subsurface soil, water bodies, and other similar applications) for studies of environmental quality. VSP is also applicable for designing sampling plans for assessing chem/rad/bio threat and hazard identification within rooms and buildings, and for designing geophysical surveys for unexploded ordnance (UXO) identification.« less

  1. Can serums be replaced by Mueller-Hinton agar in germ tube test?

    PubMed

    Atalay, M A; Koc, A N; Parkan, O M; Aydemir, G; Elmali, F; Sav, H

    2017-01-01

    The germ tube test (GTT) is inexpensive, easy, and well-defined test that differentiates Candida albicans (excluding Candida dubliniensis and Candida africana) from other species. The aim of this study was to evaluate various serums (i.e., human, rabbit, horse, and fetal bovine serum) used in the GTT and Mueller-Hinton agar (MHA). Fifty species isolated from various clinical samples that were defined as C. albicans by both conventional and DNA sequence analysis methods were included in the study. One to two colonies of C. albicans were mixed into 0.5-1 ml of fetal bovine serum, horse serum, rabbit serum, and human serum. Serums and MHA were incubated at 37°C for GTT. They were removed from the incubator and evaluated after 30 min, 1 h, 2 h, and 3 h of incubation. The GTT was accepted to be positive only if germ tube was 1/2 the width and 3 times the length of the parent yeast cell and with no constriction at the point of origin. When the use of serums and MHA for GTT was statistically evaluated, according to the positive scoring, the best results were obtained with MHA and with rabbit, horse, and fetal bovine serum, respectively. The best definition over time statistically was the third hour. It is suggested that inexpensive MHA is a fast, appropriate, and reliable medium for the probable diagnosis of GTT and C. albicans; however, additional studies are still needed to define other Candida species.

  2. CAN'T MISS--conquer any number task by making important statistics simple. Part 2. Probability, populations, samples, and normal distributions.

    PubMed

    Hansen, John P

    2003-01-01

    Healthcare quality improvement professionals need to understand and use inferential statistics to interpret sample data from their organizations. In quality improvement and healthcare research studies all the data from a population often are not available, so investigators take samples and make inferences about the population by using inferential statistics. This three-part series will give readers an understanding of the concepts of inferential statistics as well as the specific tools for calculating confidence intervals for samples of data. This article, Part 2, describes probability, populations, and samples. The uses of descriptive and inferential statistics are outlined. The article also discusses the properties and probability of normal distributions, including the standard normal distribution.

  3. An Extension of the EDGES Survey: Stellar Populations in Dark Matter Halos

    NASA Astrophysics Data System (ADS)

    van Zee, Liese

    The formation and evolution of galactic disks is one of the key questions in extragalactic astronomy today. We plan to use archival data from GALEX, Spitzer, and WISE to investigate the growth and evolution of the stellar component in a statistical sample of nearby galaxies. Data covering a broad wavelength range are critical for measurement of current star formation activity, stellar populations, and stellar distributions in nearby galaxies. In order to investigate the timescales associated with the growth of galactic disks, we will (1) investigate the structure of the underlying stellar distribution, (2) measure the ratio of current-to-past star formation activity as a function of radius, and (3) investigate the growth of the stellar disk as a function of baryon fraction and total dynamical mass. The proposed projects leverage the existing deep wide field-of-view near infrared imaging observations obtained with the Spitzer Space Telescope as part of the EDGES Survey, a Cycle 8 Exploration Science Program. The proposed analysis of multiwavelength imaging observations of a well-defined statistical sample will place strong constraints on hierarchical models of galaxy formation and evolution and will further our understanding of the stellar component of nearby galaxies.

  4. Audience Diversion Due to Cable Television: A Statistical Analysis of New Data.

    ERIC Educational Resources Information Center

    Park, Rolla Edward

    A statistical analysis of new data suggests that television broadcasting will continue to prosper, despite increasing competition from cable television carrying distant signals. Data on cable and non-cable audiences in 121 counties with well defined signal choice support generalized least squares estimates of two models: total audience and…

  5. Military Representation: The Theoretical and Practical Implications of Population Representation in the American Armed Forces

    DTIC Science & Technology

    1979-10-01

    racism " even before the Vietnam casualty statistics received attention in the national news media. In...409 In theory , then, a highly unrepresentative (in statistical terms) force could be an "approximately representative" force. Depending on the balance...of Army representation." The six-month project appeared at the outset to be a well-defined, strictly "objective," statistical evaluation of

  6. Framework for making better predictions by directly estimating variables' predictivity.

    PubMed

    Lo, Adeline; Chernoff, Herman; Zheng, Tian; Lo, Shaw-Hwa

    2016-12-13

    We propose approaching prediction from a framework grounded in the theoretical correct prediction rate of a variable set as a parameter of interest. This framework allows us to define a measure of predictivity that enables assessing variable sets for, preferably high, predictivity. We first define the prediction rate for a variable set and consider, and ultimately reject, the naive estimator, a statistic based on the observed sample data, due to its inflated bias for moderate sample size and its sensitivity to noisy useless variables. We demonstrate that the [Formula: see text]-score of the PR method of VS yields a relatively unbiased estimate of a parameter that is not sensitive to noisy variables and is a lower bound to the parameter of interest. Thus, the PR method using the [Formula: see text]-score provides an effective approach to selecting highly predictive variables. We offer simulations and an application of the [Formula: see text]-score on real data to demonstrate the statistic's predictive performance on sample data. We conjecture that using the partition retention and [Formula: see text]-score can aid in finding variable sets with promising prediction rates; however, further research in the avenue of sample-based measures of predictivity is much desired.

  7. Stochastic inference with spiking neurons in the high-conductance state

    NASA Astrophysics Data System (ADS)

    Petrovici, Mihai A.; Bill, Johannes; Bytschok, Ilja; Schemmel, Johannes; Meier, Karlheinz

    2016-10-01

    The highly variable dynamics of neocortical circuits observed in vivo have been hypothesized to represent a signature of ongoing stochastic inference but stand in apparent contrast to the deterministic response of neurons measured in vitro. Based on a propagation of the membrane autocorrelation across spike bursts, we provide an analytical derivation of the neural activation function that holds for a large parameter space, including the high-conductance state. On this basis, we show how an ensemble of leaky integrate-and-fire neurons with conductance-based synapses embedded in a spiking environment can attain the correct firing statistics for sampling from a well-defined target distribution. For recurrent networks, we examine convergence toward stationarity in computer simulations and demonstrate sample-based Bayesian inference in a mixed graphical model. This points to a new computational role of high-conductance states and establishes a rigorous link between deterministic neuron models and functional stochastic dynamics on the network level.

  8. EG-09EPIGENETIC PROFILING REVEALS A CpG HYPERMETHYLATION PHENOTYPE (CIMP) ASSOCIATED WITH WORSE PROGRESSION-FREE SURVIVAL IN MENINGIOMA

    PubMed Central

    Olar, Adriana; Wani, Khalida; Mansouri, Alireza; Zadeh, Gelareh; Wilson, Charmaine; DeMonte, Franco; Fuller, Gregory; Jones, David; Pfister, Stefan; von Deimling, Andreas; Sulman, Erik; Aldape, Kenneth

    2014-01-01

    BACKGROUND: Methylation profiling of solid tumors has revealed biologic subtypes, often with clinical implications. Methylation profiles of meningioma and their clinical implications are not well understood. METHODS: Ninety-two meningioma samples (n = 44 test set and n = 48 validation set) were profiled using the Illumina HumanMethylation450 BeadChip. Unsupervised clustering and analyses for recurrence-free survival (RFS) were performed. RESULTS: Unsupervised clustering of the test set using approximately 900 highly variable markers identified two clearly defined methylation subgroups. One of the groups (n = 19) showed global hypermethylation of a set of markers, analogous to CpG island methylator phenotype (CIMP). These findings were reproducible in the validation set, with 18/48 samples showing the CIMP-positive phenotype. Importantly, of 347 highly variable markers common to both the test and validation set analyses, 107 defined CIMP in the test set and 94 defined CIMP in the validation set, with an overlap of 83 markers between the two datasets. This number is much greater than expected by chance indicating reproducibly of the hypermethylated markers that define CIMP in meningioma. With respect to clinical correlation, the 37 CIMP-positive cases displayed significantly shorter RFS compared to the 55 non-CIMP cases (hazard ratio 2.9, p = 0.013). In an effort to develop a preliminary outcome predictor, a 155-marker subset correlated with RFS was identified in the test dataset. When interrogated in the validation dataset, this 155-marker subset showed a statistical trend (p < 0.1) towards distinguishing survival groups. CONCLUSIONS: This study defines the existence of a CIMP phenotype in meningioma, which involves a substantial proportion (37/92, 40%) of samples with clinical implications. Ongoing work will expand this cohort and examine identification of additional biologic differences (mutational and DNA copy number analysis) to further characterize the aberrant methylation subtype in meningioma. CIMP-positivity with aberrant methylation in recurrent/malignant meningioma suggests a potential therapeutic target for clinically aggressive cases.

  9. Empirical Testing of an Algorithm for Defining Somatization in Children

    PubMed Central

    Eisman, Howard D.; Fogel, Joshua; Lazarovich, Regina; Pustilnik, Inna

    2007-01-01

    Introduction A previous article proposed an algorithm for defining somatization in children by classifying them into three categories: well, medically ill, and somatizer; the authors suggested further empirical validation of the algorithm (Postilnik et al., 2006). We use the Child Behavior Checklist (CBCL) to provide this empirical validation. Method Parents of children seen in pediatric clinics completed the CBCL (n=126). The physicians of these children completed specially-designed questionnaires. The sample comprised of 62 boys and 64 girls (age range 2 to 15 years). Classification categories included: well (n=53), medically ill (n=55), and somatizer (n=18). Analysis of variance (ANOVA) was used for statistical comparisons. Discriminant function analysis was conducted with the CBCL subscales. Results There were significant differences between the classification categories for the somatic complaints (p=<0.001), social problems (p=0.004), thought problems (p=0.01), attention problems (0.006), and internalizing (p=0.003) subscales and also total (p=0.001), and total-t (p=0.001) scales of the CBCL. Discriminant function analysis showed that 78% of somatizers and 66% of well were accurately classified, while only 35% of medically ill were accurately classified. Conclusion The somatization classification algorithm proposed by Postilnik et al. (2006) shows promise for classification of children and adolescents with somatic symptoms. PMID:18421368

  10. An Analytic Solution to the Computation of Power and Sample Size for Genetic Association Studies under a Pleiotropic Mode of Inheritance.

    PubMed

    Gordon, Derek; Londono, Douglas; Patel, Payal; Kim, Wonkuk; Finch, Stephen J; Heiman, Gary A

    2016-01-01

    Our motivation here is to calculate the power of 3 statistical tests used when there are genetic traits that operate under a pleiotropic mode of inheritance and when qualitative phenotypes are defined by use of thresholds for the multiple quantitative phenotypes. Specifically, we formulate a multivariate function that provides the probability that an individual has a vector of specific quantitative trait values conditional on having a risk locus genotype, and we apply thresholds to define qualitative phenotypes (affected, unaffected) and compute penetrances and conditional genotype frequencies based on the multivariate function. We extend the analytic power and minimum-sample-size-necessary (MSSN) formulas for 2 categorical data-based tests (genotype, linear trend test [LTT]) of genetic association to the pleiotropic model. We further compare the MSSN of the genotype test and the LTT with that of a multivariate ANOVA (Pillai). We approximate the MSSN for statistics by linear models using a factorial design and ANOVA. With ANOVA decomposition, we determine which factors most significantly change the power/MSSN for all statistics. Finally, we determine which test statistics have the smallest MSSN. In this work, MSSN calculations are for 2 traits (bivariate distributions) only (for illustrative purposes). We note that the calculations may be extended to address any number of traits. Our key findings are that the genotype test usually has lower MSSN requirements than the LTT. More inclusive thresholds (top/bottom 25% vs. top/bottom 10%) have higher sample size requirements. The Pillai test has a much larger MSSN than both the genotype test and the LTT, as a result of sample selection. With these formulas, researchers can specify how many subjects they must collect to localize genes for pleiotropic phenotypes. © 2017 S. Karger AG, Basel.

  11. Psychiatric disorders in preschoolers: the structure of DSM-IV symptoms and profiles of comorbidity.

    PubMed

    Wichstrøm, Lars; Berg-Nielsen, Turid Suzanne

    2014-07-01

    Psychiatric disorders have been increasingly recognized in preschool children; at present, however, we know comparatively less about how well current diagnostic manuals capture the symptoms described in this age group and how comorbidity is patterned. Therefore, this study aimed to investigate whether the symptoms defined by the Diagnostic and Statistical Manual of Mental Disorders, fourth edition (DSM-IV) load on their respective disorders, examine whether individual symptoms exist that load particularly high or low on the disorder they allegedly define, and analyze how comorbidity clusters in individual children. Parents of a community sample of Norwegian 4-year-olds (N = 995) were interviewed using the Preschool Age Psychiatric Assessment. A confirmatory factor analysis (CFA) and a latent profile analysis (LPA) were performed on the symptoms of seven DSM disorders: attention-deficit/hyperactivity disorder (ADHD), oppositional defiant disorder, conduct disorder, major depressive disorder (MDD), generalized anxiety disorder (GAD), social phobia, and separation anxiety disorder. The results showed that the CFA solution that closely resembled the disorders delineated in the DSM-IV fitted the data best. However, vegetative symptoms did not define preschool depression. The LPA identified nine symptom profiles among preschoolers, of which four showed evidence of psychopathology: comorbid MDD/GAD ? ADHD combined type, comorbid MDD/GAD ? ADHD hyperactive/impulsive type, separation anxiety only, and social phobia only. In conclusion, the symptoms observed in preschoolers fit the DSM-IV well, and comorbidity followed specific patterns.

  12. Relative efficiency and sample size for cluster randomized trials with variable cluster sizes.

    PubMed

    You, Zhiying; Williams, O Dale; Aban, Inmaculada; Kabagambe, Edmond Kato; Tiwari, Hemant K; Cutter, Gary

    2011-02-01

    The statistical power of cluster randomized trials depends on two sample size components, the number of clusters per group and the numbers of individuals within clusters (cluster size). Variable cluster sizes are common and this variation alone may have significant impact on study power. Previous approaches have taken this into account by either adjusting total sample size using a designated design effect or adjusting the number of clusters according to an assessment of the relative efficiency of unequal versus equal cluster sizes. This article defines a relative efficiency of unequal versus equal cluster sizes using noncentrality parameters, investigates properties of this measure, and proposes an approach for adjusting the required sample size accordingly. We focus on comparing two groups with normally distributed outcomes using t-test, and use the noncentrality parameter to define the relative efficiency of unequal versus equal cluster sizes and show that statistical power depends only on this parameter for a given number of clusters. We calculate the sample size required for an unequal cluster sizes trial to have the same power as one with equal cluster sizes. Relative efficiency based on the noncentrality parameter is straightforward to calculate and easy to interpret. It connects the required mean cluster size directly to the required sample size with equal cluster sizes. Consequently, our approach first determines the sample size requirements with equal cluster sizes for a pre-specified study power and then calculates the required mean cluster size while keeping the number of clusters unchanged. Our approach allows adjustment in mean cluster size alone or simultaneous adjustment in mean cluster size and number of clusters, and is a flexible alternative to and a useful complement to existing methods. Comparison indicated that we have defined a relative efficiency that is greater than the relative efficiency in the literature under some conditions. Our measure of relative efficiency might be less than the measure in the literature under some conditions, underestimating the relative efficiency. The relative efficiency of unequal versus equal cluster sizes defined using the noncentrality parameter suggests a sample size approach that is a flexible alternative and a useful complement to existing methods.

  13. Large ensemble modeling of the last deglacial retreat of the West Antarctic Ice Sheet: comparison of simple and advanced statistical techniques

    NASA Astrophysics Data System (ADS)

    Pollard, David; Chang, Won; Haran, Murali; Applegate, Patrick; DeConto, Robert

    2016-05-01

    A 3-D hybrid ice-sheet model is applied to the last deglacial retreat of the West Antarctic Ice Sheet over the last ˜ 20 000 yr. A large ensemble of 625 model runs is used to calibrate the model to modern and geologic data, including reconstructed grounding lines, relative sea-level records, elevation-age data and uplift rates, with an aggregate score computed for each run that measures overall model-data misfit. Two types of statistical methods are used to analyze the large-ensemble results: simple averaging weighted by the aggregate score, and more advanced Bayesian techniques involving Gaussian process-based emulation and calibration, and Markov chain Monte Carlo. The analyses provide sea-level-rise envelopes with well-defined parametric uncertainty bounds, but the simple averaging method only provides robust results with full-factorial parameter sampling in the large ensemble. Results for best-fit parameter ranges and envelopes of equivalent sea-level rise with the simple averaging method agree well with the more advanced techniques. Best-fit parameter ranges confirm earlier values expected from prior model tuning, including large basal sliding coefficients on modern ocean beds.

  14. An open-source textbook for teaching climate-related risk analysis using the R computing environment

    NASA Astrophysics Data System (ADS)

    Applegate, P. J.; Keller, K.

    2015-12-01

    Greenhouse gas emissions lead to increased surface air temperatures and sea level rise. In turn, sea level rise increases the risks of flooding for people living near the world's coastlines. Our own research on assessing sea level rise-related risks emphasizes both Earth science and statistics. At the same time, the free, open-source computing environment R is growing in popularity among statisticians and scientists due to its flexibility and graphics capabilities, as well as its large library of existing functions. We have developed a set of laboratory exercises that introduce students to the Earth science and statistical concepts needed for assessing the risks presented by climate change, particularly sea-level rise. These exercises will be published as a free, open-source textbook on the Web. Each exercise begins with a description of the Earth science and/or statistical concepts that the exercise teaches, with references to key journal articles where appropriate. Next, students are asked to examine in detail a piece of existing R code, and the exercise text provides a clear explanation of how the code works. Finally, students are asked to modify the existing code to produce a well-defined outcome. We discuss our experiences in developing the exercises over two separate semesters at Penn State, plus using R Markdown to interweave explanatory text with sample code and figures in the textbook.

  15. Learning planar Ising models

    DOE PAGES

    Johnson, Jason K.; Oyen, Diane Adele; Chertkov, Michael; ...

    2016-12-01

    Inference and learning of graphical models are both well-studied problems in statistics and machine learning that have found many applications in science and engineering. However, exact inference is intractable in general graphical models, which suggests the problem of seeking the best approximation to a collection of random variables within some tractable family of graphical models. In this paper, we focus on the class of planar Ising models, for which exact inference is tractable using techniques of statistical physics. Based on these techniques and recent methods for planarity testing and planar embedding, we propose a greedy algorithm for learning the bestmore » planar Ising model to approximate an arbitrary collection of binary random variables (possibly from sample data). Given the set of all pairwise correlations among variables, we select a planar graph and optimal planar Ising model defined on this graph to best approximate that set of correlations. Finally, we demonstrate our method in simulations and for two applications: modeling senate voting records and identifying geo-chemical depth trends from Mars rover data.« less

  16. Learning planar Ising models

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Johnson, Jason K.; Oyen, Diane Adele; Chertkov, Michael

    Inference and learning of graphical models are both well-studied problems in statistics and machine learning that have found many applications in science and engineering. However, exact inference is intractable in general graphical models, which suggests the problem of seeking the best approximation to a collection of random variables within some tractable family of graphical models. In this paper, we focus on the class of planar Ising models, for which exact inference is tractable using techniques of statistical physics. Based on these techniques and recent methods for planarity testing and planar embedding, we propose a greedy algorithm for learning the bestmore » planar Ising model to approximate an arbitrary collection of binary random variables (possibly from sample data). Given the set of all pairwise correlations among variables, we select a planar graph and optimal planar Ising model defined on this graph to best approximate that set of correlations. Finally, we demonstrate our method in simulations and for two applications: modeling senate voting records and identifying geo-chemical depth trends from Mars rover data.« less

  17. Novel microbiological and spatial statistical methods to improve strength of epidemiological evidence in a community-wide waterborne outbreak.

    PubMed

    Jalava, Katri; Rintala, Hanna; Ollgren, Jukka; Maunula, Leena; Gomez-Alvarez, Vicente; Revez, Joana; Palander, Marja; Antikainen, Jenni; Kauppinen, Ari; Räsänen, Pia; Siponen, Sallamaari; Nyholm, Outi; Kyyhkynen, Aino; Hakkarainen, Sirpa; Merentie, Juhani; Pärnänen, Martti; Loginov, Raisa; Ryu, Hodon; Kuusi, Markku; Siitonen, Anja; Miettinen, Ilkka; Santo Domingo, Jorge W; Hänninen, Marja-Liisa; Pitkänen, Tarja

    2014-01-01

    Failures in the drinking water distribution system cause gastrointestinal outbreaks with multiple pathogens. A water distribution pipe breakage caused a community-wide waterborne outbreak in Vuorela, Finland, July 2012. We investigated this outbreak with advanced epidemiological and microbiological methods. A total of 473/2931 inhabitants (16%) responded to a web-based questionnaire. Water and patient samples were subjected to analysis of multiple microbial targets, molecular typing and microbial community analysis. Spatial analysis on the water distribution network was done and we applied a spatial logistic regression model. The course of the illness was mild. Drinking untreated tap water from the defined outbreak area was significantly associated with illness (RR 5.6, 95% CI 1.9-16.4) increasing in a dose response manner. The closer a person lived to the water distribution breakage point, the higher the risk of becoming ill. Sapovirus, enterovirus, single Campylobacter jejuni and EHEC O157:H7 findings as well as virulence genes for EPEC, EAEC and EHEC pathogroups were detected by molecular or culture methods from the faecal samples of the patients. EPEC, EAEC and EHEC virulence genes and faecal indicator bacteria were also detected in water samples. Microbial community sequencing of contaminated tap water revealed abundance of Arcobacter species. The polyphasic approach improved the understanding of the source of the infections, and aided to define the extent and magnitude of this outbreak.

  18. Effect of the extent of well purging on laboratory parameters of groundwater samples

    NASA Astrophysics Data System (ADS)

    Reka Mathe, Agnes; Kohler, Artur; Kovacs, Jozsef

    2017-04-01

    Chemicals reaching groundwater cause water quality deterioration. Reconnaissance and remediation demands high financial and human resources. Groundwater samples are important sources of information. Representativity of these samples is fundamental to decision making. According to relevant literature the way of sampling and the sampling equipment can affect laboratory concentrations measured in samples. Detailed and systematic research on this field is missing from even international literature. Groundwater sampling procedures are regulated worldwide. Regulations describe how to sample a groundwater monitoring well. The most common element in these regulations is well purging prior to sampling. The aim of purging the well is to avoid taking the sample from the stagnant water instead of from formation water. The stagnant water forms inside and around the well because the well casing provides direct contact with the atmosphere, changing the physico-chemical composition of the well water. Sample from the stagnant water is not representative of the formation water. Regulations regarding the extent of the purging are different. Purging is mostly defined as multiply (3-5) well volumes, and/or reaching stabilization of some purged water parameters (pH, specific conductivity, etc.). There are hints for sampling without purging. To define the necessary extent of the purging repeated pumping is conducted, triplicate samples are taken at the beginning of purging, at one, two and three times well volumes and at parameter stabilization. Triplicate samples are the means to account for laboratory errors. The subsurface is not static, the test is repeated 10 times. Up to now three tests were completed.

  19. ADAPTIVE MATCHING IN RANDOMIZED TRIALS AND OBSERVATIONAL STUDIES

    PubMed Central

    van der Laan, Mark J.; Balzer, Laura B.; Petersen, Maya L.

    2014-01-01

    SUMMARY In many randomized and observational studies the allocation of treatment among a sample of n independent and identically distributed units is a function of the covariates of all sampled units. As a result, the treatment labels among the units are possibly dependent, complicating estimation and posing challenges for statistical inference. For example, cluster randomized trials frequently sample communities from some target population, construct matched pairs of communities from those included in the sample based on some metric of similarity in baseline community characteristics, and then randomly allocate a treatment and a control intervention within each matched pair. In this case, the observed data can neither be represented as the realization of n independent random variables, nor, contrary to current practice, as the realization of n/2 independent random variables (treating the matched pair as the independent sampling unit). In this paper we study estimation of the average causal effect of a treatment under experimental designs in which treatment allocation potentially depends on the pre-intervention covariates of all units included in the sample. We define efficient targeted minimum loss based estimators for this general design, present a theorem that establishes the desired asymptotic normality of these estimators and allows for asymptotically valid statistical inference, and discuss implementation of these estimators. We further investigate the relative asymptotic efficiency of this design compared with a design in which unit-specific treatment assignment depends only on the units’ covariates. Our findings have practical implications for the optimal design and analysis of pair matched cluster randomized trials, as well as for observational studies in which treatment decisions may depend on characteristics of the entire sample. PMID:25097298

  20. The optical and near-infrared colors of galaxies, 1: The photometric data

    NASA Technical Reports Server (NTRS)

    Bershady, Matthew A.; Hereld, Mark; Kron, Richard G.; Koo, David C.; Munn, Jeffrey A.; Majewski, Steven R.

    1994-01-01

    We present optical and near-infrared photometry and spectroscopic redshifts of a well defined sample of 171 field galaxies selected from three high galactic latitude fields. This data set forms the basis for subsequent studies to characterize the trends, dispersion, and evolution of rest-frame colors and image structure. A subset of 143 galaxies constitutes a magnitude-limited sample to B approx. 19.9-20.75 (depending on field), with a median redshift of 0.14, and a maximum redshift of 0.54. This subset is statistically representative in its sampling of the apparent color distribution of galaxies. Thirty six galaxies were selected to have the reddest red-optical colors in two redshift intervals between 0.2 less than z less than 0.3. Photometric passbands are similar to U, B, V, I, and K, and sample galaxy spectral energy distributions between 0.37 and 2.2 micrometers in the observed frame, or down to 0.26 micrometers in the rest frame for the most distant galaxies. B and K images of the entire sample are assembled to form the first optical and near-infrared atlas of a statistically-representative sample of field galaxies. We discuss techniques for faint field-galaxy photometry, including a working definition of a total magnitude, and a method for matching magnitudes in different passbands and different seeing conditions to ensure reliable, integrated colors. Photographic saturation, which substantially affects the brightest 12% of our sample in the optical bands, is corrected with a model employing measured plate-density distributions for each galaxy, calibrated via similar measurements for stars as a function of known saturation level. Both the relative and absolute calibration of our photometry are demonstrated.

  1. Resolving the problem of trapped water in binding cavities: prediction of host-guest binding free energies in the SAMPL5 challenge by funnel metadynamics

    NASA Astrophysics Data System (ADS)

    Bhakat, Soumendranath; Söderhjelm, Pär

    2017-01-01

    The funnel metadynamics method enables rigorous calculation of the potential of mean force along an arbitrary binding path and thereby evaluation of the absolute binding free energy. A problem of such physical paths is that the mechanism characterizing the binding process is not always obvious. In particular, it might involve reorganization of the solvent in the binding site, which is not easily captured with a few geometrically defined collective variables that can be used for biasing. In this paper, we propose and test a simple method to resolve this trapped-water problem by dividing the process into an artificial host-desolvation step and an actual binding step. We show that, under certain circumstances, the contribution from the desolvation step can be calculated without introducing further statistical errors. We apply the method to the problem of predicting host-guest binding free energies in the SAMPL5 blind challenge, using two octa-acid hosts and six guest molecules. For one of the hosts, well-converged results are obtained and the prediction of relative binding free energies is the best among all the SAMPL5 submissions. For the other host, which has a narrower binding pocket, the statistical uncertainties are slightly higher; longer simulations would therefore be needed to obtain conclusive results.

  2. Rasch fit statistics and sample size considerations for polytomous data.

    PubMed

    Smith, Adam B; Rush, Robert; Fallowfield, Lesley J; Velikova, Galina; Sharpe, Michael

    2008-05-29

    Previous research on educational data has demonstrated that Rasch fit statistics (mean squares and t-statistics) are highly susceptible to sample size variation for dichotomously scored rating data, although little is known about this relationship for polytomous data. These statistics help inform researchers about how well items fit to a unidimensional latent trait, and are an important adjunct to modern psychometrics. Given the increasing use of Rasch models in health research the purpose of this study was therefore to explore the relationship between fit statistics and sample size for polytomous data. Data were collated from a heterogeneous sample of cancer patients (n = 4072) who had completed both the Patient Health Questionnaire - 9 and the Hospital Anxiety and Depression Scale. Ten samples were drawn with replacement for each of eight sample sizes (n = 25 to n = 3200). The Rating and Partial Credit Models were applied and the mean square and t-fit statistics (infit/outfit) derived for each model. The results demonstrated that t-statistics were highly sensitive to sample size, whereas mean square statistics remained relatively stable for polytomous data. It was concluded that mean square statistics were relatively independent of sample size for polytomous data and that misfit to the model could be identified using published recommended ranges.

  3. Rasch fit statistics and sample size considerations for polytomous data

    PubMed Central

    Smith, Adam B; Rush, Robert; Fallowfield, Lesley J; Velikova, Galina; Sharpe, Michael

    2008-01-01

    Background Previous research on educational data has demonstrated that Rasch fit statistics (mean squares and t-statistics) are highly susceptible to sample size variation for dichotomously scored rating data, although little is known about this relationship for polytomous data. These statistics help inform researchers about how well items fit to a unidimensional latent trait, and are an important adjunct to modern psychometrics. Given the increasing use of Rasch models in health research the purpose of this study was therefore to explore the relationship between fit statistics and sample size for polytomous data. Methods Data were collated from a heterogeneous sample of cancer patients (n = 4072) who had completed both the Patient Health Questionnaire – 9 and the Hospital Anxiety and Depression Scale. Ten samples were drawn with replacement for each of eight sample sizes (n = 25 to n = 3200). The Rating and Partial Credit Models were applied and the mean square and t-fit statistics (infit/outfit) derived for each model. Results The results demonstrated that t-statistics were highly sensitive to sample size, whereas mean square statistics remained relatively stable for polytomous data. Conclusion It was concluded that mean square statistics were relatively independent of sample size for polytomous data and that misfit to the model could be identified using published recommended ranges. PMID:18510722

  4. Evaluation of SLAR and thematic mapper MSS data for forest cover mapping using computer-aided analysis techniques

    NASA Technical Reports Server (NTRS)

    Hoffer, R. M. (Principal Investigator); Knowlton, D. J.; Dean, M. E.

    1981-01-01

    A set of training statistics for the 30 meter resolution simulated thematic mapper MSS data was generated based on land use/land cover classes. In addition to this supervised data set, a nonsupervised multicluster block of training statistics is being defined in order to compare the classification results and evaluate the effect of the different training selection methods on classification performance. Two test data sets, defined using a stratified sampling procedure incorporating a grid system with dimensions of 50 lines by 50 columns, and another set based on an analyst supervised set of test fields were used to evaluate the classifications of the TMS data. The supervised training data set generated training statistics, and a per point Gaussian maximum likelihood classification of the 1979 TMS data was obtained. The August 1980 MSS data was radiometrically adjusted. The SAR data was redigitized and the SAR imagery was qualitatively analyzed.

  5. 78 FR 73839 - 1,1,1,2 Tetrafluoroethane From the People's Republic of China: Initiation of Countervailing Duty...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-12-09

    ... using a statistically valid sampling method to poll the industry. Section 771(4)(A) of the Act defines... subheading 2903.39.2020. Although the HTSUS subheading and CAS registry number are provided for convenience...

  6. Summary of selected U.S. Geological survey data on domestic well water quality for the Centers for Disease Control's National Environmental Public Health Tracking Program

    USGS Publications Warehouse

    Bartholomay, Roy C.; Carter, Janet M.; Qi, Sharon L.; Squillace, Paul J.; Rowe, Gary L.

    2007-01-01

    About 10 to 30 percent of the population in most States uses domestic (private) water supply. In many States, the total number of people served by domestic supplies can be in the millions. The water quality of domestic supplies is inconsistently regulated and generally not well characterized. The U.S. Geological Survey (USGS) has two water-quality data sets in the National Water Information System (NWIS) database that can be used to help define the water quality of domestic-water supplies: (1) data from the National Water-Quality Assessment (NAWQA) Program, and (2) USGS State data. Data from domestic wells from the NAWQA Program were collected to meet one of the Program's objectives, which was to define the water quality of major aquifers in the United States. These domestic wells were located primarily in rural areas. Water-quality conditions in these major aquifers as defined by the NAWQA data can be compared because of the consistency of the NAWQA sampling design, sampling protocols, and water-quality analyses. The NWIS database is a repository of USGS water data collected for a variety of projects; consequently, project objectives and analytical methods vary. This variability can bias statistical summaries of contaminant occurrence and concentrations; nevertheless, these data can be used to define the geographic distribution of contaminants. Maps created using NAWQA and USGS State data in NWIS can show geographic areas where contaminant concentrations may be of potential human-health concern by showing concentrations relative to human-health water-quality benchmarks. On the basis of national summaries of detection frequencies and concentrations relative to U.S. Environmental Protection Agency (USEPA) human-health benchmarks for trace elements, pesticides, and volatile organic compounds, 28 water-quality constituents were identified as contaminants of potential human-health concern. From this list, 11 contaminants were selected for summarization of water-quality data in 16 States (grantee States) that were funded by the Environmental Public Health Tracking (EPHT) Program of the Centers for Disease Control and Prevention (CDC). Only data from domestic-water supplies were used in this summary because samples from these wells are most relevant to human exposure for the targeted population. Using NAWQA data, the concentrations of the 11 contaminants were compared to USEPA human-health benchmarks. Using NAWQA and USGS State data in NWIS, the geographic distribution of the contaminants were mapped for the 16 grantee States. Radon, arsenic, manganese, nitrate, strontium, and uranium had the largest percentages of samples with concentrations greater than their human-health benchmarks. In contrast, organic compounds (pesticides and volatile organic compounds) had the lowest percentages of samples with concentrations greater than human-health benchmarks. Results of data retrievals and spatial analysis were compiled for each of the 16 States and are presented in State summaries for each State. Example summary tables, graphs, and maps based on USGS data for New Jersey are presented to illustrate how USGS water-quality and associated ancillary geospatial data can be used by the CDC to address goals and objectives of the EPHT Program.

  7. Statistics and bioinformatics in nutritional sciences: analysis of complex data in the era of systems biology⋆

    PubMed Central

    Fu, Wenjiang J.; Stromberg, Arnold J.; Viele, Kert; Carroll, Raymond J.; Wu, Guoyao

    2009-01-01

    Over the past two decades, there have been revolutionary developments in life science technologies characterized by high throughput, high efficiency, and rapid computation. Nutritionists now have the advanced methodologies for the analysis of DNA, RNA, protein, low-molecular-weight metabolites, as well as access to bioinformatics databases. Statistics, which can be defined as the process of making scientific inferences from data that contain variability, has historically played an integral role in advancing nutritional sciences. Currently, in the era of systems biology, statistics has become an increasingly important tool to quantitatively analyze information about biological macromolecules. This article describes general terms used in statistical analysis of large, complex experimental data. These terms include experimental design, power analysis, sample size calculation, and experimental errors (type I and II errors) for nutritional studies at population, tissue, cellular, and molecular levels. In addition, we highlighted various sources of experimental variations in studies involving microarray gene expression, real-time polymerase chain reaction, proteomics, and other bioinformatics technologies. Moreover, we provided guidelines for nutritionists and other biomedical scientists to plan and conduct studies and to analyze the complex data. Appropriate statistical analyses are expected to make an important contribution to solving major nutrition-associated problems in humans and animals (including obesity, diabetes, cardiovascular disease, cancer, ageing, and intrauterine fetal retardation). PMID:20233650

  8. Statistical literacy and sample survey results

    NASA Astrophysics Data System (ADS)

    McAlevey, Lynn; Sullivan, Charles

    2010-10-01

    Sample surveys are widely used in the social sciences and business. The news media almost daily quote from them, yet they are widely misused. Using students with prior managerial experience embarking on an MBA course, we show that common sample survey results are misunderstood even by those managers who have previously done a statistics course. In general, they fare no better than managers who have never studied statistics. There are implications for teaching, especially in business schools, as well as for consulting.

  9. Methods for evaluating temporal groundwater quality data and results of decadal-scale changes in chloride, dissolved solids, and nitrate concentrations in groundwater in the United States, 1988-2010

    USGS Publications Warehouse

    Lindsey, Bruce D.; Rupert, Michael G.

    2012-01-01

    Decadal-scale changes in groundwater quality were evaluated by the U.S. Geological Survey National Water-Quality Assessment (NAWQA) Program. Samples of groundwater collected from wells during 1988-2000 - a first sampling event representing the decade ending the 20th century - were compared on a pair-wise basis to samples from the same wells collected during 2001-2010 - a second sampling event representing the decade beginning the 21st century. The data set consists of samples from 1,236 wells in 56 well networks, representing major aquifers and urban and agricultural land-use areas, with analytical results for chloride, dissolved solids, and nitrate. Statistical analysis was done on a network basis rather than by individual wells. Although spanning slightly more or less than a 10-year period, the two-sample comparison between the first and second sampling events is referred to as an analysis of decadal-scale change based on a step-trend analysis. The 22 principal aquifers represented by these 56 networks account for nearly 80 percent of the estimated withdrawals of groundwater used for drinking-water supply in the Nation. Well networks where decadal-scale changes in concentrations were statistically significant were identified using the Wilcoxon-Pratt signed-rank test. For the statistical analysis of chloride, dissolved solids, and nitrate concentrations at the network level, more than half revealed no statistically significant change over the decadal period. However, for networks that had statistically significant changes, increased concentrations outnumbered decreased concentrations by a large margin. Statistically significant increases of chloride concentrations were identified for 43 percent of 56 networks. Dissolved solids concentrations increased significantly in 41 percent of the 54 networks with dissolved solids data, and nitrate concentrations increased significantly in 23 percent of 56 networks. At least one of the three - chloride, dissolved solids, or nitrate - had a statistically significant increase in concentration in 66 percent of the networks. Statistically significant decreases in concentrations were identified in 4 percent of the networks for chloride, 2 percent of the networks for dissolved solids, and 9 percent of the networks for nitrate. A larger percentage of urban land-use networks had statistically significant increases in chloride, dissolved solids, and nitrate concentrations than agricultural land-use networks. In order to assess the magnitude of statistically significant changes, the median of the differences between constituent concentrations from the first full-network sampling event and those from the second full-network sampling event was calculated using the Turnbull method. The largest median decadal increases in chloride concentrations were in networks in the Upper Illinois River Basin (67 mg/L) and in the New England Coastal Basins (34 mg/L), whereas the largest median decadal decrease in chloride concentrations was in the Upper Snake River Basin (1 mg/L). The largest median decadal increases in dissolved solids concentrations were in networks in the Rio Grande Valley (260 mg/L) and the Upper Illinois River Basin (160 mg/L). The largest median decadal decrease in dissolved solids concentrations was in the Apalachicola-Chattahoochee-Flint River Basin (6.0 mg/L). The largest median decadal increases in nitrate as nitrogen (N) concentrations were in networks in the South Platte River Basin (2.0 mg/L as N) and the San Joaquin-Tulare Basins (1.0 mg/L as N). The largest median decadal decrease in nitrate concentrations was in the Santee River Basin and Coastal Drainages (0.63 mg/L). The magnitude of change in networks with statistically significant increases typically was much larger than the magnitude of change in networks with statistically significant decreases. The magnitude of change was greatest for chloride in the urban land-use networks and greatest for dissolved solids and nitrate in the agricultural land-use networks. Analysis of data from all networks combined indicated statistically significant increases for chloride, dissolved solids, and nitrate. Although chloride, dissolved solids, and nitrate concentrations were typically less than the drinking-water standards and guidelines, a statistical test was used to determine whether or not the proportion of samples exceeding the drinking-water standard or guideline changed significantly between the first and second full-network sampling events. The proportion of samples exceeding the U.S. Environmental Protection Agency (USEPA) Secondary Maximum Contaminant Level for dissolved solids (500 milligrams per liter) increased significantly between the first and second full-network sampling events when evaluating all networks combined at the national level. Also, for all networks combined, the proportion of samples exceeding the USEPA Maximum Contaminant Level (MCL) of 10 mg/L as N for nitrate increased significantly. One network in the Delmarva Peninsula had a significant increase in the proportion of samples exceeding the MCL for nitrate. A subset of 261 wells was sampled every other year (biennially) to evaluate decadal-scale changes using a time-series analysis. The analysis of the biennial data set showed that changes were generally similar to the findings from the analysis of decadal-scale change that was based on a step-trend analysis. Because of the small number of wells in a network with biennial data (typically 4-5 wells), the time-series analysis is more useful for understanding water-quality responses to changes in site-specific conditions rather than as an indicator of the change for the entire network.

  10. Iron isotopes in ancient and modern komatiites: Evidence in support of an oxidised mantle from Archean to present

    NASA Astrophysics Data System (ADS)

    Hibbert, K. E. J.; Williams, H. M.; Kerr, A. C.; Puchtel, I. S.

    2012-03-01

    The mantle of the modern Earth is relatively oxidised compared to the initially reducing conditions inferred for core formation. The timing of the oxidation of the mantle is not conclusively resolved but has important implications for the timing of the development of the hydrosphere and atmosphere. In order to examine the timing of this oxidation event, we present iron isotope data from three exceptionally well preserved komatiite localities, Belingwe (2.7 Ga), Vetreny (2.4 Ga) and Gorgona (0.089 Ga). Measurements of Fe isotope compositions of whole-rock samples are complemented by the analysis of olivine, spinel and pyroxene separates. Bulk-rock and olivine Fe isotope compositions (δ57Fe) define clear linear correlations with indicators of magmatic differentiation (Mg#, Cr#). The mean Fe isotope compositions of the 2.7-2.4 Ga and 0.089 Ga samples are statistically distinct and this difference can be explained by greater extent of partial melting represented by the older samples and higher mantle ambient temperatures in the Archean and early Proterozoic relative to the present day. Significantly, samples of all ages define continuous positive linear correlations between bulk rock δ57Fe and V/Sc and δ57Fe and V, and between V/Sc and V with TiO2, providing evidence for the incompatible behaviour of V (relative to Sc) and of isotopically heavy Fe. Partial melting models calculated using partition coefficients for V at oxygen fugacities (fO2s) of 0 and + 1 relative to the fayalite-magnetite-quartz buffer (FMQ) best match the data arrays, which are defined by all samples, from late Archean to Tertiary. These data, therefore, provide evidence for komatiite generation under moderately oxidising conditions since the late Archean, and argue against a change in mantle fO2 concomitant with atmospheric oxygenation at ~ 2.4 Ga.

  11. The NASA/AFRL Meter Class Autonomous Telescope

    NASA Technical Reports Server (NTRS)

    Cowardin, H.; Lederer, S.; Buckalew, B.; Frith, J.; Hickson, P.; Glesne, T.; Anz-Meador, P.; Barker, E.; Stansbery, G.; Kervin, P.

    2016-01-01

    For the past decade, the NASA Orbital Debris Program Office (ODPO) has relied on using various ground-based telescopes in Chile to acquire statistical survey data as well as photometric and spectroscopic data of orbital debris in geosynchronous Earth orbit (GEO). The statistical survey data have been used to supply the Orbital Debris Engineering Model (ORDEM) v.3.0 with debris detections in GEO to better model the environment at altitudes where radar detections are limited. The data produced for the statistical survey ranged from 30 to 40 nights per year, which only accounted for 10% of the possible observing time. Data collection was restricted by ODPO resources and weather conditions. In order to improve the statistical sampling in GEO, as well as observe and sample other orbits, NASA's ODPO with support from the Air Force Research Laboratory (AFRL), has constructed a new observatory dedicated to orbital debris - the Meter Class Autonomous Telescope (MCAT) on Ascension Island. This location provides MCAT with the unique ability to access targets orbiting at an altitude of less than 1,000 km and low inclinations (< 20 deg). This orbital regime currently has little to no coverage by the U.S. Space Surveillance Network. Unlike previous ODPO optical assets, the ability to operate autonomously will allow rapid response observations of break-up events, an observing mode that was only available via radar tasking prior to MCAT's deployment. The primary goal of MCAT is to statistically characterize GEO via daily tasking files uploaded from ODPO. These tasking files define which operating mode to follow, providing the field center, rates, and/or targets to observe over the entire observing period. The system is also capable of tracking fast-moving targets in low Earth orbit (LEO), middle Earth orbit (MEO), as well as highly eccentric orbits like geostationary transfer orbits. On 25 August 2015, MCAT successfully acquired scientific first light, imaging the Bug Nebula and tracked objects in LEO, MEO, and GEO. NASA is working towards characterizing the system and thoroughly testing the integrated hardware and software control to achieve fully autonomous operations by late 2016. This paper will review the history and current status of the MCAT project, the details of the telescope system, and its five currently manifested operating modes.

  12. How things fall apart: understanding the nature of internalizing through its relationship with impairment.

    PubMed

    Markon, Kristian E

    2010-08-01

    The literature suggests that internalizing psychopathology relates to impairment incrementally and gradually. However, the form of this relationship has not been characterized. This form is critical to understanding internalizing psychopathology, as it is possible that internalizing may accelerate in effect at some level of severity, defining a natural boundary of abnormality. Here, a novel method-semiparametric structural equation modeling-was used to model the relationship between internalizing and impairment in a sample of 8,580 individuals from the 2000 British Office for National Statistics Survey of Psychiatric Morbidity, a large, population-representative study of psychopathology. This method allows one to model relationships between latent internalizing and impairment without assuming any particular form a priori and to compare models in which the relationship is constant and linear. Results suggest that the relationship between internalizing and impairment is in fact linear and constant across the entire range of internalizing variation and that it is impossible to nonarbitrarily define a specific level of internalizing beyond which consequences suddenly become catastrophic in nature. Results demonstrate the phenomenological continuity of internalizing psychopathology, highlight the importance of impairment as well as symptoms, and have clear implications for defining mental disorder. Copyright 2010 APA, all rights reserved

  13. The Amount of Media and Information Literacy Among Isfahan University of Medical Sciences' Students Using Iranian Media and Information Literacy Questionnaire (IMILQ).

    PubMed

    Ashrafi-Rizi, Hasan; Ramezani, Amir; Koupaei, Hamed Aghajani; Kazempour, Zahra

    2014-12-01

    Media and Information literacy (MIL) enables people to interpret and make informed judgments as users of information and media, as well as to become skillful creators and producers of information and media messages in their own right. The purpose of this research was to determine the amount of Media and Information Literacy among Isfahan University of Medical Sciences' students using Iranian Media and Information Literacy Questionnaire (IMILQ). This is an applied analytical survey research in which the data were collected by a researcher made questionnaire, provided based on specialists' viewpoints and valid scientific works. Its validity and reliability were confirmed by Library and Information Sciences specialists and Cronbach's alpha (r=0.89) respectively. Statistical population consisted of all students in Isfahan University of Medical Sciences (6000 cases) and the samples were 361. Sampling method was random stratified sampling. Data were analyzed by descriptive and inferential statistics. The findings showed that the mean level of Media and Information Literacy among Isfahan University of Medical Sciences' students was 3.34±0.444 (higher than average). The highest mean was promotion of scientific degree with 3.84±0.975 and the lowest mean was difficulties in starting research with 2.50±1.08. There was significant difference between educational degree, college type and family's income and amount of Media and Information Literacy. The results showed that the students didn't have enough skills in starting the research, defining the research subject as well as confining the research subject. In general, all students and education practitioners should pay special attention to factors affecting in improving Media and Information Literacy as a main capability in using printed and electronic media.

  14. The Amount of Media and Information Literacy Among Isfahan University of Medical Sciences’ Students Using Iranian Media and Information Literacy Questionnaire (IMILQ)

    PubMed Central

    Ashrafi-rizi, Hasan; Ramezani, Amir; Koupaei, Hamed Aghajani; Kazempour, Zahra

    2014-01-01

    Introduction: Media and Information literacy (MIL) enables people to interpret and make informed judgments as users of information and media, as well as to become skillful creators and producers of information and media messages in their own right. The purpose of this research was to determine the amount of Media and Information Literacy among Isfahan University of Medical Sciences’ students using Iranian Media and Information Literacy Questionnaire (IMILQ). Methods: This is an applied analytical survey research in which the data were collected by a researcher made questionnaire, provided based on specialists’ viewpoints and valid scientific works. Its validity and reliability were confirmed by Library and Information Sciences specialists and Cronbach’s alpha (r=0.89) respectively. Statistical population consisted of all students in Isfahan University of Medical Sciences (6000 cases) and the samples were 361. Sampling method was random stratified sampling. Data were analyzed by descriptive and inferential statistics. Results: The findings showed that the mean level of Media and Information Literacy among Isfahan University of Medical Sciences’ students was 3.34±0.444 (higher than average). The highest mean was promotion of scientific degree with 3.84±0.975 and the lowest mean was difficulties in starting research with 2.50±1.08. There was significant difference between educational degree, college type and family’s income and amount of Media and Information Literacy. Conclusion: The results showed that the students didn’t have enough skills in starting the research, defining the research subject as well as confining the research subject. In general, all students and education practitioners should pay special attention to factors affecting in improving Media and Information Literacy as a main capability in using printed and electronic media. PMID:25684848

  15. Statistical investigation of avalanches of three-dimensional small-world networks and their boundary and bulk cross-sections

    NASA Astrophysics Data System (ADS)

    Najafi, M. N.; Dashti-Naserabadi, H.

    2018-03-01

    In many situations we are interested in the propagation of energy in some portions of a three-dimensional system with dilute long-range links. In this paper, a sandpile model is defined on the three-dimensional small-world network with real dissipative boundaries and the energy propagation is studied in three dimensions as well as the two-dimensional cross-sections. Two types of cross-sections are defined in the system, one in the bulk and another in the system boundary. The motivation of this is to make clear how the statistics of the avalanches in the bulk cross-section tend to the statistics of the dissipative avalanches, defined in the boundaries as the concentration of long-range links (α ) increases. This trend is numerically shown to be a power law in a manner described in the paper. Two regimes of α are considered in this work. For sufficiently small α s the dominant behavior of the system is just like that of the regular BTW, whereas for the intermediate values the behavior is nontrivial with some exponents that are reported in the paper. It is shown that the spatial extent up to which the statistics is similar to the regular BTW model scales with α just like the dissipative BTW model with the dissipation factor (mass in the corresponding ghost model) m2˜α for the three-dimensional system as well as its two-dimensional cross-sections.

  16. TRAM (Transcriptome Mapper): database-driven creation and analysis of transcriptome maps from multiple sources

    PubMed Central

    2011-01-01

    Background Several tools have been developed to perform global gene expression profile data analysis, to search for specific chromosomal regions whose features meet defined criteria as well as to study neighbouring gene expression. However, most of these tools are tailored for a specific use in a particular context (e.g. they are species-specific, or limited to a particular data format) and they typically accept only gene lists as input. Results TRAM (Transcriptome Mapper) is a new general tool that allows the simple generation and analysis of quantitative transcriptome maps, starting from any source listing gene expression values for a given gene set (e.g. expression microarrays), implemented as a relational database. It includes a parser able to assign univocal and updated gene symbols to gene identifiers from different data sources. Moreover, TRAM is able to perform intra-sample and inter-sample data normalization, including an original variant of quantile normalization (scaled quantile), useful to normalize data from platforms with highly different numbers of investigated genes. When in 'Map' mode, the software generates a quantitative representation of the transcriptome of a sample (or of a pool of samples) and identifies if segments of defined lengths are over/under-expressed compared to the desired threshold. When in 'Cluster' mode, the software searches for a set of over/under-expressed consecutive genes. Statistical significance for all results is calculated with respect to genes localized on the same chromosome or to all genome genes. Transcriptome maps, showing differential expression between two sample groups, relative to two different biological conditions, may be easily generated. We present the results of a biological model test, based on a meta-analysis comparison between a sample pool of human CD34+ hematopoietic progenitor cells and a sample pool of megakaryocytic cells. Biologically relevant chromosomal segments and gene clusters with differential expression during the differentiation toward megakaryocyte were identified. Conclusions TRAM is designed to create, and statistically analyze, quantitative transcriptome maps, based on gene expression data from multiple sources. The release includes FileMaker Pro database management runtime application and it is freely available at http://apollo11.isto.unibo.it/software/, along with preconfigured implementations for mapping of human, mouse and zebrafish transcriptomes. PMID:21333005

  17. MAP: an iterative experimental design methodology for the optimization of catalytic search space structure modeling.

    PubMed

    Baumes, Laurent A

    2006-01-01

    One of the main problems in high-throughput research for materials is still the design of experiments. At early stages of discovery programs, purely exploratory methodologies coupled with fast screening tools should be employed. This should lead to opportunities to find unexpected catalytic results and identify the "groups" of catalyst outputs, providing well-defined boundaries for future optimizations. However, very few new papers deal with strategies that guide exploratory studies. Mostly, traditional designs, homogeneous covering, or simple random samplings are exploited. Typical catalytic output distributions exhibit unbalanced datasets for which an efficient learning is hardly carried out, and interesting but rare classes are usually unrecognized. Here is suggested a new iterative algorithm for the characterization of the search space structure, working independently of learning processes. It enhances recognition rates by transferring catalysts to be screened from "performance-stable" space zones to "unsteady" ones which necessitate more experiments to be well-modeled. The evaluation of new algorithm attempts through benchmarks is compulsory due to the lack of past proofs about their efficiency. The method is detailed and thoroughly tested with mathematical functions exhibiting different levels of complexity. The strategy is not only empirically evaluated, the effect or efficiency of sampling on future Machine Learning performances is also quantified. The minimum sample size required by the algorithm for being statistically discriminated from simple random sampling is investigated.

  18. Sampling Errors in Monthly Rainfall Totals for TRMM and SSM/I, Based on Statistics of Retrieved Rain Rates and Simple Models

    NASA Technical Reports Server (NTRS)

    Bell, Thomas L.; Kundu, Prasun K.; Einaudi, Franco (Technical Monitor)

    2000-01-01

    Estimates from TRMM satellite data of monthly total rainfall over an area are subject to substantial sampling errors due to the limited number of visits to the area by the satellite during the month. Quantitative comparisons of TRMM averages with data collected by other satellites and by ground-based systems require some estimate of the size of this sampling error. A method of estimating this sampling error based on the actual statistics of the TRMM observations and on some modeling work has been developed. "Sampling error" in TRMM monthly averages is defined here relative to the monthly total a hypothetical satellite permanently stationed above the area would have reported. "Sampling error" therefore includes contributions from the random and systematic errors introduced by the satellite remote sensing system. As part of our long-term goal of providing error estimates for each grid point accessible to the TRMM instruments, sampling error estimates for TRMM based on rain retrievals from TRMM microwave (TMI) data are compared for different times of the year and different oceanic areas (to minimize changes in the statistics due to algorithmic differences over land and ocean). Changes in sampling error estimates due to changes in rain statistics due 1) to evolution of the official algorithms used to process the data, and 2) differences from other remote sensing systems such as the Defense Meteorological Satellite Program (DMSP) Special Sensor Microwave/Imager (SSM/I), are analyzed.

  19. DESCARTES' RULE OF SIGNS AND THE IDENTIFIABILITY OF POPULATION DEMOGRAPHIC MODELS FROM GENOMIC VARIATION DATA.

    PubMed

    Bhaskar, Anand; Song, Yun S

    2014-01-01

    The sample frequency spectrum (SFS) is a widely-used summary statistic of genomic variation in a sample of homologous DNA sequences. It provides a highly efficient dimensional reduction of large-scale population genomic data and its mathematical dependence on the underlying population demography is well understood, thus enabling the development of efficient inference algorithms. However, it has been recently shown that very different population demographies can actually generate the same SFS for arbitrarily large sample sizes. Although in principle this nonidentifiability issue poses a thorny challenge to statistical inference, the population size functions involved in the counterexamples are arguably not so biologically realistic. Here, we revisit this problem and examine the identifiability of demographic models under the restriction that the population sizes are piecewise-defined where each piece belongs to some family of biologically-motivated functions. Under this assumption, we prove that the expected SFS of a sample uniquely determines the underlying demographic model, provided that the sample is sufficiently large. We obtain a general bound on the sample size sufficient for identifiability; the bound depends on the number of pieces in the demographic model and also on the type of population size function in each piece. In the cases of piecewise-constant, piecewise-exponential and piecewise-generalized-exponential models, which are often assumed in population genomic inferences, we provide explicit formulas for the bounds as simple functions of the number of pieces. Lastly, we obtain analogous results for the "folded" SFS, which is often used when there is ambiguity as to which allelic type is ancestral. Our results are proved using a generalization of Descartes' rule of signs for polynomials to the Laplace transform of piecewise continuous functions.

  20. DESCARTES’ RULE OF SIGNS AND THE IDENTIFIABILITY OF POPULATION DEMOGRAPHIC MODELS FROM GENOMIC VARIATION DATA1

    PubMed Central

    Bhaskar, Anand; Song, Yun S.

    2016-01-01

    The sample frequency spectrum (SFS) is a widely-used summary statistic of genomic variation in a sample of homologous DNA sequences. It provides a highly efficient dimensional reduction of large-scale population genomic data and its mathematical dependence on the underlying population demography is well understood, thus enabling the development of efficient inference algorithms. However, it has been recently shown that very different population demographies can actually generate the same SFS for arbitrarily large sample sizes. Although in principle this nonidentifiability issue poses a thorny challenge to statistical inference, the population size functions involved in the counterexamples are arguably not so biologically realistic. Here, we revisit this problem and examine the identifiability of demographic models under the restriction that the population sizes are piecewise-defined where each piece belongs to some family of biologically-motivated functions. Under this assumption, we prove that the expected SFS of a sample uniquely determines the underlying demographic model, provided that the sample is sufficiently large. We obtain a general bound on the sample size sufficient for identifiability; the bound depends on the number of pieces in the demographic model and also on the type of population size function in each piece. In the cases of piecewise-constant, piecewise-exponential and piecewise-generalized-exponential models, which are often assumed in population genomic inferences, we provide explicit formulas for the bounds as simple functions of the number of pieces. Lastly, we obtain analogous results for the “folded” SFS, which is often used when there is ambiguity as to which allelic type is ancestral. Our results are proved using a generalization of Descartes’ rule of signs for polynomials to the Laplace transform of piecewise continuous functions. PMID:28018011

  1. Ultrasound criteria and guided fine-needle aspiration diagnostic yields in small animal peritoneal, mesenteric and omental disease.

    PubMed

    Feeney, Daniel A; Ober, Christopher P; Snyder, Laura A; Hill, Sara A; Jessen, Carl R

    2013-01-01

    Peritoneal, mesenteric, and omental diseases are important causes of morbidity and mortality in humans and animals, although information in the veterinary literature is limited. The purposes of this retrospective study were to determine whether objectively applied ultrasound interpretive criteria are statistically useful in differentiating among cytologically defined normal, inflammatory, and neoplastic peritoneal conditions in dogs and cats. A second goal was to determine the cytologically interpretable yield on ultrasound-guided, fine-needle sampling of peritoneal, mesenteric, or omental structures. Sonographic criteria agreed upon by the authors were retrospectively and independently applied by two radiologists to the available ultrasound images without knowledge of the cytologic diagnosis and statistically compared to the ultrasound-guided, fine-needle aspiration cytologic interpretations. A total of 72 dogs and 49 cats with abdominal peritoneal, mesenteric, or omental (peritoneal) surface or effusive disease and 17 dogs and 3 cats with no cytologic evidence of inflammation or neoplasia were included. The optimized, ultrasound criteria-based statistical model created independently for each radiologist yielded an equation-based diagnostic category placement accuracy of 63.2-69.9% across the two involved radiologists. Regional organ-associated masses or nodules as well as aggregated bowel and peritoneal thickening were more associated with peritoneal neoplasia whereas localized, severely complex fluid collections were more associated with inflammatory peritoneal disease. The cytologically interpretable yield for ultrasound-guided fine-needle sampling was 72.3% with no difference between species, making this a worthwhile clinical procedure. © 2013 Veterinary Radiology & Ultrasound.

  2. The SDSS-III Multi-object Apo Radial-velocity Exoplanet Large-area Survey

    NASA Astrophysics Data System (ADS)

    Ge, Jian; Mahadevan, S.; Lee, B.; Wan, X.; Zhao, B.; van Eyken, J.; Kane, S.; Guo, P.; Ford, E. B.; Agol, E.; Gaudi, S.; Fleming, S.; Crepp, J.; Cohen, R.; Groot, J.; Galvez, M.; Liu, J.; Ford, H.; Schneider, D.; Seager, S.; Hawley, S. L.; Weinberg, D.; Eisenstein, D.

    2007-12-01

    As part of SDSS-III survey in 2008-2014, the Multi-object APO Radial-Velocity Exoplanet Large-area Survey (MARVELS) will conduct the largest ground-based Doppler planet survey to date using the SDSS telescope and new generation multi-object Doppler instruments with 120 object capability and 10-20 m/s Doppler precision. The baseline survey plan is to monitor a total of 11,000 V=8-12 stars ( 10,000 main sequence stars and 1000 giant stars) over 800 square degrees over the 6 years. The primary goal is to produce a large, statistically well defined sample of giant planets ( 200) with a wide range of masses ( 0.2-10 Jupiter masses) and orbits (1 day-2 years) drawn from a large of host stars with a diverse set of masses, compositions, and ages for studying the diversity of extrasolar planets and constraining planet formation, migration & dynamical evolution of planetary systems. The survey data will also be used for providing a statistical sample for theoretical comparison and discovering rare systems and identifying signposts for lower-mass or more distant planets. Early science results from the pilot program will be reported. We would like to thank the SDSS MC for allocation of the telescope time and the W.M. Keck Foundation, NSF, NASA and UF for support.

  3. Estimation of pyrethroid pesticide intake using regression ...

    EPA Pesticide Factsheets

    Population-based estimates of pesticide intake are needed to characterize exposure for particular demographic groups based on their dietary behaviors. Regression modeling performed on measurements of selected pesticides in composited duplicate diet samples allowed (1) estimation of pesticide intakes for a defined demographic community, and (2) comparison of dietary pesticide intakes between the composite and individual samples. Extant databases were useful for assigning individual samples to composites, but they could not provide the breadth of information needed to facilitate measurable levels in every composite. Composite sample measurements were found to be good predictors of pyrethroid pesticide levels in their individual sample constituents where sufficient measurements are available above the method detection limit. Statistical inference shows little evidence of differences between individual and composite measurements and suggests that regression modeling of food groups based on composite dietary samples may provide an effective tool for estimating dietary pesticide intake for a defined population. The research presented in the journal article will improve community's ability to determine exposures through the dietary route with a less burdensome and costly method.

  4. Ichthyoplankton abundance and variance in a large river system concerns for long-term monitoring

    USGS Publications Warehouse

    Holland-Bartels, Leslie E.; Dewey, Michael R.; Zigler, Steven J.

    1995-01-01

    System-wide spatial patterns of ichthyoplankton abundance and variability were assessed in the upper Mississippi and lower Illinois rivers to address the experimental design and statistical confidence in density estimates. Ichthyoplankton was sampled from June to August 1989 in primary milieus (vegetated and non-vegated backwaters and impounded areas, main channels and main channel borders) in three navigation pools (8, 13 and 26) of the upper Mississippi River and in a downstream reach of the Illinois River. Ichthyoplankton densities varied among stations of similar aquatic landscapes (milieus) more than among subsamples within a station. An analysis of sampling effort indicated that the collection of single samples at many stations in a given milieu type is statistically and economically preferable to the collection of multiple subsamples at fewer stations. Cluster analyses also revealed that stations only generally grouped by their preassigned milieu types. Pilot studies such as this can define station groupings and sources of variation beyond an a priori habitat classification. Thus the minimum intensity of sampling required to achieve a desired statistical confidence can be identified before implementing monitoring efforts.

  5. Roadmap for Navy Family Research.

    DTIC Science & Technology

    1980-08-01

    of methodological limitations, including: small, often non -representative or narrowly defined samples; inadequate statistical controls, inadequate...1-1 1.2 Overview of the Research Roadmap ..................... 1-2 2. Methodology ...the Office of Naval Research by the Westinghouse Public Applied Systems Division, and is designed to provide the Navy with a systematic framework for

  6. 78 FR 51133 - Submission for OMB Review; Comment Request

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-08-20

    ... a currently valid OMB control number. National Agricultural Statistics Service Title: Wheat and... surveys. This project is conducted as a cooperative effort with the U.S. Wheat and Barley Scab Initiative... of the Information: The survey will use a sampling universe defined as producers that harvest wheat...

  7. Quasi-Monochromatic Visual Environments and the Resting Point of Accommodation

    DTIC Science & Technology

    1988-01-01

    accommodation. No statistically significant differences were revealed to support the possibility of color mediated differential regression to resting...discussed with respect to the general findings of the total sample as well as the specific behavior of individual participants. The summarized statistics ...remaining ten varied considerably with respect to the averaged trends reported in the above descriptive statistics as well as with respect to precision

  8. Dental fear and caries in 6-12 year old children in Greece. Determination of dental fear cut-off points.

    PubMed

    Boka, V; Arapostathis, K; Karagiannis, V; Kotsanos, N; van Loveren, C; Veerkamp, J

    2017-03-01

    To present: the normative data on dental fear and caries status; the dental fear cut-off points of young children in the city of Thessaloniki, Greece. Study Design: This is a cross-sectional study with two independent study groups. A first representative sample consisted of 1484 children from 15 primary public schools of Thessaloniki. A second sample consisted of 195 randomly selected age-matched children, all patients of the Postgraduate Paediatric Dental Clinic of Aristotle University of Thessaloniki. First sample: In order to select data on dental fear and caries, dental examination took place in the classroom with disposable mirrors and a penlight. All the children completed the Dental Subscale of the Children's Fear Survey Schedule (CFSS-DS). Second sample: In order to define the cut-off points of the CFSS-DS, dental treatment of the 195 children was performed at the University Clinic. Children⁁s dental fear was assessed using the CFSS-DS and their behaviour during dental treatment was observed by one calibrated examiner using the Venham scale. Statistical analysis of the data was performed with IBM SPSS Statistics 20 at a statistical significance level of <0.05. First sample: The mean CFSS-DS score was 27.1±10.8. Age was significantly (p<0.05) related to dental fear. Mean differences between boys and girls were not significant. Caries was not correlated with dental fear. Second sample: CFSS-DS< 33 was defined as 'no dental fear', scores 33-37 as 'borderline' and scores > 37 as 'dental fear'. In the first sample, 84.6% of the children did not suffer from dental fear (CFSS-DS<33). Dental fear was correlated to age and not to caries and gender. The dental fear cut-off point for the CFSS-DS was estimated at 37 for 6-12 year old children (33-37 borderlines).

  9. Emergent Irreversibility and Entanglement Spectrum Statistics

    NASA Astrophysics Data System (ADS)

    Chamon, Claudio; Hamma, Alioscia; Mucciolo, Eduardo R.

    2014-06-01

    We study the problem of irreversibility when the dynamical evolution of a many-body system is described by a stochastic quantum circuit. Such evolution is more general than a Hamiltonian one, and since energy levels are not well defined, the well-established connection between the statistical fluctuations of the energy spectrum and irreversibility cannot be made. We show that the entanglement spectrum provides a more general connection. Irreversibility is marked by a failure of a disentangling algorithm and is preceded by the appearance of Wigner-Dyson statistical fluctuations in the entanglement spectrum. This analysis can be done at the wave-function level and offers an alternative route to study quantum chaos and quantum integrability.

  10. Statistical computation of tolerance limits

    NASA Technical Reports Server (NTRS)

    Wheeler, J. T.

    1993-01-01

    Based on a new theory, two computer codes were developed specifically to calculate the exact statistical tolerance limits for normal distributions within unknown means and variances for the one-sided and two-sided cases for the tolerance factor, k. The quantity k is defined equivalently in terms of the noncentral t-distribution by the probability equation. Two of the four mathematical methods employ the theory developed for the numerical simulation. Several algorithms for numerically integrating and iteratively root-solving the working equations are written to augment the program simulation. The program codes generate some tables of k's associated with the varying values of the proportion and sample size for each given probability to show accuracy obtained for small sample sizes.

  11. Automated Classification and Analysis of Non-metallic Inclusion Data Sets

    NASA Astrophysics Data System (ADS)

    Abdulsalam, Mohammad; Zhang, Tongsheng; Tan, Jia; Webler, Bryan A.

    2018-05-01

    The aim of this study is to utilize principal component analysis (PCA), clustering methods, and correlation analysis to condense and examine large, multivariate data sets produced from automated analysis of non-metallic inclusions. Non-metallic inclusions play a major role in defining the properties of steel and their examination has been greatly aided by automated analysis in scanning electron microscopes equipped with energy dispersive X-ray spectroscopy. The methods were applied to analyze inclusions on two sets of samples: two laboratory-scale samples and four industrial samples from a near-finished 4140 alloy steel components with varying machinability. The laboratory samples had well-defined inclusions chemistries, composed of MgO-Al2O3-CaO, spinel (MgO-Al2O3), and calcium aluminate inclusions. The industrial samples contained MnS inclusions as well as (Ca,Mn)S + calcium aluminate oxide inclusions. PCA could be used to reduce inclusion chemistry variables to a 2D plot, which revealed inclusion chemistry groupings in the samples. Clustering methods were used to automatically classify inclusion chemistry measurements into groups, i.e., no user-defined rules were required.

  12. Experimental Investigations of Non-Stationary Properties In Radiometer Receivers Using Measurements of Multiple Calibration References

    NASA Technical Reports Server (NTRS)

    Racette, Paul; Lang, Roger; Zhang, Zhao-Nan; Zacharias, David; Krebs, Carolyn A. (Technical Monitor)

    2002-01-01

    Radiometers must be periodically calibrated because the receiver response fluctuates. Many techniques exist to correct for the time varying response of a radiometer receiver. An analytical technique has been developed that uses generalized least squares regression (LSR) to predict the performance of a wide variety of calibration algorithms. The total measurement uncertainty including the uncertainty of the calibration can be computed using LSR. The uncertainties of the calibration samples used in the regression are based upon treating the receiver fluctuations as non-stationary processes. Signals originating from the different sources of emission are treated as simultaneously existing random processes. Thus, the radiometer output is a series of samples obtained from these random processes. The samples are treated as random variables but because the underlying processes are non-stationary the statistics of the samples are treated as non-stationary. The statistics of the calibration samples depend upon the time for which the samples are to be applied. The statistics of the random variables are equated to the mean statistics of the non-stationary processes over the interval defined by the time of calibration sample and when it is applied. This analysis opens the opportunity for experimental investigation into the underlying properties of receiver non stationarity through the use of multiple calibration references. In this presentation we will discuss the application of LSR to the analysis of various calibration algorithms, requirements for experimental verification of the theory, and preliminary results from analyzing experiment measurements.

  13. Reliability and longitudinal change of detrital-zircon age spectra in the Snake River system, Idaho and Wyoming: An example of reproducing the bumpy barcode

    NASA Astrophysics Data System (ADS)

    Link, Paul Karl; Fanning, C. Mark; Beranek, Luke P.

    2005-12-01

    Detrital-zircon age-spectra effectively define provenance in Holocene and Neogene fluvial sands from the Snake River system of the northern Rockies, U.S.A. SHRIMP U-Pb dates have been measured for forty-six samples (about 2700 zircon grains) of fluvial and aeolian sediment. The detrital-zircon age distributions are repeatable and demonstrate predictable longitudinal variation. By lumping multiple samples to attain populations of several hundred grains, we recognize distinctive, provenance-defining zircon-age distributions or "barcodes," for fluvial sedimentary systems of several scales, within the upper and middle Snake River system. Our detrital-zircon studies effectively define the geochronology of the northern Rocky Mountains. The composite detrital-zircon grain distribution of the middle Snake River consists of major populations of Neogene, Eocene, and Cretaceous magmatic grains plus intermediate and small grain populations of multiply recycled Grenville (˜950 to 1300 Ma) grains and Yavapai-Mazatzal province grains (˜1600 to 1800 Ma) recycled through the upper Belt Supergroup and Cretaceous sandstones. A wide range of older Paleoproterozoic and Archean grains are also present. The best-case scenario for using detrital-zircon populations to isolate provenance is when there is a point-source pluton with known age, that is only found in one location or drainage. We find three such zircon age-populations in fluvial sediments downstream from the point-source plutons: Ordovician in the southern Beaverhead Mountains, Jurassic in northern Nevada, and Oligocene in the Albion Mountains core complex of southern Idaho. Large detrital-zircon age-populations derived from regionally well-defined, magmatic or recycled sedimentary, sources also serve to delimit the provenance of Neogene fluvial systems. In the Snake River system, defining populations include those derived from Cretaceous Atlanta lobe of the Idaho batholith (80 to 100 Ma), Eocene Challis Volcanic Group and associated plutons (˜45 to 52 Ma), and Neogene rhyolitic Yellowstone-Snake River Plain volcanics (˜0 to 17 Ma). For first-order drainage basins containing these zircon-rich source terranes, or containing a point-source pluton, a 60-grain random sample is sufficient to define the dominant provenance. The most difficult age-distributions to analyze are those that contain multiple small zircon age-populations and no defining large populations. Examples of these include streams draining the Proterozoic and Paleozoic Cordilleran miogeocline in eastern Idaho and Pleistocene loess on the Snake River Plain. For such systems, large sample bases of hundreds of grains, plus the use of statistical methods, may be necessary to distinguish detrital-zircon age-spectra.

  14. Statistics 101 for Radiologists.

    PubMed

    Anvari, Arash; Halpern, Elkan F; Samir, Anthony E

    2015-10-01

    Diagnostic tests have wide clinical applications, including screening, diagnosis, measuring treatment effect, and determining prognosis. Interpreting diagnostic test results requires an understanding of key statistical concepts used to evaluate test efficacy. This review explains descriptive statistics and discusses probability, including mutually exclusive and independent events and conditional probability. In the inferential statistics section, a statistical perspective on study design is provided, together with an explanation of how to select appropriate statistical tests. Key concepts in recruiting study samples are discussed, including representativeness and random sampling. Variable types are defined, including predictor, outcome, and covariate variables, and the relationship of these variables to one another. In the hypothesis testing section, we explain how to determine if observed differences between groups are likely to be due to chance. We explain type I and II errors, statistical significance, and study power, followed by an explanation of effect sizes and how confidence intervals can be used to generalize observed effect sizes to the larger population. Statistical tests are explained in four categories: t tests and analysis of variance, proportion analysis tests, nonparametric tests, and regression techniques. We discuss sensitivity, specificity, accuracy, receiver operating characteristic analysis, and likelihood ratios. Measures of reliability and agreement, including κ statistics, intraclass correlation coefficients, and Bland-Altman graphs and analysis, are introduced. © RSNA, 2015.

  15. Convergence and Efficiency of Adaptive Importance Sampling Techniques with Partial Biasing

    NASA Astrophysics Data System (ADS)

    Fort, G.; Jourdain, B.; Lelièvre, T.; Stoltz, G.

    2018-04-01

    We propose a new Monte Carlo method to efficiently sample a multimodal distribution (known up to a normalization constant). We consider a generalization of the discrete-time Self Healing Umbrella Sampling method, which can also be seen as a generalization of well-tempered metadynamics. The dynamics is based on an adaptive importance technique. The importance function relies on the weights (namely the relative probabilities) of disjoint sets which form a partition of the space. These weights are unknown but are learnt on the fly yielding an adaptive algorithm. In the context of computational statistical physics, the logarithm of these weights is, up to an additive constant, the free-energy, and the discrete valued function defining the partition is called the collective variable. The algorithm falls into the general class of Wang-Landau type methods, and is a generalization of the original Self Healing Umbrella Sampling method in two ways: (i) the updating strategy leads to a larger penalization strength of already visited sets in order to escape more quickly from metastable states, and (ii) the target distribution is biased using only a fraction of the free-energy, in order to increase the effective sample size and reduce the variance of importance sampling estimators. We prove the convergence of the algorithm and analyze numerically its efficiency on a toy example.

  16. TNO/Centaurs grouping tested with asteroid data sets

    NASA Astrophysics Data System (ADS)

    Fulchignoni, M.; Birlan, M.; Barucci, M. A.

    2001-11-01

    Recently, we have discussed the possible subdivision in few groups of a sample of 22 TNO and Centaurs for which the BVRIJ photometry were available (Barucci et al., 2001, A&A, 371,1150). We obtained this results using the multivariate statistics adopted to define the current asteroid taxonomy, namely the Principal Components Analysis and the G-mode method (Tholen & Barucci, 1989, in ASTEROIDS II). How these methods work with a very small statistical sample as the TNO/Centaurs one? Theoretically, the number of degrees of freedom of the sample is correct. In fact it is 88 in our case and have to be larger then 50 to cope with the requirements of the G-mode. Does the random sampling of the small number of members of a large population contain enough information to reveal some structure in the population? We extracted several samples of 22 asteroids out of a data-base of 86 objects of known taxonomic type for which BVRIJ photometry is available from ECAS (Zellner et al. 1985, ICARUS 61, 355), SMASS II (S.W. Bus, 1999, PhD Thesis, MIT), and the Bell et al. Atlas of the asteroid infrared spectra. The objects constituting the first sample were selected in order to give a good representation of the major asteroid taxonomic classes (at least three samples each class): C,S,D,A, and G. Both methods were able to distinguish all these groups confirming the validity of the adopted methods. The S class is hard to individuate as a consequence of the choice of I and J variables, which imply a lack of information on the absorption band at 1 micron. The other samples were obtained by random choice of the objects. Not all the major groups were well represented (less than three samples per groups), but the general trend of the asteroid taxonomy has been always obtained. We conclude that the quoted grouping of TNO/Centaurs is representative of some physico-chemical structure of the outer solar system small body population.

  17. [Situational low self-esteem in pregnant women: an analysis of accuracy].

    PubMed

    Cavalcante, Joyce Carolle Bezerra; de Sousa, Vanessa Emille Carvalho; Lopes, Marcos Venícios de Oliveira

    2012-01-01

    To investigate the accuracy of defining characteristics of Situational low self-esteem we developed a cross-sectional study, with 52 pregnant women assisted in a family centre. The NANDA-I taxonomy was used as well as the Rosenberg's scale. The diagnosis was present in 32.7% of the sample and all characteristics presented statistical significance, except "Reports verbally situational challenge to its own value". The characteristics "Indecisive behavior" and "Helplessness expressions" had 82.35% of sensitivity. On the other hand, the characteristics "Expression of feelings of worthlessness" and "Reports verbally situational challenge to its own value" were the more specific, with 94.29% of specificity. These results can contribute with the nursing practice because the identification of accurate characteristics is essential to a secure inference.

  18. Binding of carboxylate and trimethylammonium salts to octa-acid and TEMOA deep-cavity cavitands

    NASA Astrophysics Data System (ADS)

    Sullivan, Matthew R.; Sokkalingam, Punidha; Nguyen, Thong; Donahue, James P.; Gibb, Bruce C.

    2017-01-01

    In participation of the fifth statistical assessment of modeling of proteins and ligands (SAMPL5), the strength of association of six guests ( 3- 8) to two hosts ( 1 and 2) were measured by 1H NMR and ITC. Each host possessed a unique and well-defined binding pocket, whilst the wide array of amphiphilic guests possessed binding moieties that included: a terminal alkyne, nitro-arene, alkyl halide and cyano-arene groups. Solubilizing head groups for the guests included both positively charged trimethylammonium and negatively charged carboxylate functionality. Measured association constants ( K a ) covered five orders of magnitude, ranging from 56 M-1 for guest 6 binding with host 2 up to 7.43 × 106 M-1 for guest 6 binding to host 1.

  19. Time series, periodograms, and significance

    NASA Astrophysics Data System (ADS)

    Hernandez, G.

    1999-05-01

    The geophysical literature shows a wide and conflicting usage of methods employed to extract meaningful information on coherent oscillations from measurements. This makes it difficult, if not impossible, to relate the findings reported by different authors. Therefore, we have undertaken a critical investigation of the tests and methodology used for determining the presence of statistically significant coherent oscillations in periodograms derived from time series. Statistical significance tests are only valid when performed on the independent frequencies present in a measurement. Both the number of possible independent frequencies in a periodogram and the significance tests are determined by the number of degrees of freedom, which is the number of true independent measurements, present in the time series, rather than the number of sample points in the measurement. The number of degrees of freedom is an intrinsic property of the data, and it must be determined from the serial coherence of the time series. As part of this investigation, a detailed study has been performed which clearly illustrates the deleterious effects that the apparently innocent and commonly used processes of filtering, de-trending, and tapering of data have on periodogram analysis and the consequent difficulties in the interpretation of the statistical significance thus derived. For the sake of clarity, a specific example of actual field measurements containing unevenly-spaced measurements, gaps, etc., as well as synthetic examples, have been used to illustrate the periodogram approach, and pitfalls, leading to the (statistical) significance tests for the presence of coherent oscillations. Among the insights of this investigation are: (1) the concept of a time series being (statistically) band limited by its own serial coherence and thus having a critical sampling rate which defines one of the necessary requirements for the proper statistical design of an experiment; (2) the design of a critical test for the maximum number of significant frequencies which can be used to describe a time series, while retaining intact the variance of the test sample; (3) a demonstration of the unnecessary difficulties that manipulation of the data brings into the statistical significance interpretation of said data; and (4) the resolution and correction of the apparent discrepancy in significance results obtained by the use of the conventional Lomb-Scargle significance test, when compared with the long-standing Schuster-Walker and Fisher tests.

  20. Novel Microbiological and Spatial Statistical Methods to Improve Strength of Epidemiological Evidence in a Community-Wide Waterborne Outbreak

    PubMed Central

    Jalava, Katri; Rintala, Hanna; Ollgren, Jukka; Maunula, Leena; Gomez-Alvarez, Vicente; Revez, Joana; Palander, Marja; Antikainen, Jenni; Kauppinen, Ari; Räsänen, Pia; Siponen, Sallamaari; Nyholm, Outi; Kyyhkynen, Aino; Hakkarainen, Sirpa; Merentie, Juhani; Pärnänen, Martti; Loginov, Raisa; Ryu, Hodon; Kuusi, Markku; Siitonen, Anja; Miettinen, Ilkka; Santo Domingo, Jorge W.; Hänninen, Marja-Liisa; Pitkänen, Tarja

    2014-01-01

    Failures in the drinking water distribution system cause gastrointestinal outbreaks with multiple pathogens. A water distribution pipe breakage caused a community-wide waterborne outbreak in Vuorela, Finland, July 2012. We investigated this outbreak with advanced epidemiological and microbiological methods. A total of 473/2931 inhabitants (16%) responded to a web-based questionnaire. Water and patient samples were subjected to analysis of multiple microbial targets, molecular typing and microbial community analysis. Spatial analysis on the water distribution network was done and we applied a spatial logistic regression model. The course of the illness was mild. Drinking untreated tap water from the defined outbreak area was significantly associated with illness (RR 5.6, 95% CI 1.9–16.4) increasing in a dose response manner. The closer a person lived to the water distribution breakage point, the higher the risk of becoming ill. Sapovirus, enterovirus, single Campylobacter jejuni and EHEC O157:H7 findings as well as virulence genes for EPEC, EAEC and EHEC pathogroups were detected by molecular or culture methods from the faecal samples of the patients. EPEC, EAEC and EHEC virulence genes and faecal indicator bacteria were also detected in water samples. Microbial community sequencing of contaminated tap water revealed abundance of Arcobacter species. The polyphasic approach improved the understanding of the source of the infections, and aided to define the extent and magnitude of this outbreak. PMID:25147923

  1. Genome-wide association analysis of secondary imaging phenotypes from the Alzheimer's disease neuroimaging initiative study.

    PubMed

    Zhu, Wensheng; Yuan, Ying; Zhang, Jingwen; Zhou, Fan; Knickmeyer, Rebecca C; Zhu, Hongtu

    2017-02-01

    The aim of this paper is to systematically evaluate a biased sampling issue associated with genome-wide association analysis (GWAS) of imaging phenotypes for most imaging genetic studies, including the Alzheimer's Disease Neuroimaging Initiative (ADNI). Specifically, the original sampling scheme of these imaging genetic studies is primarily the retrospective case-control design, whereas most existing statistical analyses of these studies ignore such sampling scheme by directly correlating imaging phenotypes (called the secondary traits) with genotype. Although it has been well documented in genetic epidemiology that ignoring the case-control sampling scheme can produce highly biased estimates, and subsequently lead to misleading results and suspicious associations, such findings are not well documented in imaging genetics. We use extensive simulations and a large-scale imaging genetic data analysis of the Alzheimer's Disease Neuroimaging Initiative (ADNI) data to evaluate the effects of the case-control sampling scheme on GWAS results based on some standard statistical methods, such as linear regression methods, while comparing it with several advanced statistical methods that appropriately adjust for the case-control sampling scheme. Copyright © 2016 Elsevier Inc. All rights reserved.

  2. Statistical benchmark for BosonSampling

    NASA Astrophysics Data System (ADS)

    Walschaers, Mattia; Kuipers, Jack; Urbina, Juan-Diego; Mayer, Klaus; Tichy, Malte Christopher; Richter, Klaus; Buchleitner, Andreas

    2016-03-01

    Boson samplers—set-ups that generate complex many-particle output states through the transmission of elementary many-particle input states across a multitude of mutually coupled modes—promise the efficient quantum simulation of a classically intractable computational task, and challenge the extended Church-Turing thesis, one of the fundamental dogmas of computer science. However, as in all experimental quantum simulations of truly complex systems, one crucial problem remains: how to certify that a given experimental measurement record unambiguously results from enforcing the claimed dynamics, on bosons, fermions or distinguishable particles? Here we offer a statistical solution to the certification problem, identifying an unambiguous statistical signature of many-body quantum interference upon transmission across a multimode, random scattering device. We show that statistical analysis of only partial information on the output state allows to characterise the imparted dynamics through particle type-specific features of the emerging interference patterns. The relevant statistical quantifiers are classically computable, define a falsifiable benchmark for BosonSampling, and reveal distinctive features of many-particle quantum dynamics, which go much beyond mere bunching or anti-bunching effects.

  3. Krypton and xenon in lunar fines

    NASA Technical Reports Server (NTRS)

    Basford, J. R.; Dragon, J. C.; Pepin, R. O.; Coscio, M. R., Jr.; Murthy, V. R.

    1973-01-01

    Data from grain-size separates, stepwise-heated fractions, and bulk analyses of 20 samples of fines and breccias from five lunar sites are used to define three-isotope and ordinate intercept correlations in an attempt to resolve the lunar heavy rare gas system in a statistically valid approach. Tables of concentrations and isotope compositions are given.

  4. Substance Abuse Counselors and Moral Reasoning: Hypothetical and Authentic Dilemmas

    ERIC Educational Resources Information Center

    Sias, Shari M.

    2009-01-01

    This exploratory study examined the assumption that the level of moral reasoning (Defining Issues Test; J. R. Rest, 1986) used in solving hypothetical and authentic dilemmas is similar for substance abuse counselors (N = 188). The statistical analyses used were paired-sample t tests, Pearson product-moment correlation, and simultaneous multiple…

  5. Recent Reliability Reporting Practices in "Psychological Assessment": Recognizing the People behind the Data

    ERIC Educational Resources Information Center

    Green, Carlton E.; Chen, Cynthia E.; Helms, Janet E.; Henze, Kevin T.

    2011-01-01

    Helms, Henze, Sass, and Mifsud (2006) defined good practices for internal consistency reporting, interpretation, and analysis consistent with an alpha-as-data perspective. Their viewpoint (a) expands on previous arguments that reliability coefficients are group-level summary statistics of samples' responses rather than stable properties of scales…

  6. Chemical quality of bottom sediments in selected streams, Jefferson County, Kentucky, April-July 1992

    USGS Publications Warehouse

    Moore, B.L.; Evaldi, R.D.

    1995-01-01

    Bottom sediments from 25 stream sites in Jefferson County, Ky., were analyzed for percent volatile solids and concentrations of nutrients, major metals, trace elements, miscellaneous inorganic compounds, and selected organic compounds. Statistical high outliers of the constituent concentrations analyzed for in the bottom sediments were defined as a measure of possible elevated concentrations. Statistical high outliers were determined for at least 1 constituent at each of 12 sampling sites in Jefferson County. Of the 10 stream basins sampled in Jefferson County, the Middle Fork Beargrass Basin, Cedar Creek Basin, and Harrods Creek Basin were the only three basins where a statistical high outlier was not found for any of the measured constituents. In the Pennsylvania Run Basin, total volatile solids, nitrate plus nitrite, and endrin constituents were statistical high outliers. Pond Creek was the only basin where five constituents were statistical high outliers-barium, beryllium, cadmium, chromium, and silver. Nitrate plus nitrite and copper constituents were the only statistical high outliers found in the Mill Creek Basin. In the Floyds Fork Basin, nitrate plus nitrite, phosphorus, mercury, and silver constituents were the only statistical high outliers. Ammonia was the only statistical high outlier found in the South Fork Beargrass Basin. In the Goose Creek Basin, mercury and silver constituents were the only statistical high outliers. Cyanide was the only statistical high outlier in the Muddy Fork Basin.

  7. The Statistics of Visual Representation

    NASA Technical Reports Server (NTRS)

    Jobson, Daniel J.; Rahman, Zia-Ur; Woodell, Glenn A.

    2002-01-01

    The experience of retinex image processing has prompted us to reconsider fundamental aspects of imaging and image processing. Foremost is the idea that a good visual representation requires a non-linear transformation of the recorded (approximately linear) image data. Further, this transformation appears to converge on a specific distribution. Here we investigate the connection between numerical and visual phenomena. Specifically the questions explored are: (1) Is there a well-defined consistent statistical character associated with good visual representations? (2) Does there exist an ideal visual image? And (3) what are its statistical properties?

  8. Risk analysis in cohort studies with heterogeneous strata. A global chi2-test for dose-response relationship, generalizing the Mantel-Haenszel procedure.

    PubMed

    Ahlborn, W; Tuz, H J; Uberla, K

    1990-03-01

    In cohort studies the Mantel-Haenszel estimator ORMH is computed from sample data and is used as a point estimator of relative risk. Test-based confidence intervals are estimated with the help of the asymptotic chi-squared distributed MH-statistic chi 2MHS. The Mantel-extension-chi-squared is used as a test statistic for a dose-response relationship. Both test statistics--the Mantel-Haenszel-chi as well as the Mantel-extension-chi--assume homogeneity of risk across strata, which is rarely present. Also an extended nonparametric statistic, proposed by Terpstra, which is based on the Mann-Whitney-statistics assumes homogeneity of risk across strata. We have earlier defined four risk measures RRkj (k = 1,2,...,4) in the population and considered their estimates and the corresponding asymptotic distributions. In order to overcome the homogeneity assumption we use the delta-method to get "test-based" confidence intervals. Because the four risk measures RRkj are presented as functions of four weights gik we give, consequently, the asymptotic variances of these risk estimators also as functions of the weights gik in a closed form. Approximations to these variances are given. For testing a dose-response relationship we propose a new class of chi 2(1)-distributed global measures Gk and the corresponding global chi 2-test. In contrast to the Mantel-extension-chi homogeneity of risk across strata must not be assumed. These global test statistics are of the Wald type for composite hypotheses.(ABSTRACT TRUNCATED AT 250 WORDS)

  9. Evaluation of a New Mean Scaled and Moment Adjusted Test Statistic for SEM

    ERIC Educational Resources Information Center

    Tong, Xiaoxiao; Bentler, Peter M.

    2013-01-01

    Recently a new mean scaled and skewness adjusted test statistic was developed for evaluating structural equation models in small samples and with potentially nonnormal data, but this statistic has received only limited evaluation. The performance of this statistic is compared to normal theory maximum likelihood and 2 well-known robust test…

  10. Sample Skewness as a Statistical Measurement of Neuronal Tuning Sharpness

    PubMed Central

    Samonds, Jason M.; Potetz, Brian R.; Lee, Tai Sing

    2014-01-01

    We propose using the statistical measurement of the sample skewness of the distribution of mean firing rates of a tuning curve to quantify sharpness of tuning. For some features, like binocular disparity, tuning curves are best described by relatively complex and sometimes diverse functions, making it difficult to quantify sharpness with a single function and parameter. Skewness provides a robust nonparametric measure of tuning curve sharpness that is invariant with respect to the mean and variance of the tuning curve and is straightforward to apply to a wide range of tuning, including simple orientation tuning curves and complex object tuning curves that often cannot even be described parametrically. Because skewness does not depend on a specific model or function of tuning, it is especially appealing to cases of sharpening where recurrent interactions among neurons produce sharper tuning curves that deviate in a complex manner from the feedforward function of tuning. Since tuning curves for all neurons are not typically well described by a single parametric function, this model independence additionally allows skewness to be applied to all recorded neurons, maximizing the statistical power of a set of data. We also compare skewness with other nonparametric measures of tuning curve sharpness and selectivity. Compared to these other nonparametric measures tested, skewness is best used for capturing the sharpness of multimodal tuning curves defined by narrow peaks (maximum) and broad valleys (minima). Finally, we provide a more formal definition of sharpness using a shape-based information gain measure and derive and show that skewness is correlated with this definition. PMID:24555451

  11. A robust clustering algorithm for identifying problematic samples in genome-wide association studies.

    PubMed

    Bellenguez, Céline; Strange, Amy; Freeman, Colin; Donnelly, Peter; Spencer, Chris C A

    2012-01-01

    High-throughput genotyping arrays provide an efficient way to survey single nucleotide polymorphisms (SNPs) across the genome in large numbers of individuals. Downstream analysis of the data, for example in genome-wide association studies (GWAS), often involves statistical models of genotype frequencies across individuals. The complexities of the sample collection process and the potential for errors in the experimental assay can lead to biases and artefacts in an individual's inferred genotypes. Rather than attempting to model these complications, it has become a standard practice to remove individuals whose genome-wide data differ from the sample at large. Here we describe a simple, but robust, statistical algorithm to identify samples with atypical summaries of genome-wide variation. Its use as a semi-automated quality control tool is demonstrated using several summary statistics, selected to identify different potential problems, and it is applied to two different genotyping platforms and sample collections. The algorithm is written in R and is freely available at www.well.ox.ac.uk/chris-spencer chris.spencer@well.ox.ac.uk Supplementary data are available at Bioinformatics online.

  12. Clinical competence of Guatemalan and Mexican physicians for family dysfunction management.

    PubMed

    Cabrera-Pivaral, Carlos Enrique; Orozco-Valerio, María de Jesús; Celis-de la Rosa, Alfredo; Covarrubias-Bermúdez, María de Los Ángeles; Zavala-González, Marco Antonio

    2017-01-01

    To evaluate the clinical competence of Mexican and Guatemalan physicians to management the family dysfunction. Cross comparative study in four care units first in Guadalajara, Mexico, and four in Guatemala, Guatemala, based on a purposeful sampling, involving 117 and 100 physicians, respectively. Clinical competence evaluated by validated instrument integrated for 187 items. Non-parametric descriptive and inferential statistical analysis was performed. The percentage of Mexican physicians with high clinical competence was 13.7%, medium 53%, low 24.8% and defined by random 8.5%. For the Guatemalan physicians'14% was high, average 63%, and 23% defined by random. There were no statistically significant differences between healthcare country units, but between the medium of Mexicans (0.55) and Guatemalans (0.55) (p = 0.02). The proportion of the high clinical competency of Mexican physicians' was as Guatemalans.

  13. Reliable mortality statistics for Turkey: Are we there yet?

    PubMed

    Özdemir, Raziye; Rao, Chalapati; Öcek, Zeliha; Dinç Horasan, Gönül

    2015-06-10

    The Turkish government has implemented several reforms to improve the Turkish Statistical Institute Death Reporting System (TURKSTAT-DRS) since 2009. However, there has been no assessment to evaluate the impact of these reforms on causes of death statistics. This study attempted to analyse the impact of these reforms on the TURKSTAT-DRS for Turkey, and in the case of Izmir, one of the most developed provinces in Turkey. The evaluation framework comprised three main components each with specific criteria. Firstly, data from TURKSTAT for Turkey and Izmir for the periods 2001-2008 and 2009-2013 were assessed in terms of the following dimensions that represent quality of mortality statistics (a. completeness of death registration, b. trends in proportions of deaths with ill-defined causes). Secondly, the quality of information recorded on individual death certificates from Izmir in 2010 was analysed for a. missing information, b. timeliness of death notifications and c. characteristics of deaths with ill-defined causes. Finally, TURKSTAT data were analysed to estimate life tables and summary mortality indicators for Turkey and Izmir, as well as the leading causes-of-death in Turkey in 2013. Registration of adult deaths in Izmir as well as at the national level for Turkey has considerably improved since the introduction of reforms in 2009, along with marked decline in the proportions of deaths assigned ill-defined causes. Death certificates from Izmir indicated significant gaps in recorded information for demographic as well as epidemiological variables, particularly for infant deaths, and in the detailed recording of causes of death. Life expectancy at birth estimated from local data is 3-4 years higher than similar estimates for Turkey from international studies, and this requires further investigation and confirmation. The TURKSTAT-DRS is now an improved source of mortality and cause of death statistics for Turkey. The reliability and validity of TURKSTAT data needs to be established through a detailed research program to evaluate completeness of death registration and validity of registered causes of death. Similar evaluation and data analysis of mortality indicators is required at regular intervals at national and sub-national level, to increase confidence in their utility as primary data for epidemiology and health policy.

  14. Optimizing image registration and infarct definition in stroke research.

    PubMed

    Harston, George W J; Minks, David; Sheerin, Fintan; Payne, Stephen J; Chappell, Michael; Jezzard, Peter; Jenkinson, Mark; Kennedy, James

    2017-03-01

    Accurate representation of final infarct volume is essential for assessing the efficacy of stroke interventions in imaging-based studies. This study defines the impact of image registration methods used at different timepoints following stroke, and the implications for infarct definition in stroke research. Patients presenting with acute ischemic stroke were imaged serially using magnetic resonance imaging. Infarct volume was defined manually using four metrics: 24-h b1000 imaging; 1-week and 1-month T2-weighted FLAIR; and automatically using predefined thresholds of ADC at 24 h. Infarct overlap statistics and volumes were compared across timepoints following both rigid body and nonlinear image registration to the presenting MRI. The effect of nonlinear registration on a hypothetical trial sample size was calculated. Thirty-seven patients were included. Nonlinear registration improved infarct overlap statistics and consistency of total infarct volumes across timepoints, and reduced infarct volumes by 4.0 mL (13.1%) and 7.1 mL (18.2%) at 24 h and 1 week, respectively, compared to rigid body registration. Infarct volume at 24 h, defined using a predetermined ADC threshold, was less sensitive to infarction than b1000 imaging. 1-week T2-weighted FLAIR imaging was the most accurate representation of final infarct volume. Nonlinear registration reduced hypothetical trial sample size, independent of infarct volume, by an average of 13%. Nonlinear image registration may offer the opportunity of improving the accuracy of infarct definition in serial imaging studies compared to rigid body registration, helping to overcome the challenges of anatomical distortions at subacute timepoints, and reducing sample size for imaging-based clinical trials.

  15. Limit order book and its modeling in terms of Gibbs Grand-Canonical Ensemble

    NASA Astrophysics Data System (ADS)

    Bicci, Alberto

    2016-12-01

    In the domain of so called Econophysics some attempts have been already made for applying the theory of thermodynamics and statistical mechanics to economics and financial markets. In this paper a similar approach is made from a different perspective, trying to model the limit order book and price formation process of a given stock by the Grand-Canonical Gibbs Ensemble for the bid and ask orders. The application of the Bose-Einstein statistics to this ensemble allows then to derive the distribution of the sell and buy orders as a function of price. As a consequence we can define in a meaningful way expressions for the temperatures of the ensembles of bid orders and of ask orders, which are a function of minimum bid, maximum ask and closure prices of the stock as well as of the exchanged volume of shares. It is demonstrated that the difference between the ask and bid orders temperatures can be related to the VAO (Volume Accumulation Oscillator), an indicator empirically defined in Technical Analysis of stock markets. Furthermore the derived distributions for aggregate bid and ask orders can be subject to well defined validations against real data, giving a falsifiable character to the model.

  16. Dynamics and Statistical Mechanics of Rotating and non-Rotating Vortical Flows

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lim, Chjan

    Three projects were analyzed with the overall aim of developing a computational/analytical model for estimating values of the energy, angular momentum, enstrophy and total variation of fluid height at phase transitions between disordered and self-organized flow states in planetary atmospheres. It is believed that these transitions in equilibrium statistical mechanics models play a role in the construction of large-scale, stable structures including super-rotation in the Venusian atmosphere and the formation of the Great Red Spot on Jupiter. Exact solutions of the spherical energy-enstrophy models for rotating planetary atmospheres by Kac's method of steepest descent predicted phase transitions to super-rotating solid-bodymore » flows at high energy to enstrophy ratio for all planetary spins and to sub-rotating modes if the planetary spin is large enough. These canonical statistical ensembles are well-defined for the long-range energy interactions that arise from 2D fluid flows on compact oriented manifolds such as the surface of the sphere and torus. This is because in Fourier space available through Hodge theory, the energy terms are exactly diagonalizable and hence has zero range, leading to well-defined heat baths.« less

  17. Remote sensing-aided systems for snow qualification, evapotranspiration estimation, and their application in hydrologic models

    NASA Technical Reports Server (NTRS)

    Korram, S.

    1977-01-01

    The design of general remote sensing-aided methodologies was studied to provide the estimates of several important inputs to water yield forecast models. These input parameters are snow area extent, snow water content, and evapotranspiration. The study area is Feather River Watershed (780,000 hectares), Northern California. The general approach involved a stepwise sequence of identification of the required information, sample design, measurement/estimation, and evaluation of results. All the relevent and available information types needed in the estimation process are being defined. These include Landsat, meteorological satellite, and aircraft imagery, topographic and geologic data, ground truth data, and climatic data from ground stations. A cost-effective multistage sampling approach was employed in quantification of all the required parameters. The physical and statistical models for both snow quantification and evapotranspiration estimation was developed. These models use the information obtained by aerial and ground data through appropriate statistical sampling design.

  18. Complement Activation on Platelets Correlates with a Decrease in Circulating Immature Platelets in Patients with Immune Thrombocytopenic Purpura

    PubMed Central

    Peerschke, Ellinor I.B.; Andemariam, Biree; Yin, Wei; Bussel, James B.

    2010-01-01

    The role of the complement system in immune thrombocytopenic purpura (ITP) is not well defined. We examined plasma from 79 patients with ITP, 50 healthy volunteers, and 25 patients with non-immune mediated thrombocytopenia, to investigate their complement activation/fixation capacity (CAC) on immobilized heterologous platelets. Enhanced CAC was found in 46 plasma samples (59%) from patients with ITP, but no samples from patients with non-immune mediated thrombocytopenia. Plasma from healthy volunteers was used for comparison. In patients with ITP, an enhanced plasma CAC was associated with a decreased circulating absolute immature platelet fraction (A-IPF) (<15 × 109/L) (p = 0.027) and thrombocytopenia (platelet count less than 100K/μl) (p= 0.024). The positive predictive value of an enhanced CAC for a low A-IPF was 93%, with a specificity of 77%. The specificity and positive predictive values increased to 100% when plasma CAC was defined strictly by enhanced C1q and/or C4d deposition on test platelets. Although no statistically significant correlation emerged between CAC and response to different pharmacologic therapies, an enhanced response to splenectomy was noted (p <0.063). Thus, complement fixation may contribute to the thrombocytopenia of ITP by enhancing clearance of opsonized platelets from the circulation, and/or directly damaging platelets and megakaryocytes. PMID:19925495

  19. Detrital dating of Asian orogenesis: insights and caveats

    NASA Astrophysics Data System (ADS)

    Burbank, D. W.

    2007-12-01

    Technological advances over the past two decades have facilitated increasingly routine application of single- crystal dating and cosmogenic nuclide dating to studies of orogenic erosion. Both approaches commonly utilize grab samples of detrital sediment, either modern or ancient. Whereas detrital cosmogenic data are typically used to define mean erosion rates for upstream catchments, single-crystal ages are used both to discern provenance and to define lag times: interval between isotopic closure and deposition. Recent results from dating modern fluvial sediments illuminate key concepts that underpin interpretations of results from older strata: the fidelity of the detrital signal, its evolution through an orogen, its relationship to discrete source areas, and its temporal evolution. Despite the increasing availability of dates and rates for detrial grains, relatively few studies have addressed the sources of uncertainty that modulate the precision and accuracy with which detrital results should be interpreted. Such uncertainties derive not only from sampling statistics and measurement uncertainties, but also from both geomorphic sources (seasonal variation in sediment supply and source, changes in glacial cover, the impact of stochastic geomorphic events, such as landslides), as well as tectonic ones (time-dependent deformation and thermal models, particle paths through the orogen). A better understanding of the impact of these uncertainties will underpin more reliable and less speculative interpretations of future dating results from both ancient and modern detrital fluvial sediments.

  20. THEMATIC ACCURACY OF THE 1992 NATIONAL LAND-COVER DATA (NLCD) FOR THE EASTERN UNITED STATES: STATISTICAL METHODOLOGY AND REGIONAL RESULTS

    EPA Science Inventory

    The accuracy of the National Land Cover Data (NLCD) map is assessed via a probability sampling design incorporating three levels of stratification and two stages of selection. Agreement between the map and reference land-cover labels is defined as a match between the primary or a...

  1. Detection of coliform bacteria and Escherichia coli by multiplex polymerase chain reaction: comparison with defined substrate and plating methods for water quality monitoring.

    PubMed Central

    Bej, A K; McCarty, S C; Atlas, R M

    1991-01-01

    Multiplex polymerase chain reaction (PCR) and gene probe detection of target lacZ and uidA genes were used to detect total coliform bacteria and Escherichia coli, respectively, for determining water quality. In tests of environmental water samples, the lacZ PCR method gave results statistically equivalent to those of the plate count and defined substrate methods accepted by the U.S. Environmental Protection Agency for water quality monitoring and the uidA PCR method was more sensitive than 4-methylumbelliferyl-beta-D-glucuronide-based defined substrate tests for specific detection of E. coli. Images PMID:1768116

  2. The Bootstrap, the Jackknife, and the Randomization Test: A Sampling Taxonomy.

    PubMed

    Rodgers, J L

    1999-10-01

    A simple sampling taxonomy is defined that shows the differences between and relationships among the bootstrap, the jackknife, and the randomization test. Each method has as its goal the creation of an empirical sampling distribution that can be used to test statistical hypotheses, estimate standard errors, and/or create confidence intervals. Distinctions between the methods can be made based on the sampling approach (with replacement versus without replacement) and the sample size (replacing the whole original sample versus replacing a subset of the original sample). The taxonomy is useful for teaching the goals and purposes of resampling schemes. An extension of the taxonomy implies other possible resampling approaches that have not previously been considered. Univariate and multivariate examples are presented.

  3. Attitudes towards euthanasia.

    PubMed Central

    Winget, C; Kapp, F T; Yeaworth, R C

    1977-01-01

    There are an infinite variety of attitudes to euthanasia, each individual response to the concept being influenced by many factors. Consequently there is a literature on the subject ranging from the popular article to papers in specialized journals. This study, however, has taken a well defined sample of people, inviting them to answer a questionnaire which was designed to elicit their attitudes to euthanasia in a way which could be analysed statistically. Nor surprisingly attitudes appeared to 'harden' as those answering the questionnaire grew more experienced in dealing with patients and also more professionally established. Thus it was found that of the seven groups questioned practising physicians showed more positive attitudes to euthanasia and their responses did not differ significantly from those of senior medical students. It is these groups which actually or potentially have to resolve the clinical dilemma posed by the dying patient. PMID:859163

  4. Comparison of the effects of filtration and preservation methods on analyses for strontium-90 in ground water

    USGS Publications Warehouse

    Knobel, L.L.; DeWayne, Cecil L.; Wegner, S.J.; Moore, L.L.

    1992-01-01

    From 1952 to 1988, about 140 curies of strontium-90 were discharged in liquid waste to disposal ponds and wells at the INEL (Idaho National Engineering Laboratory). Water from four wells was sampled as part of the U.S. Geological Survey's quality-assurance program to evaluate the effects of filtration and preservation methods on strontium-90 concentrations in ground water at the INEL. Water from each well was filtered through eithera 0.45- or a 0.1-micrometer membrane filter; unfiltered samples also were collected. Two sets of filtered and two sets of unfiltered water samples were collected at each well. One of the two sets of water samples was field acidified. Strontium-90 concentrations ranged from below the reporting level to 52 ?? 4 picocuries per liter. Descriptive statistics were used to determine reproducibility of the analytical results for strontium-90 concentrations in water from each well. Comparisons were made with unfiltered, acidified samples at each well. Analytical results for strontium-90 concentrations in water from well 88 were not in statistical agreement between the unfiltered, acidified sample and the filtered (0.45 micrometer), acidified sample. The strontium-90 concentration for water from well 88 was less than the reporting level. For water from wells with strontium-90 concentrations at or above the reporting level, 94 percent or more of the strontium-90 is in true solution or in colloidal particles smaller than 0.1 micrometer. These results suggest that changes in filtration and preservation methods used for sample collection do not significantly affect reproducibility of strontium-90 analyses in ground water at the INEL.

  5. Comparison of the effects of filtration and preservation methods on analyses for strontium-90 in ground water.

    PubMed

    Knobel, L L; Cecil, L D; Wegner, S J; Moore, L L

    1992-01-01

    From 1952 to 1988, about 140 curies of strontium-90 were discharged in liquid waste to disposal ponds and wells at the INEL (Idaho National Engineering Laboratory). Water from four wells was sampled as part of the U.S. Geological Survey's quality-assurance program to evaluate the effects of filtration and preservation methods on strontium-90 concentrations in ground water at the INEL. Water from each well was filtered through either a 0.45- or a 0.1-micrometer membrane filter; unfiltered samples also were collected. Two sets of filtered and two sets of unfiltered water samples were collected at each well. One of the two sets of water samples was field acidified.Strontium-90 concentrations ranged from below the reporting level to 52±4 picocuries per liter. Descriptive statistics were used to determine reproducibility of the analytical results for strontium-90 concentrations in water from each well. Comparisons were made with unfiltered, acidified samples at each well. Analytical results for strontium-90 concentrations in water from well 88 were not in statistical agreement between the unfiltered, acidified sample and the filtered (0.45 micrometer), acidified sample. The strontium-90 concentration for water from well 88 was less than the reporting level.For water from wells with strontium-90 concentrations at or above the reporting level, 94 percent or more of the strontium-90 is in true solution or in colloidal particles smaller than 0.1 micrometer. These results suggest that changes in filtration and preservation methods used for sample collection do not significantly affect reproducibility of strontium-90 analyses in ground water at the INEL.

  6. Estimating true human and animal host source contribution in quantitative microbial source tracking using the Monte Carlo method.

    PubMed

    Wang, Dan; Silkie, Sarah S; Nelson, Kara L; Wuertz, Stefan

    2010-09-01

    Cultivation- and library-independent, quantitative PCR-based methods have become the method of choice in microbial source tracking. However, these qPCR assays are not 100% specific and sensitive for the target sequence in their respective hosts' genome. The factors that can lead to false positive and false negative information in qPCR results are well defined. It is highly desirable to have a way of removing such false information to estimate the true concentration of host-specific genetic markers and help guide the interpretation of environmental monitoring studies. Here we propose a statistical model based on the Law of Total Probability to predict the true concentration of these markers. The distributions of the probabilities of obtaining false information are estimated from representative fecal samples of known origin. Measurement error is derived from the sample precision error of replicated qPCR reactions. Then, the Monte Carlo method is applied to sample from these distributions of probabilities and measurement error. The set of equations given by the Law of Total Probability allows one to calculate the distribution of true concentrations, from which their expected value, confidence interval and other statistical characteristics can be easily evaluated. The output distributions of predicted true concentrations can then be used as input to watershed-wide total maximum daily load determinations, quantitative microbial risk assessment and other environmental models. This model was validated by both statistical simulations and real world samples. It was able to correct the intrinsic false information associated with qPCR assays and output the distribution of true concentrations of Bacteroidales for each animal host group. Model performance was strongly affected by the precision error. It could perform reliably and precisely when the standard deviation of the precision error was small (≤ 0.1). Further improvement on the precision of sample processing and qPCR reaction would greatly improve the performance of the model. This methodology, built upon Bacteroidales assays, is readily transferable to any other microbial source indicator where a universal assay for fecal sources of that indicator exists. Copyright © 2010 Elsevier Ltd. All rights reserved.

  7. Defining standardized protocols for determining the efficacy of a postmilking teat disinfectant following experimental exposure of teats to mastitis pathogens.

    PubMed

    Schukken, Y H; Rauch, B J; Morelli, J

    2013-04-01

    The objective of this paper was to define standardized protocols for determining the efficacy of a postmilking teat disinfectant following experimental exposure of teats to both Staphylococcus aureus and Streptococcus agalactiae. The standardized protocols describe the selection of cows and herds and define the critical points in performing experimental exposure, performing bacterial culture, evaluating the culture results, and finally performing statistical analyses and reporting of the results. The protocols define both negative control and positive control trials. For negative control trials, the protocol states that an efficacy of reducing new intramammary infections (IMI) of at least 40% is required for a teat disinfectant to be considered effective. For positive control trials, noninferiority to a control disinfectant with a published efficacy of reducing new IMI of at least 70% is required. Sample sizes for both negative and positive control trials are calculated. Positive control trials are expected to require a large trial size. Statistical analysis methods are defined and, in the proposed methods, the rate of IMI may be analyzed using generalized linear mixed models. The efficacy of the test product can be evaluated while controlling for important covariates and confounders in the trial. Finally, standards for reporting are defined and reporting considerations are discussed. The use of the defined protocol is shown through presentation of the results of a recent trial of a test product against a negative control. Copyright © 2013 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.

  8. Power of tests for comparing trend curves with application to national immunization survey (NIS).

    PubMed

    Zhao, Zhen

    2011-02-28

    To develop statistical tests for comparing trend curves of study outcomes between two socio-demographic strata across consecutive time points, and compare statistical power of the proposed tests under different trend curves data, three statistical tests were proposed. For large sample size with independent normal assumption among strata and across consecutive time points, the Z and Chi-square test statistics were developed, which are functions of outcome estimates and the standard errors at each of the study time points for the two strata. For small sample size with independent normal assumption, the F-test statistic was generated, which is a function of sample size of the two strata and estimated parameters across study period. If two trend curves are approximately parallel, the power of Z-test is consistently higher than that of both Chi-square and F-test. If two trend curves cross at low interaction, the power of Z-test is higher than or equal to the power of both Chi-square and F-test; however, at high interaction, the powers of Chi-square and F-test are higher than that of Z-test. The measurement of interaction of two trend curves was defined. These tests were applied to the comparison of trend curves of vaccination coverage estimates of standard vaccine series with National Immunization Survey (NIS) 2000-2007 data. Copyright © 2011 John Wiley & Sons, Ltd.

  9. Compounding approach for univariate time series with nonstationary variances

    NASA Astrophysics Data System (ADS)

    Schäfer, Rudi; Barkhofen, Sonja; Guhr, Thomas; Stöckmann, Hans-Jürgen; Kuhl, Ulrich

    2015-12-01

    A defining feature of nonstationary systems is the time dependence of their statistical parameters. Measured time series may exhibit Gaussian statistics on short time horizons, due to the central limit theorem. The sample statistics for long time horizons, however, averages over the time-dependent variances. To model the long-term statistical behavior, we compound the local distribution with the distribution of its parameters. Here, we consider two concrete, but diverse, examples of such nonstationary systems: the turbulent air flow of a fan and a time series of foreign exchange rates. Our main focus is to empirically determine the appropriate parameter distribution for the compounding approach. To this end, we extract the relevant time scales by decomposing the time signals into windows and determine the distribution function of the thus obtained local variances.

  10. Compounding approach for univariate time series with nonstationary variances.

    PubMed

    Schäfer, Rudi; Barkhofen, Sonja; Guhr, Thomas; Stöckmann, Hans-Jürgen; Kuhl, Ulrich

    2015-12-01

    A defining feature of nonstationary systems is the time dependence of their statistical parameters. Measured time series may exhibit Gaussian statistics on short time horizons, due to the central limit theorem. The sample statistics for long time horizons, however, averages over the time-dependent variances. To model the long-term statistical behavior, we compound the local distribution with the distribution of its parameters. Here, we consider two concrete, but diverse, examples of such nonstationary systems: the turbulent air flow of a fan and a time series of foreign exchange rates. Our main focus is to empirically determine the appropriate parameter distribution for the compounding approach. To this end, we extract the relevant time scales by decomposing the time signals into windows and determine the distribution function of the thus obtained local variances.

  11. Groundwater-Quality Data in the South Coast Range-Coastal Study Unit, 2008: Results from the California GAMA Program

    USGS Publications Warehouse

    Mathany, Timothy M.; Burton, Carmen A.; Land, Michael; Belitz, Kenneth

    2010-01-01

    Groundwater quality in the approximately 766-square-mile South Coast Range-Coastal (SCRC) study unit was investigated from May to December 2008, as part of the Priority Basins Project of the Groundwater Ambient Monitoring and Assessment (GAMA) Program. The GAMA Priority Basins Project was developed in response to legislative mandates (Supplemental Report of the 1999 Budget Act 1999-00 Fiscal Year; and, the Groundwater Quality Monitoring Act of 2001 [Sections 10780-10782.3 of the California Water Code, Assembly Bill 599]) to assess and monitor the quality of groundwater in California, and is being conducted by the U.S. Geological Survey (USGS) in cooperation with the California State Water Resources Control Board (SWRCB). The SCRC study unit was the 25th study unit to be sampled as part of the GAMA Priority Basins Project. The SCRC study unit was designed to provide a spatially unbiased assessment of untreated groundwater quality in the primary aquifer systems and to facilitate statistically consistent comparisons of untreated groundwater quality throughout California. The primary aquifer systems (hereinafter referred to as primary aquifers) were defined as that part of the aquifer corresponding to the perforation interval of wells listed in the California Department of Public Health (CDPH) database for the SCRC study unit. The quality of groundwater in shallow or deep water-bearing zones may differ from the quality of groundwater in the primary aquifers; shallow groundwater may be more vulnerable to surficial contamination. In the SCRC study unit, groundwater samples were collected from 70 wells in two study areas (Basins and Uplands) in Santa Barbara and San Luis Obispo Counties. Fifty-five of the wells were selected using a spatially distributed, randomized grid-based method to provide statistical representation of the study unit (grid wells), and 15 wells were selected to aid in evaluation of specific water-quality issues (understanding wells). In addition to the 70 wells sampled, 3 surface-water samples were collected in streams near 2 of the sampled wells in order to better comprehend the interaction between groundwater and surface water in the area. The groundwater samples were analyzed for organic constituents (volatile organic compounds [VOC], pesticides and pesticide degradates, polar pesticides and metabolites, and pharmaceutical compounds), constituents of special interest (perchlorate, N-nitrosodimethylamine [NDMA], and 1,2,3-TCP), naturally occurring inorganic constituents (trace elements, nutrients, dissolved organic carbon [DOC], major and minor ions, silica, total dissolved solids [TDS], and alkalinity), and radioactive constituents (gross alpha and gross beta radioactivity). Naturally occurring isotopes (stable isotopes of hydrogen and oxygen in water, stable isotopes of nitrogen and oxygen in dissolved nitrate, stable isotopes of sulfur in dissolved sulfate, stable isotopes of carbon in dissolved inorganic carbon, activities of tritium, and carbon-14 abundance), and dissolved gases (including noble gases) also were measured to help identify the sources and ages of the sampled groundwater. In total, 298 constituents and field water-quality indicators were investigated. Three types of quality-control samples (blanks, replicates, and matrix-spikes) were collected at approximately 3 to 12 percent of the wells in the SCRC study unit, and the results for these samples were used to evaluate the quality of the data for the groundwater samples. Field blanks rarely contained detectable concentrations of any constituent, suggesting that contamination from sample collection procedures was not a significant source of bias in the data for the groundwater samples. Differences between replicate samples generally were less than 10 percent relative and/or standard deviation, indicating acceptable analytical reproducibility. Matrix-spike recoveries were within the acceptable range (70 to 130 percent) for approximately 84

  12. "Non-hydrolytic" sol-gel synthesis of molybdenum sulfides

    NASA Astrophysics Data System (ADS)

    Leidich, Saskia; Buechele, Dominique; Lauenstein, Raphael; Kluenker, Martin; Lind, Cora

    2016-10-01

    Non-hydrolytic sol-gel reactions provide a low temperature solution based synthetic approach to solid-state materials. In this paper, reactions between molybdenum chloride and hexamethyldisilthiane in chloroform were explored, which gave access to both MoS2 and Mo2S3 after heat treatment of as-recovered amorphous samples to 600-1000 °C. Interesting morphologies were obtained for MoS2, ranging from fused spherical particles to well-defined nanoplatelets and nanoflakes. Both 2H- and 3R-MoS2 were observed, which formed thin hexagonal and triangular platelets, respectively. The platelets exhibited thicknesses of 10-30 nm, which corresponds to 15-50 MoS2 layers. No attempts to prevent agglomeration were made, however, well separated platelets were observed for many samples. Heating at 1000 °C led to formation of Mo2S3 for samples that showed well-defined MoS2 at lower temperatures, while less crystalline samples had a tendency to retain the MoS2 structure.

  13. Reduction of Complications of Local Anaesthesia in Dental Healthcare Setups by Application of the Six Sigma Methodology: A Statistical Quality Improvement Technique.

    PubMed

    Akifuddin, Syed; Khatoon, Farheen

    2015-12-01

    Health care faces challenges due to complications, inefficiencies and other concerns that threaten the safety of patients. The purpose of his study was to identify causes of complications encountered after administration of local anaesthesia for dental and oral surgical procedures and to reduce the incidence of complications by introduction of six sigma methodology. DMAIC (Define, Measure, Analyse, Improve and Control) process of Six Sigma was taken into consideration to reduce the incidence of complications encountered after administration of local anaesthesia injections for dental and oral surgical procedures using failure mode and effect analysis. Pareto analysis was taken into consideration to analyse the most recurring complications. Paired z-sample test using Minitab Statistical Inference and Fisher's exact test was used to statistically analyse the obtained data. The p-value <0.05 was considered as significant value. Total 54 systemic and 62 local complications occurred during three months of analyse and measure phase. Syncope, failure of anaesthesia, trismus, auto mordeduras and pain at injection site was found to be most recurring complications. Cumulative defective percentage was 7.99 in case of pre-improved data and decreased to 4.58 in the control phase. Estimate for difference was 0.0341228 and 95% lower bound for difference was 0.0193966. p-value was found to be highly significant with p= 0.000. The application of six sigma improvement methodology in healthcare tends to deliver consistently better results to the patients as well as hospitals and results in better patient compliance as well as satisfaction.

  14. Descriptive statistics.

    PubMed

    Nick, Todd G

    2007-01-01

    Statistics is defined by the Medical Subject Headings (MeSH) thesaurus as the science and art of collecting, summarizing, and analyzing data that are subject to random variation. The two broad categories of summarizing and analyzing data are referred to as descriptive and inferential statistics. This chapter considers the science and art of summarizing data where descriptive statistics and graphics are used to display data. In this chapter, we discuss the fundamentals of descriptive statistics, including describing qualitative and quantitative variables. For describing quantitative variables, measures of location and spread, for example the standard deviation, are presented along with graphical presentations. We also discuss distributions of statistics, for example the variance, as well as the use of transformations. The concepts in this chapter are useful for uncovering patterns within the data and for effectively presenting the results of a project.

  15. Comparing generalized ensemble methods for sampling of systems with many degrees of freedom

    DOE PAGES

    Lincoff, James; Sasmal, Sukanya; Head-Gordon, Teresa

    2016-11-03

    Here, we compare two standard replica exchange methods using temperature and dielectric constant as the scaling variables for independent replicas against two new corresponding enhanced sampling methods based on non-equilibrium statistical cooling (temperature) or descreening (dielectric). We test the four methods on a rough 1D potential as well as for alanine dipeptide in water, for which their relatively small phase space allows for the ability to define quantitative convergence metrics. We show that both dielectric methods are inferior to the temperature enhanced sampling methods, and in turn show that temperature cool walking (TCW) systematically outperforms the standard temperature replica exchangemore » (TREx) method. We extend our comparisons of the TCW and TREx methods to the 5 residue met-enkephalin peptide, in which we evaluate the Kullback-Leibler divergence metric to show that the rate of convergence between two independent trajectories is faster for TCW compared to TREx. Finally we apply the temperature methods to the 42 residue amyloid-β peptide in which we find non-negligible differences in the disordered ensemble using TCW compared to the standard TREx. All four methods have been made available as software through the OpenMM Omnia software consortium.« less

  16. Comparing generalized ensemble methods for sampling of systems with many degrees of freedom.

    PubMed

    Lincoff, James; Sasmal, Sukanya; Head-Gordon, Teresa

    2016-11-07

    We compare two standard replica exchange methods using temperature and dielectric constant as the scaling variables for independent replicas against two new corresponding enhanced sampling methods based on non-equilibrium statistical cooling (temperature) or descreening (dielectric). We test the four methods on a rough 1D potential as well as for alanine dipeptide in water, for which their relatively small phase space allows for the ability to define quantitative convergence metrics. We show that both dielectric methods are inferior to the temperature enhanced sampling methods, and in turn show that temperature cool walking (TCW) systematically outperforms the standard temperature replica exchange (TREx) method. We extend our comparisons of the TCW and TREx methods to the 5 residue met-enkephalin peptide, in which we evaluate the Kullback-Leibler divergence metric to show that the rate of convergence between two independent trajectories is faster for TCW compared to TREx. Finally we apply the temperature methods to the 42 residue amyloid-β peptide in which we find non-negligible differences in the disordered ensemble using TCW compared to the standard TREx. All four methods have been made available as software through the OpenMM Omnia software consortium (http://www.omnia.md/).

  17. Fine-scale phylogenetic architecture of a complex bacterial community.

    PubMed

    Acinas, Silvia G; Klepac-Ceraj, Vanja; Hunt, Dana E; Pharino, Chanathip; Ceraj, Ivica; Distel, Daniel L; Polz, Martin F

    2004-07-29

    Although molecular data have revealed the vast scope of microbial diversity, two fundamental questions remain unanswered even for well-defined natural microbial communities: how many bacterial types co-exist, and are such types naturally organized into phylogenetically discrete units of potential ecological significance? It has been argued that without such information, the environmental function, population biology and biogeography of microorganisms cannot be rigorously explored. Here we address these questions by comprehensive sampling of two large 16S ribosomal RNA clone libraries from a coastal bacterioplankton community. We show that compensation for artefacts generated by common library construction techniques reveals fine-scale patterns of community composition. At least 516 ribotypes (unique rRNA sequences) were detected in the sample and, by statistical extrapolation, at least 1,633 co-existing ribotypes in the sampled population. More than 50% of the ribotypes fall into discrete clusters containing less than 1% sequence divergence. This pattern cannot be accounted for by interoperon variation, indicating a large predominance of closely related taxa in this community. We propose that such microdiverse clusters arise by selective sweeps and persist because competitive mechanisms are too weak to purge diversity from within them.

  18. Characterization and reconstruction of 3D stochastic microstructures via supervised learning.

    PubMed

    Bostanabad, R; Chen, W; Apley, D W

    2016-12-01

    The need for computational characterization and reconstruction of volumetric maps of stochastic microstructures for understanding the role of material structure in the processing-structure-property chain has been highlighted in the literature. Recently, a promising characterization and reconstruction approach has been developed where the essential idea is to convert the digitized microstructure image into an appropriate training dataset to learn the stochastic nature of the morphology by fitting a supervised learning model to the dataset. This compact model can subsequently be used to efficiently reconstruct as many statistically equivalent microstructure samples as desired. The goal of this paper is to build upon the developed approach in three major directions by: (1) extending the approach to characterize 3D stochastic microstructures and efficiently reconstruct 3D samples, (2) improving the performance of the approach by incorporating user-defined predictors into the supervised learning model, and (3) addressing potential computational issues by introducing a reduced model which can perform as effectively as the full model. We test the extended approach on three examples and show that the spatial dependencies, as evaluated via various measures, are well preserved in the reconstructed samples. © 2016 The Authors Journal of Microscopy © 2016 Royal Microscopical Society.

  19. The AMIGA sample of isolated galaxies. IV. A catalogue of neighbours around isolated galaxies

    NASA Astrophysics Data System (ADS)

    Verley, S.; Odewahn, S. C.; Verdes-Montenegro, L.; Leon, S.; Combes, F.; Sulentic, J.; Bergond, G.; Espada, D.; García, E.; Lisenfeld, U.; Sabater, J.

    2007-08-01

    Context: Studies of the effects of environment on galaxy properties and evolution require well defined control samples. Such isolated galaxy samples have up to now been small or poorly defined. The AMIGA project (Analysis of the interstellar Medium of Isolated GAlaxies) represents an attempt to define a statistically useful sample of the most isolated galaxies in the local (z ≤ 0.05) Universe. Aims: A suitable large sample for the AMIGA project already exists, the Catalogue of Isolated Galaxies (CIG, Karachentseva, 1973, Astrofizicheskie Issledovaniia Izvestiya Spetsial'noj Astrofizicheskoj Observatorii, 8, 3; 1050 galaxies), and we use this sample as a starting point to refine and perform a better quantification of its isolation properties. Methods: Digitised POSS-I E images were analysed out to a minimum projected radius R ≥ 0.5 Mpc around 950 CIG galaxies (those within Vr = 1500 km s-1 were excluded). We identified all galaxy candidates in each field brighter than B = 17.5 with a high degree of confidence using the LMORPHO software. We generated a catalogue of approximately 54 000 potential neighbours (redshifts exist for ≈30% of this sample). Results: Six hundred sixty-six galaxies pass and two hundred eighty-four fail the original CIG isolation criterion. The available redshift data confirm that our catalogue involves a largely background population rather than physically associated neighbours. We find that the exclusion of neighbours within a factor of four in size around each CIG galaxy, employed in the original isolation criterion, corresponds to Δ Vr ≈ 18 000 km s-1 indicating that it was a conservative limit. Conclusions: Galaxies in the CIG have been found to show different degrees of isolation. We conclude that a quantitative measure of this is mandatory. It will be the subject of future work based on the catalogue of neighbours obtained here. Full Table [see full text] is only available in electronic form at the CDS via anonymous ftp to cdsarc.u-strasbg.fr (130.79.128.5) or via http://cdsweb.u-strasbg.fr/cgi-bin/qcat?J/A+A/470/505 and from http://www.iaa.es/AMIGA.html. Figure 4 is only available in electronic form at http://www.aanda.org

  20. Monitoring the healing process of rat bones using Raman spectroscopy

    NASA Astrophysics Data System (ADS)

    Gamulin, O.; Serec, K.; Bilić, V.; Balarin, M.; Kosović, M.; Drmić, D.; Brčić, L.; Seiwerth, S.; Sikirić, P.

    2013-07-01

    The healing effect of BPC 157 on rat femoral head osteonecrosis was monitored by Raman spectroscopy. Three groups of rats were defined: an injured group treated with BPC 157 (10 μg/kg/daily ip), an injured control group (treated with saline, 5 ml/kg/daily ip), and an uninjured healthy group. The spectra were recorded and the healing effect assessed on samples harvested from animals which were sacrificed 3 and 6 weeks after being injured. The statistical analysis of the recorded spectra showed statistical differences between the BPC 157-treated, control, and healthy groups of animals. In particular, after 6 weeks the spectral resemblance between the healthy and BPC 157 samples indicated a positive BPC 157 influence on the healing process of rat femoral head.

  1. Quasi-Supervised Scoring of Human Sleep in Polysomnograms Using Augmented Input Variables

    PubMed Central

    Yaghouby, Farid; Sunderam, Sridhar

    2015-01-01

    The limitations of manual sleep scoring make computerized methods highly desirable. Scoring errors can arise from human rater uncertainty or inter-rater variability. Sleep scoring algorithms either come as supervised classifiers that need scored samples of each state to be trained, or as unsupervised classifiers that use heuristics or structural clues in unscored data to define states. We propose a quasi-supervised classifier that models observations in an unsupervised manner but mimics a human rater wherever training scores are available. EEG, EMG, and EOG features were extracted in 30s epochs from human-scored polysomnograms recorded from 42 healthy human subjects (18 to 79 years) and archived in an anonymized, publicly accessible database. Hypnograms were modified so that: 1. Some states are scored but not others; 2. Samples of all states are scored but not for transitional epochs; and 3. Two raters with 67% agreement are simulated. A framework for quasi-supervised classification was devised in which unsupervised statistical models—specifically Gaussian mixtures and hidden Markov models—are estimated from unlabeled training data, but the training samples are augmented with variables whose values depend on available scores. Classifiers were fitted to signal features incorporating partial scores, and used to predict scores for complete recordings. Performance was assessed using Cohen's K statistic. The quasi-supervised classifier performed significantly better than an unsupervised model and sometimes as well as a completely supervised model despite receiving only partial scores. The quasi-supervised algorithm addresses the need for classifiers that mimic scoring patterns of human raters while compensating for their limitations. PMID:25679475

  2. Quasi-supervised scoring of human sleep in polysomnograms using augmented input variables.

    PubMed

    Yaghouby, Farid; Sunderam, Sridhar

    2015-04-01

    The limitations of manual sleep scoring make computerized methods highly desirable. Scoring errors can arise from human rater uncertainty or inter-rater variability. Sleep scoring algorithms either come as supervised classifiers that need scored samples of each state to be trained, or as unsupervised classifiers that use heuristics or structural clues in unscored data to define states. We propose a quasi-supervised classifier that models observations in an unsupervised manner but mimics a human rater wherever training scores are available. EEG, EMG, and EOG features were extracted in 30s epochs from human-scored polysomnograms recorded from 42 healthy human subjects (18-79 years) and archived in an anonymized, publicly accessible database. Hypnograms were modified so that: 1. Some states are scored but not others; 2. Samples of all states are scored but not for transitional epochs; and 3. Two raters with 67% agreement are simulated. A framework for quasi-supervised classification was devised in which unsupervised statistical models-specifically Gaussian mixtures and hidden Markov models--are estimated from unlabeled training data, but the training samples are augmented with variables whose values depend on available scores. Classifiers were fitted to signal features incorporating partial scores, and used to predict scores for complete recordings. Performance was assessed using Cohen's Κ statistic. The quasi-supervised classifier performed significantly better than an unsupervised model and sometimes as well as a completely supervised model despite receiving only partial scores. The quasi-supervised algorithm addresses the need for classifiers that mimic scoring patterns of human raters while compensating for their limitations. Copyright © 2015 Elsevier Ltd. All rights reserved.

  3. Review of Well Operator Files for Hydraulically Fractured Oil and Gas Production Wells: Well Design and Construction Fact Sheet

    EPA Pesticide Factsheets

    EPA reviewed a statistically representative sample of oil and gas production wells reported by nine service companies to help understand the role of well design and construction practices preventing pathways for subsurface fluid movement.

  4. Seasonal Variation of Total Mercury Burden in the American Alligator (Alligator Mississippiensis) at Merritt Island National Wildlife Refuge (MINWR), Florida

    NASA Technical Reports Server (NTRS)

    Nilsen, Frances M.; Dorsey, Jonathan E.; Long, Stephen E.; Schock, Tracey B.; Bowden, John A.; Lowers, Russell H.; Guillette, Louis J., Jr.

    2016-01-01

    Seasonal variation of mercury (Hg) is not well studied in free-ranging wildlife. Atmospheric deposition patterns of Hg have been studied in detail and have been modeled for both global and specific locations with great accuracy and correlates to environment impact. However, monitoring these trends in wildlife is complicated due to local environmental parameters (e.g., rainfall, humidity, pH, bacterial composition) that can affect the transformation of atmospheric Hg to the biologically available forms. Here, we utilized an abundant and healthy population of American alligators (Alligator mississippiensis) at Merritt Island National Wildlife Refuge (MINWR), FL, and assessed Hg burden in whole blood samples over a span of 7 years (2007 2014; n 174) in an effort to assess seasonal variation of total [Hg]. While the majority of this population is assumed healthy, 18 individuals with low body mass indices (BMI, defined in this study) were captured throughout the 7 year sampling period. These individual alligators exhibited [Hg] that were not consistent with the observed overall seasonal [Hg] variation, and were statistically different from the healthy population of alligators. The alligators with low BMI had elevated concentrations of Hg compared to their age/sex/season matched counterparts with normal BMI. Statistically significant differences were found between the winter and spring seasons for animals with normal BMI. The data in this report supports the conclusion that organismal total [Hg] do fluctuate directly with seasonal deposition rates as well as other seasonal environmental parameters, such as average rainfall and prevailing wind direction. This study highlights the unique environment of MINWR to permit annual assessment of apex predators, such as the American alligator, to determine detailed environmental impact of contaminants of concern.

  5. Flocculation kinetics and aggregate structure of kaolinite mixtures in laminar tube flow.

    PubMed

    Vaezi G, Farid; Sanders, R Sean; Masliyah, Jacob H

    2011-03-01

    Flocculation is commonly used in various solid-liquid separation processes in chemical and mineral industries to separate desired products or to treat waste streams. This paper presents an experimental technique to study flocculation processes in laminar tube flow. This approach allows for more realistic estimation of the shear rate to which an aggregate is exposed, as compared to more complicated shear fields (e.g. stirred tanks). A direct sampling method is used to minimize the effect of sampling on the aggregate structure. A combination of aggregate settling velocity and image analysis was used to quantify the structure of the aggregate. Aggregate size, density, and fractal dimension were found to be the most important aggregate structural parameters. The two methods used to determine aggregate fractal dimension were in good agreement. The effects of advective flow through an aggregate's porous structure and transition-regime drag coefficient on the evaluation of aggregate density were considered. The technique was applied to investigate the flocculation kinetics and the evolution of the aggregate structure of kaolin particles with an anionic flocculant under conditions similar to those of oil sands fine tailings. Aggregates were formed using a well controlled two-stage aggregation process. Detailed statistical analysis was performed to investigate the establishment of dynamic equilibrium condition in terms of aggregate size and density evolution. An equilibrium steady state condition was obtained within 90 s of the start of flocculation; after which no further change in aggregate structure was observed. Although longer flocculation times inside the shear field could conceivably cause aggregate structure conformation, statistical analysis indicated that this did not occur for the studied conditions. The results show that the technique and experimental conditions employed here produce aggregates having a well-defined, reproducible structure. Copyright © 2011. Published by Elsevier Inc.

  6. Evolution of high-mass star-forming regions .

    NASA Astrophysics Data System (ADS)

    Giannetti, A.; Leurini, S.; Wyrowski, F.; Urquhart, J.; König, C.; Csengeri, T.; Güsten, R.; Menten, K. M.

    Observational identification of a coherent evolutionary sequence for high-mass star-forming regions is still missing. We use the progressive heating of the gas caused by the feedback of high-mass young stellar objects to prove the statistical validity of the most common schemes used to observationally define an evolutionary sequence for high-mass clumps, and identify which physical process dominates in the different phases. From the spectroscopic follow-ups carried out towards the TOP100 sample between 84 and 365 km s^-1 giga hertz, we selected several multiplets of CH3CN, CH3CCH, and CH3OH lines to derive the physical properties of the gas in the clumps along the evolutionary sequence. We demonstrate that the evolutionary sequence is statistically valid, and we define intervals in L/M separating the compression, collapse and accretion, and disruption phases. The first hot cores and ZAMS stars appear at L/M≈10usk {L_ȯ}msun-1

  7. [The research protocol III. Study population].

    PubMed

    Arias-Gómez, Jesús; Villasís-Keever, Miguel Ángel; Miranda-Novales, María Guadalupe

    2016-01-01

    The study population is defined as a set of cases, determined, limited, and accessible, that will constitute the subjects for the selection of the sample, and must fulfill several characteristics and distinct criteria. The objectives of this manuscript are focused on specifying each one of the elements required to make the selection of the participants of a research project, during the elaboration of the protocol, including the concepts of study population, sample, selection criteria and sampling methods. After delineating the study population, the researcher must specify the criteria that each participant has to comply. The criteria that include the specific characteristics are denominated selection or eligibility criteria. These criteria are inclusion, exclusion and elimination, and will delineate the eligible population. The sampling methods are divided in two large groups: 1) probabilistic or random sampling and 2) non-probabilistic sampling. The difference lies in the employment of statistical methods to select the subjects. In every research, it is necessary to establish at the beginning the specific number of participants to be included to achieve the objectives of the study. This number is the sample size, and can be calculated or estimated with mathematical formulas and statistic software.

  8. Defining window-boundaries for genomic analyses using smoothing spline techniques

    DOE PAGES

    Beissinger, Timothy M.; Rosa, Guilherme J.M.; Kaeppler, Shawn M.; ...

    2015-04-17

    High-density genomic data is often analyzed by combining information over windows of adjacent markers. Interpretation of data grouped in windows versus at individual locations may increase statistical power, simplify computation, reduce sampling noise, and reduce the total number of tests performed. However, use of adjacent marker information can result in over- or under-smoothing, undesirable window boundary specifications, or highly correlated test statistics. We introduce a method for defining windows based on statistically guided breakpoints in the data, as a foundation for the analysis of multiple adjacent data points. This method involves first fitting a cubic smoothing spline to the datamore » and then identifying the inflection points of the fitted spline, which serve as the boundaries of adjacent windows. This technique does not require prior knowledge of linkage disequilibrium, and therefore can be applied to data collected from individual or pooled sequencing experiments. Moreover, in contrast to existing methods, an arbitrary choice of window size is not necessary, since these are determined empirically and allowed to vary along the genome.« less

  9. Unintentional Injuries, Violence, and the Health of Young People

    ERIC Educational Resources Information Center

    Centers for Disease Control and Prevention, 2006

    2006-01-01

    This fact sheet defines unintentional injuries and violence as the terms are used by the CDC and provides statistics on the leading causes of injury mortality and morbidity among children and adolescents, as well as information on the context of injury occurrence. (Contains 2 tables.)

  10. Large ensemble modeling of last deglacial retreat of the West Antarctic Ice Sheet: comparison of simple and advanced statistical techniques

    NASA Astrophysics Data System (ADS)

    Pollard, D.; Chang, W.; Haran, M.; Applegate, P.; DeConto, R.

    2015-11-01

    A 3-D hybrid ice-sheet model is applied to the last deglacial retreat of the West Antarctic Ice Sheet over the last ~ 20 000 years. A large ensemble of 625 model runs is used to calibrate the model to modern and geologic data, including reconstructed grounding lines, relative sea-level records, elevation-age data and uplift rates, with an aggregate score computed for each run that measures overall model-data misfit. Two types of statistical methods are used to analyze the large-ensemble results: simple averaging weighted by the aggregate score, and more advanced Bayesian techniques involving Gaussian process-based emulation and calibration, and Markov chain Monte Carlo. Results for best-fit parameter ranges and envelopes of equivalent sea-level rise with the simple averaging method agree quite well with the more advanced techniques, but only for a large ensemble with full factorial parameter sampling. Best-fit parameter ranges confirm earlier values expected from prior model tuning, including large basal sliding coefficients on modern ocean beds. Each run is extended 5000 years into the "future" with idealized ramped climate warming. In the majority of runs with reasonable scores, this produces grounding-line retreat deep into the West Antarctic interior, and the analysis provides sea-level-rise envelopes with well defined parametric uncertainty bounds.

  11. Data-adaptive test statistics for microarray data.

    PubMed

    Mukherjee, Sach; Roberts, Stephen J; van der Laan, Mark J

    2005-09-01

    An important task in microarray data analysis is the selection of genes that are differentially expressed between different tissue samples, such as healthy and diseased. However, microarray data contain an enormous number of dimensions (genes) and very few samples (arrays), a mismatch which poses fundamental statistical problems for the selection process that have defied easy resolution. In this paper, we present a novel approach to the selection of differentially expressed genes in which test statistics are learned from data using a simple notion of reproducibility in selection results as the learning criterion. Reproducibility, as we define it, can be computed without any knowledge of the 'ground-truth', but takes advantage of certain properties of microarray data to provide an asymptotically valid guide to expected loss under the true data-generating distribution. We are therefore able to indirectly minimize expected loss, and obtain results substantially more robust than conventional methods. We apply our method to simulated and oligonucleotide array data. By request to the corresponding author.

  12. Defining the ecological hydrology of Taiwan Rivers using multivariate statistical methods

    NASA Astrophysics Data System (ADS)

    Chang, Fi-John; Wu, Tzu-Ching; Tsai, Wen-Ping; Herricks, Edwin E.

    2009-09-01

    SummaryThe identification and verification of ecohydrologic flow indicators has found new support as the importance of ecological flow regimes is recognized in modern water resources management, particularly in river restoration and reservoir management. An ecohydrologic indicator system reflecting the unique characteristics of Taiwan's water resources and hydrology has been developed, the Taiwan ecohydrological indicator system (TEIS). A major challenge for the water resources community is using the TEIS to provide environmental flow rules that improve existing water resources management. This paper examines data from the extensive network of flow monitoring stations in Taiwan using TEIS statistics to define and refine environmental flow options in Taiwan. Multivariate statistical methods were used to examine TEIS statistics for 102 stations representing the geographic and land use diversity of Taiwan. The Pearson correlation coefficient showed high multicollinearity between the TEIS statistics. Watersheds were separated into upper and lower-watershed locations. An analysis of variance indicated significant differences between upstream, more natural, and downstream, more developed, locations in the same basin with hydrologic indicator redundancy in flow change and magnitude statistics. Issues of multicollinearity were examined using a Principal Component Analysis (PCA) with the first three components related to general flow and high/low flow statistics, frequency and time statistics, and quantity statistics. These principle components would explain about 85% of the total variation. A major conclusion is that managers must be aware of differences among basins, as well as differences within basins that will require careful selection of management procedures to achieve needed flow regimes.

  13. Evaluation of standardized and applied variables in predicting treatment outcomes of polytrauma patients.

    PubMed

    Aksamija, Goran; Mulabdic, Adi; Rasic, Ismar; Muhovic, Samir; Gavric, Igor

    2011-01-01

    Polytrauma is defined as an injury where they are affected by at least two different organ systems or body, with at least one life-threatening injuries. Given the multilevel model care of polytrauma patients within KCUS are inevitable weaknesses in the management of this category of patients. To determine the dynamics of existing procedures in treatment of polytrauma patients on admission to KCUS, and based on statistical analysis of variables applied to determine and define the factors that influence the final outcome of treatment, and determine their mutual relationship, which may result in eliminating the flaws in the approach to the problem. The study was based on 263 polytrauma patients. Parametric and non-parametric statistical methods were used. Basic statistics were calculated, based on the calculated parameters for the final achievement of research objectives, multicoleration analysis, image analysis, discriminant analysis and multifactorial analysis were used. From the universe of variables for this study we selected sample of n = 25 variables, of which the first two modular, others belong to the common measurement space (n = 23) and in this paper defined as a system variable methods, procedures and assessments of polytrauma patients. After the multicoleration analysis, since the image analysis gave a reliable measurement results, we started the analysis of eigenvalues, that is defining the factors upon which they obtain information about the system solve the problem of the existing model and its correlation with treatment outcome. The study singled out the essential factors that determine the current organizational model of care, which may affect the treatment and better outcome of polytrauma patients. This analysis has shown the maximum correlative relationships between these practices and contributed to development guidelines that are defined by isolated factors.

  14. General Constraints on Sampling Wildlife on FIA Plots

    Treesearch

    Larissa L. Bailey; John R. Sauer; James D. Nichols; Paul H. Geissler

    2005-01-01

    This paper reviews the constraints to sampling wildlife populations at FIA points. Wildlife sampling programs must have well-defined goals and provide information adequate to meet those goals. Investigators should choose a State variable based on information needs and the spatial sampling scale. We discuss estimation-based methods for three State variables: species...

  15. From fields to objects: A review of geographic boundary analysis

    NASA Astrophysics Data System (ADS)

    Jacquez, G. M.; Maruca, S.; Fortin, M.-J.

    Geographic boundary analysis is a relatively new approach unfamiliar to many spatial analysts. It is best viewed as a technique for defining objects - geographic boundaries - on spatial fields, and for evaluating the statistical significance of characteristics of those boundary objects. This is accomplished using null spatial models representative of the spatial processes expected in the absence of boundary-generating phenomena. Close ties to the object-field dialectic eminently suit boundary analysis to GIS data. The majority of existing spatial methods are field-based in that they describe, estimate, or predict how attributes (variables defining the field) vary through geographic space. Such methods are appropriate for field representations but not object representations. As the object-field paradigm gains currency in geographic information science, appropriate techniques for the statistical analysis of objects are required. The methods reviewed in this paper are a promising foundation. Geographic boundary analysis is clearly a valuable addition to the spatial statistical toolbox. This paper presents the philosophy of, and motivations for geographic boundary analysis. It defines commonly used statistics for quantifying boundaries and their characteristics, as well as simulation procedures for evaluating their significance. We review applications of these techniques, with the objective of making this promising approach accessible to the GIS-spatial analysis community. We also describe the implementation of these methods within geographic boundary analysis software: GEM.

  16. Fluid Inclusion Gas Analysis

    DOE Data Explorer

    Dilley, Lorie

    2013-01-01

    Fluid inclusion gas analysis for wells in various geothermal areas. Analyses used in developing fluid inclusion stratigraphy for wells and defining fluids across the geothermal fields. Each sample has mass spectrum counts for 180 chemical species.

  17. Meta-Analysis of Inquiry-Based Instruction Research

    NASA Astrophysics Data System (ADS)

    Hasanah, N.; Prasetyo, A. P. B.; Rudyatmi, E.

    2017-04-01

    Inquiry-based instruction in biology has been the focus of educational research conducted by Unnes biology department students in collaboration with their university supervisors. This study aimed to describe the methodological aspects, inquiry teaching methods critically, and to analyse the results claims, of the selected four student research reports, grounded in inquiry, based on the database of Unnes biology department 2014. Four experimental quantitative research of 16 were selected as research objects by purposive sampling technique. Data collected through documentation study was qualitatively analysed regarding methods used, quality of inquiry syntax, and finding claims. Findings showed that the student research was still the lack of relevant aspects of research methodology, namely in appropriate sampling procedures, limited validity tests of all research instruments, and the limited parametric statistic (t-test) not supported previously by data normality tests. Their consistent inquiry syntax supported the four mini-thesis claims that inquiry-based teaching influenced their dependent variables significantly. In other words, the findings indicated that positive claims of the research results were not fully supported by good research methods, and well-defined inquiry procedures implementation.

  18. Further developments in cloud statistics for computer simulations

    NASA Technical Reports Server (NTRS)

    Chang, D. T.; Willand, J. H.

    1972-01-01

    This study is a part of NASA's continued program to provide global statistics of cloud parameters for computer simulation. The primary emphasis was on the development of the data bank of the global statistical distributions of cloud types and cloud layers and their applications in the simulation of the vertical distributions of in-cloud parameters such as liquid water content. These statistics were compiled from actual surface observations as recorded in Standard WBAN forms. Data for a total of 19 stations were obtained and reduced. These stations were selected to be representative of the 19 primary cloud climatological regions defined in previous studies of cloud statistics. Using the data compiled in this study, a limited study was conducted of the hemogeneity of cloud regions, the latitudinal dependence of cloud-type distributions, the dependence of these statistics on sample size, and other factors in the statistics which are of significance to the problem of simulation. The application of the statistics in cloud simulation was investigated. In particular, the inclusion of the new statistics in an expanded multi-step Monte Carlo simulation scheme is suggested and briefly outlined.

  19. Microwave resonances in dielectric samples probed in Corbino geometry: simulation and experiment.

    PubMed

    Felger, M Maximilian; Dressel, Martin; Scheffler, Marc

    2013-11-01

    The Corbino approach, where the sample of interest terminates a coaxial cable, is a well-established method for microwave spectroscopy. If the sample is dielectric and if the probe geometry basically forms a conductive cavity, this combination can sustain well-defined microwave resonances that are detrimental for broadband measurements. Here, we present detailed simulations and measurements to investigate the resonance frequencies as a function of sample and probe size and of sample permittivity. This allows a quantitative optimization to increase the frequency of the lowest-lying resonance.

  20. Publication bias in situ.

    PubMed

    Phillips, Carl V

    2004-08-05

    Publication bias, as typically defined, refers to the decreased likelihood of studies' results being published when they are near the null, not statistically significant, or otherwise "less interesting." But choices about how to analyze the data and which results to report create a publication bias within the published results, a bias I label "publication bias in situ" (PBIS). PBIS may create much greater bias in the literature than traditionally defined publication bias (the failure to publish any result from a study). The causes of PBIS are well known, consisting of various decisions about reporting that are influenced by the data. But its impact is not generally appreciated, and very little attention is devoted to it. What attention there is consists largely of rules for statistical analysis that are impractical and do not actually reduce the bias in reported estimates. PBIS cannot be reduced by statistical tools because it is not fundamentally a problem of statistics, but rather of non-statistical choices and plain language interpretations. PBIS should be recognized as a phenomenon worthy of study - it is extremely common and probably has a huge impact on results reported in the literature - and there should be greater systematic efforts to identify and reduce it. The paper presents examples, including results of a recent HIV vaccine trial, that show how easily PBIS can have a large impact on reported results, as well as how there can be no simple answer to it. PBIS is a major problem, worthy of substantially more attention than it receives. There are ways to reduce the bias, but they are very seldom employed because they are largely unrecognized.

  1. Impact of Satellite Viewing-Swath Width on Global and Regional Aerosol Optical Thickness Statistics and Trends

    NASA Technical Reports Server (NTRS)

    Colarco, P. R.; Kahn, R. A.; Remer, L. A.; Levy, R. C.

    2014-01-01

    We use the Moderate Resolution Imaging Spectroradiometer (MODIS) satellite aerosol optical thickness (AOT) product to assess the impact of reduced swath width on global and regional AOT statistics and trends. Alongtrack and across-track sampling strategies are employed, in which the full MODIS data set is sub-sampled with various narrow-swath (approximately 400-800 km) and single pixel width (approximately 10 km) configurations. Although view-angle artifacts in the MODIS AOT retrieval confound direct comparisons between averages derived from different sub-samples, careful analysis shows that with many portions of the Earth essentially unobserved, spatial sampling introduces uncertainty in the derived seasonal-regional mean AOT. These AOT spatial sampling artifacts comprise up to 60%of the full-swath AOT value under moderate aerosol loading, and can be as large as 0.1 in some regions under high aerosol loading. Compared to full-swath observations, narrower swath and single pixel width sampling exhibits a reduced ability to detect AOT trends with statistical significance. On the other hand, estimates of the global, annual mean AOT do not vary significantly from the full-swath values as spatial sampling is reduced. Aggregation of the MODIS data at coarse grid scales (10 deg) shows consistency in the aerosol trends across sampling strategies, with increased statistical confidence, but quantitative errors in the derived trends are found even for the full-swath data when compared to high spatial resolution (0.5 deg) aggregations. Using results of a model-derived aerosol reanalysis, we find consistency in our conclusions about a seasonal-regional spatial sampling artifact in AOT Furthermore, the model shows that reduced spatial sampling can amount to uncertainty in computed shortwave top-ofatmosphere aerosol radiative forcing of 2-3 W m(sup-2). These artifacts are lower bounds, as possibly other unconsidered sampling strategies would perform less well. These results suggest that future aerosol satellite missions having significantly less than full-swath viewing are unlikely to sample the true AOT distribution well enough to obtain the statistics needed to reduce uncertainty in aerosol direct forcing of climate.

  2. Content Analysis of Chemistry Curricula in Germany Case Study: Chemical Reactions

    ERIC Educational Resources Information Center

    Timofte, Roxana S.

    2015-01-01

    Curriculum-assessment alignment is a well known foundation for good practice in educational assessment, for items' curricular validity purposes. Nowadays instruments are designed to measure pupils' competencies in one or more areas of competence. Sub-competence areas could be defined theoretically and statistical analysis of empirical data by…

  3. Alaska national hydrography dataset positional accuracy assessment study

    USGS Publications Warehouse

    Arundel, Samantha; Yamamoto, Kristina H.; Constance, Eric; Mantey, Kim; Vinyard-Houx, Jeremy

    2013-01-01

    Initial visual assessments Wide range in the quality of fit between features in NHD and these new image sources. No statistical analysis has been performed to actually quantify accuracy Determining absolute accuracy is cost prohibitive (must collect independent, well defined test points) Quantitative analysis of relative positional error is feasible.

  4. Difficulties in learning and teaching statistics: teacher views

    NASA Astrophysics Data System (ADS)

    Koparan, Timur

    2015-01-01

    The purpose of this study is to define teacher views about the difficulties in learning and teaching middle school statistics subjects. To serve this aim, a number of interviews were conducted with 10 middle school maths teachers in 2011-2012 school year in the province of Trabzon. Of the qualitative descriptive research methods, the semi-structured interview technique was applied in the research. In accordance with the aim, teacher opinions about the statistics subjects were examined and analysed. Similar responses from the teachers were grouped and evaluated. The teachers stated that it was positive that middle school statistics subjects were taught gradually in every grade but some difficulties were experienced in the teaching of this subject. The findings are presented in eight themes which are context, sample, data representation, central tendency and dispersion measure, probability, variance, and other difficulties.

  5. Rasch model based analysis of the Force Concept Inventory

    NASA Astrophysics Data System (ADS)

    Planinic, Maja; Ivanjek, Lana; Susac, Ana

    2010-06-01

    The Force Concept Inventory (FCI) is an important diagnostic instrument which is widely used in the field of physics education research. It is therefore very important to evaluate and monitor its functioning using different tools for statistical analysis. One of such tools is the stochastic Rasch model, which enables construction of linear measures for persons and items from raw test scores and which can provide important insight in the structure and functioning of the test (how item difficulties are distributed within the test, how well the items fit the model, and how well the items work together to define the underlying construct). The data for the Rasch analysis come from the large-scale research conducted in 2006-07, which investigated Croatian high school students’ conceptual understanding of mechanics on a representative sample of 1676 students (age 17-18 years). The instrument used in research was the FCI. The average FCI score for the whole sample was found to be (27.7±0.4)% , indicating that most of the students were still non-Newtonians at the end of high school, despite the fact that physics is a compulsory subject in Croatian schools. The large set of obtained data was analyzed with the Rasch measurement computer software WINSTEPS 3.66. Since the FCI is routinely used as pretest and post-test on two very different types of population (non-Newtonian and predominantly Newtonian), an additional predominantly Newtonian sample ( N=141 , average FCI score of 64.5%) of first year students enrolled in introductory physics course at University of Zagreb was also analyzed. The Rasch model based analysis suggests that the FCI has succeeded in defining a sufficiently unidimensional construct for each population. The analysis of fit of data to the model found no grossly misfitting items which would degrade measurement. Some items with larger misfit and items with significantly different difficulties in the two samples of students do require further examination. The analysis revealed some problems with item distribution in the FCI and suggested that the FCI may function differently in non-Newtonian and predominantly Newtonian population. Some possible improvements of the test are suggested.

  6. Approximate sample sizes required to estimate length distributions

    USGS Publications Warehouse

    Miranda, L.E.

    2007-01-01

    The sample sizes required to estimate fish length were determined by bootstrapping from reference length distributions. Depending on population characteristics and species-specific maximum lengths, 1-cm length-frequency histograms required 375-1,200 fish to estimate within 10% with 80% confidence, 2.5-cm histograms required 150-425 fish, proportional stock density required 75-140 fish, and mean length required 75-160 fish. In general, smaller species, smaller populations, populations with higher mortality, and simpler length statistics required fewer samples. Indices that require low sample sizes may be suitable for monitoring population status, and when large changes in length are evident, additional sampling effort may be allocated to more precisely define length status with more informative estimators. ?? Copyright by the American Fisheries Society 2007.

  7. How Sample Size Affects a Sampling Distribution

    ERIC Educational Resources Information Center

    Mulekar, Madhuri S.; Siegel, Murray H.

    2009-01-01

    If students are to understand inferential statistics successfully, they must have a profound understanding of the nature of the sampling distribution. Specifically, they must comprehend the determination of the expected value and standard error of a sampling distribution as well as the meaning of the central limit theorem. Many students in a high…

  8. Uranium resource assessment through statistical analysis of exploration geochemical and other data. Final report. [Codes EVAL, SURE

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Koch, G.S. Jr.; Howarth, R.J.; Schuenemeyer, J.H.

    1981-02-01

    We have developed a procedure that can help quadrangle evaluators to systematically summarize and use hydrogeochemical and stream sediment reconnaissance (HSSR) and occurrence data. Although we have not provided an independent estimate of uranium endowment, we have devised a methodology that will provide this independent estimate when additional calibration is done by enlarging the study area. Our statistical model for evaluation (system EVAL) ranks uranium endowment for each quadrangle. Because using this model requires experience in geology, statistics, and data analysis, we have also devised a simplified model, presented in the package SURE, a System for Uranium Resource Evaluation. Wemore » have developed and tested these models for the four quadrangles in southern Colorado that comprise the study area; to investigate their generality, the models should be applied to other quandrangles. Once they are calibrated with accepted uranium endowments for several well-known quadrangles, the models can be used to give independent estimates for less-known quadrangles. The point-oriented models structure the objective comparison of the quandrangles on the bases of: (1) Anomalies (a) derived from stream sediments, (b) derived from waters (stream, well, pond, etc.), (2) Geology (a) source rocks, as defined by the evaluator, (b) host rocks, as defined by the evaluator, and (3) Aerial radiometric anomalies.« less

  9. The relationship between cross-sectional shapes and FTIR profiles in synthetic wig fibers and their discriminating abilities - An evidential value perspective.

    PubMed

    Joslin Yogi, Theresa A; Penrod, Michael; Holt, Melinda; Buzzini, Patrick

    2018-02-01

    Wig fragments or fibers may occasionally be recognized as potential physical evidence during criminal investigations. While analytical methods traditionally adopted for the examination of textile fibers are utilized for the characterizations and comparisons of wig specimens, it is essential to understand in deeper detail the valuable contribution of features of these non-routine evidentiary materials as well as the relationship of the gathered analytical data. This study explores the dependence between the microscopic features of cross-sectional shapes and the polymer type gathered by Fourier transform infrared (FTIR) spectroscopy. The discriminating power of the two methods of cross-sectioning and FTIR spectroscopy was also investigated. Forty-one synthetic wigs varying in both quality and price were collected: twenty-three brown, twelve blondes and six black samples. The collected samples were observed using light microscopy methods (bright field illumination and polarized light), before obtaining cross-sections using the Joliff method and analyze them using FTIR spectroscopy. The forty-one samples were divided into ten groups based on one or more of the ten types of cross-sectional shapes that were observed. The majority of encountered cross-sectional shapes were defined as horseshoe, dog bone and lobular. Infrared spectroscopy confirmed modacrylic to be the most prevalent fiber type. Blends of modacrylic and polyvinyl chloride fibers were also observed as well as polypropylene wig samples. The Goodman and Kruskal lambda statistical test was used and showed that the cross-sectional shape and infrared profile were related. From an evidentiary value perspective, this finding has implications when addressing questions about a common source between questioned wig specimens and a wig reference sample. Copyright © 2017 Elsevier B.V. All rights reserved.

  10. How conservative is Fisher's exact test? A quantitative evaluation of the two-sample comparative binomial trial.

    PubMed

    Crans, Gerald G; Shuster, Jonathan J

    2008-08-15

    The debate as to which statistical methodology is most appropriate for the analysis of the two-sample comparative binomial trial has persisted for decades. Practitioners who favor the conditional methods of Fisher, Fisher's exact test (FET), claim that only experimental outcomes containing the same amount of information should be considered when performing analyses. Hence, the total number of successes should be fixed at its observed level in hypothetical repetitions of the experiment. Using conditional methods in clinical settings can pose interpretation difficulties, since results are derived using conditional sample spaces rather than the set of all possible outcomes. Perhaps more importantly from a clinical trial design perspective, this test can be too conservative, resulting in greater resource requirements and more subjects exposed to an experimental treatment. The actual significance level attained by FET (the size of the test) has not been reported in the statistical literature. Berger (J. R. Statist. Soc. D (The Statistician) 2001; 50:79-85) proposed assessing the conservativeness of conditional methods using p-value confidence intervals. In this paper we develop a numerical algorithm that calculates the size of FET for sample sizes, n, up to 125 per group at the two-sided significance level, alpha = 0.05. Additionally, this numerical method is used to define new significance levels alpha(*) = alpha+epsilon, where epsilon is a small positive number, for each n, such that the size of the test is as close as possible to the pre-specified alpha (0.05 for the current work) without exceeding it. Lastly, a sample size and power calculation example are presented, which demonstrates the statistical advantages of implementing the adjustment to FET (using alpha(*) instead of alpha) in the two-sample comparative binomial trial. 2008 John Wiley & Sons, Ltd

  11. Estimation of aquifer scale proportion using equal area grids: assessment of regional scale groundwater quality

    USGS Publications Warehouse

    Belitz, Kenneth; Jurgens, Bryant C.; Landon, Matthew K.; Fram, Miranda S.; Johnson, Tyler D.

    2010-01-01

    The proportion of an aquifer with constituent concentrations above a specified threshold (high concentrations) is taken as a nondimensional measure of regional scale water quality. If computed on the basis of area, it can be referred to as the aquifer scale proportion. A spatially unbiased estimate of aquifer scale proportion and a confidence interval for that estimate are obtained through the use of equal area grids and the binomial distribution. Traditionally, the confidence interval for a binomial proportion is computed using either the standard interval or the exact interval. Research from the statistics literature has shown that the standard interval should not be used and that the exact interval is overly conservative. On the basis of coverage probability and interval width, the Jeffreys interval is preferred. If more than one sample per cell is available, cell declustering is used to estimate the aquifer scale proportion, and Kish's design effect may be useful for estimating an effective number of samples. The binomial distribution is also used to quantify the adequacy of a grid with a given number of cells for identifying a small target, defined as a constituent that is present at high concentrations in a small proportion of the aquifer. Case studies illustrate a consistency between approaches that use one well per grid cell and many wells per cell. The methods presented in this paper provide a quantitative basis for designing a sampling program and for utilizing existing data.

  12. Gastroschisis: antenatal sonographic predictors of adverse neonatal outcome.

    PubMed

    Page, Rachael; Ferraro, Zachary Michael; Moretti, Felipe; Fung, Karen Fung Kee

    2014-01-01

    The aim of this review was to identify clinically significant ultrasound predictors of adverse neonatal outcome in fetal gastroschisis. A quasi-systematic review was conducted in PubMed and Ovid using the key terms "gastroschisis," "predictors," "outcome," and "ultrasound." A total of 18 papers were included. The most common sonographic predictors were intra-abdominal bowel dilatation (IABD), intrauterine growth restriction (IUGR), and bowel dilatation not otherwise specified (NOS). Three ultrasound markers were consistently found to be statistically insignificant with respect to predicting adverse outcome including abdominal circumference, stomach herniation and dilatation, and extra-abdominal bowel dilatation (EABD). Gastroschisis is associated with several comorbidities, yet there is much discrepancy in the literature regarding which specific ultrasound markers best predict adverse neonatal outcomes. Future research should include prospective trials with larger sample sizes and use well-defined and consistent definitions of the adverse outcomes investigated with consideration given to IABD.

  13. Sampling and counting genome rearrangement scenarios

    PubMed Central

    2015-01-01

    Background Even for moderate size inputs, there are a tremendous number of optimal rearrangement scenarios, regardless what the model is and which specific question is to be answered. Therefore giving one optimal solution might be misleading and cannot be used for statistical inferring. Statistically well funded methods are necessary to sample uniformly from the solution space and then a small number of samples are sufficient for statistical inferring. Contribution In this paper, we give a mini-review about the state-of-the-art of sampling and counting rearrangement scenarios, focusing on the reversal, DCJ and SCJ models. Above that, we also give a Gibbs sampler for sampling most parsimonious labeling of evolutionary trees under the SCJ model. The method has been implemented and tested on real life data. The software package together with example data can be downloaded from http://www.renyi.hu/~miklosi/SCJ-Gibbs/ PMID:26452124

  14. Sigsearch: a new term for post hoc unplanned search for statistically significant relationships with the intent to create publishable findings.

    PubMed

    Hashim, Muhammad Jawad

    2010-09-01

    Post-hoc secondary data analysis with no prespecified hypotheses has been discouraged by textbook authors and journal editors alike. Unfortunately no single term describes this phenomenon succinctly. I would like to coin the term "sigsearch" to define this practice and bring it within the teaching lexicon of statistics courses. Sigsearch would include any unplanned, post-hoc search for statistical significance using multiple comparisons of subgroups. It would also include data analysis with outcomes other than the prespecified primary outcome measure of a study as well as secondary data analyses of earlier research.

  15. Constructing a Reward-Related Quality of Life Statistic in Daily Life—a Proof of Concept Study Using Positive Affect

    PubMed Central

    Verhagen, Simone J. W.; Simons, Claudia J. P.; van Zelst, Catherine; Delespaul, Philippe A. E. G.

    2017-01-01

    Background: Mental healthcare needs person-tailored interventions. Experience Sampling Method (ESM) can provide daily life monitoring of personal experiences. This study aims to operationalize and test a measure of momentary reward-related Quality of Life (rQoL). Intuitively, quality of life improves by spending more time on rewarding experiences. ESM clinical interventions can use this information to coach patients to find a realistic, optimal balance of positive experiences (maximize reward) in daily life. rQoL combines the frequency of engaging in a relevant context (a ‘behavior setting’) with concurrent (positive) affect. High rQoL occurs when the most frequent behavior settings are combined with positive affect or infrequent behavior settings co-occur with low positive affect. Methods: Resampling procedures (Monte Carlo experiments) were applied to assess the reliability of rQoL using various behavior setting definitions under different sampling circumstances, for real or virtual subjects with low-, average- and high contextual variability. Furthermore, resampling was used to assess whether rQoL is a distinct concept from positive affect. Virtual ESM beep datasets were extracted from 1,058 valid ESM observations for virtual and real subjects. Results: Behavior settings defined by Who-What contextual information were most informative. Simulations of at least 100 ESM observations are needed for reliable assessment. Virtual ESM beep datasets of a real subject can be defined by Who-What-Where behavior setting combinations. Large sample sizes are necessary for reliable rQoL assessments, except for subjects with low contextual variability. rQoL is distinct from positive affect. Conclusion: rQoL is a feasible concept. Monte Carlo experiments should be used to assess the reliable implementation of an ESM statistic. Future research in ESM should asses the behavior of summary statistics under different sampling situations. This exploration is especially relevant in clinical implementation, where often only small datasets are available. PMID:29163294

  16. Constructing a Reward-Related Quality of Life Statistic in Daily Life-a Proof of Concept Study Using Positive Affect.

    PubMed

    Verhagen, Simone J W; Simons, Claudia J P; van Zelst, Catherine; Delespaul, Philippe A E G

    2017-01-01

    Background: Mental healthcare needs person-tailored interventions. Experience Sampling Method (ESM) can provide daily life monitoring of personal experiences. This study aims to operationalize and test a measure of momentary reward-related Quality of Life (rQoL). Intuitively, quality of life improves by spending more time on rewarding experiences. ESM clinical interventions can use this information to coach patients to find a realistic, optimal balance of positive experiences (maximize reward) in daily life. rQoL combines the frequency of engaging in a relevant context (a 'behavior setting') with concurrent (positive) affect. High rQoL occurs when the most frequent behavior settings are combined with positive affect or infrequent behavior settings co-occur with low positive affect. Methods: Resampling procedures (Monte Carlo experiments) were applied to assess the reliability of rQoL using various behavior setting definitions under different sampling circumstances, for real or virtual subjects with low-, average- and high contextual variability. Furthermore, resampling was used to assess whether rQoL is a distinct concept from positive affect. Virtual ESM beep datasets were extracted from 1,058 valid ESM observations for virtual and real subjects. Results: Behavior settings defined by Who-What contextual information were most informative. Simulations of at least 100 ESM observations are needed for reliable assessment. Virtual ESM beep datasets of a real subject can be defined by Who-What-Where behavior setting combinations. Large sample sizes are necessary for reliable rQoL assessments, except for subjects with low contextual variability. rQoL is distinct from positive affect. Conclusion: rQoL is a feasible concept. Monte Carlo experiments should be used to assess the reliable implementation of an ESM statistic. Future research in ESM should asses the behavior of summary statistics under different sampling situations. This exploration is especially relevant in clinical implementation, where often only small datasets are available.

  17. Towards well-defined gold nanomaterials via diafiltration and aptamer mediated synthesis

    NASA Astrophysics Data System (ADS)

    Sweeney, Scott Francis

    Gold nanoparticles have garnered recent attention due to their intriguing size- and shape-dependent properties. Routine access to well-defined gold nanoparticle samples in terms of core diameter, shape, peripheral functionality and purity is required in order to carry out fundamental studies of their properties and to utilize these properties in future applications. For this reason, the development of methods for preparing well-defined gold nanoparticle samples remains an area of active research in materials science. In this dissertation, two methods, diafiltration and aptamer mediated synthesis, are explored as possible routes towards well-defined gold nanoparticle samples. It is shown that diafiltration has considerable potential for the efficient and convenient purification and size separation of water-soluble nanoparticles. The suitability of diafiltration for (i) the purification of water-soluble gold nanoparticles, (ii) the separation of a bimodal distribution of nanoparticles into fractions, (iii) the fractionation of a polydisperse sample and (iv) the isolation of [rimers from monomers and aggregates is studied. NMR, thermogravimetric analysis (TGA), and X-ray photoelectron spectroscopy (XPS) measurements demonstrate that diafiltration produces highly pure nanoparticles. UV-visible spectroscopic and transmission electron microscopic analyses show that diafiltration offers the ability to separate nanoparticles of disparate core size, including linked nanoparticles. These results demonstrate the applicability of diafiltration for the rapid and green preparation of high-purity gold nanoparticle samples and the size separation of heterogeneous nanoparticle samples. In the second half of the dissertation, the identification of materials specific aptamers and their use to synthesize shaped gold nanoparticles is explored. The use of in vitro selection for identifying materials specific peptide and oligonucleotide aptamers is reviewed, outlining the specific requirements of in vitro selection for materials and the ways in which the field can be advanced. A promising new technique, in vitro selection on surfaces (ISOS), is developed and the discovery using ISOS of RNA aptamers that bind to evaporated gold is discussed. Analysis of the isolated gold binding RNA aptamers indicates that they are highly structured with single-stranded polyadenosine binding motifs. These aptamers, and similarly isolated peptide aptamers, are briefly explored for their ability to synthesize gold nanoparticles. This dissertation contains both previously published and unpublished co-authored material.

  18. Student and Teacher Factors as Predictors of Statistics Achievement in Federal School of Statistics Ibadan

    ERIC Educational Resources Information Center

    Adetona, Abel Adekanmi

    2017-01-01

    The study aimed at assessing how students and teachers factor taken together influence students' achievement in Statistics as well as their relative contribution to the prediction. Two research questions were raised and purposive sampling was adopted to select national diploma year 2 students since they are already in their final level in the…

  19. Validation of Statistical Sampling Algorithms in Visual Sample Plan (VSP): Summary Report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nuffer, Lisa L; Sego, Landon H.; Wilson, John E.

    2009-02-18

    The U.S. Department of Homeland Security, Office of Technology Development (OTD) contracted with a set of U.S. Department of Energy national laboratories, including the Pacific Northwest National Laboratory (PNNL), to write a Remediation Guidance for Major Airports After a Chemical Attack. The report identifies key activities and issues that should be considered by a typical major airport following an incident involving release of a toxic chemical agent. Four experimental tasks were identified that would require further research in order to supplement the Remediation Guidance. One of the tasks, Task 4, OTD Chemical Remediation Statistical Sampling Design Validation, dealt with statisticalmore » sampling algorithm validation. This report documents the results of the sampling design validation conducted for Task 4. In 2005, the Government Accountability Office (GAO) performed a review of the past U.S. responses to Anthrax terrorist cases. Part of the motivation for this PNNL report was a major GAO finding that there was a lack of validated sampling strategies in the U.S. response to Anthrax cases. The report (GAO 2005) recommended that probability-based methods be used for sampling design in order to address confidence in the results, particularly when all sample results showed no remaining contamination. The GAO also expressed a desire that the methods be validated, which is the main purpose of this PNNL report. The objective of this study was to validate probability-based statistical sampling designs and the algorithms pertinent to within-building sampling that allow the user to prescribe or evaluate confidence levels of conclusions based on data collected as guided by the statistical sampling designs. Specifically, the designs found in the Visual Sample Plan (VSP) software were evaluated. VSP was used to calculate the number of samples and the sample location for a variety of sampling plans applied to an actual release site. Most of the sampling designs validated are probability based, meaning samples are located randomly (or on a randomly placed grid) so no bias enters into the placement of samples, and the number of samples is calculated such that IF the amount and spatial extent of contamination exceeds levels of concern, at least one of the samples would be taken from a contaminated area, at least X% of the time. Hence, "validation" of the statistical sampling algorithms is defined herein to mean ensuring that the "X%" (confidence) is actually met.« less

  20. Chemical freezeout parameters within generic nonextensive statistics

    NASA Astrophysics Data System (ADS)

    Tawfik, Abdel; Yassin, Hayam; Abo Elyazeed, Eman R.

    2018-06-01

    The particle production in relativistic heavy-ion collisions seems to be created in a dynamically disordered system which can be best described by an extended exponential entropy. In distinguishing between the applicability of this and Boltzmann-Gibbs (BG) in generating various particle-ratios, generic (non)extensive statistics is introduced to the hadron resonance gas model. Accordingly, the degree of (non)extensivity is determined by the possible modifications in the phase space. Both BG extensivity and Tsallis nonextensivity are included as very special cases defined by specific values of the equivalence classes (c, d). We found that the particle ratios at energies ranging between 3.8 and 2760 GeV are best reproduced by nonextensive statistics, where c and d range between ˜ 0.9 and ˜ 1 . The present work aims at illustrating that the proposed approach is well capable to manifest the statistical nature of the system on interest. We don't aim at highlighting deeper physical insights. In other words, while the resulting nonextensivity is neither BG nor Tsallis, the freezeout parameters are found very compatible with BG and accordingly with the well-known freezeout phase-diagram, which is in an excellent agreement with recent lattice calculations. We conclude that the particle production is nonextensive but should not necessarily be accompanied by a radical change in the intensive or extensive thermodynamic quantities, such as internal energy and temperature. Only, the two critical exponents defining the equivalence classes (c, d) are the physical parameters characterizing the (non)extensivity.

  1. Community-Oriented Counterterrorism: Incorporating National Homeland Security Mandates into the Local Community Policing Philosophy

    DTIC Science & Technology

    2014-12-01

    both 2000 and 2007 Bw·eau of Justice Statistics Law Enforcement Management and Administrative Statistics sw’Veys. These agencies incmporate most...responded to a variety of community policing and homeland security questions in both 2000 and 2007 Bureau of Justice Statistics Law Enforcement...Management and Administrative Statistics surveys. These agencies incorporate most major U.S. police departments as well as a representative sample of smaller

  2. Formulating appropriate statistical hypotheses for treatment comparison in clinical trial design and analysis.

    PubMed

    Huang, Peng; Ou, Ai-hua; Piantadosi, Steven; Tan, Ming

    2014-11-01

    We discuss the problem of properly defining treatment superiority through the specification of hypotheses in clinical trials. The need to precisely define the notion of superiority in a one-sided hypothesis test problem has been well recognized by many authors. Ideally designed null and alternative hypotheses should correspond to a partition of all possible scenarios of underlying true probability models P={P(ω):ω∈Ω} such that the alternative hypothesis Ha={P(ω):ω∈Ωa} can be inferred upon the rejection of null hypothesis Ho={P(ω):ω∈Ω(o)} However, in many cases, tests are carried out and recommendations are made without a precise definition of superiority or a specification of alternative hypothesis. Moreover, in some applications, the union of probability models specified by the chosen null and alternative hypothesis does not constitute a completed model collection P (i.e., H(o)∪H(a) is smaller than P). This not only imposes a strong non-validated assumption of the underlying true models, but also leads to different superiority claims depending on which test is used instead of scientific plausibility. Different ways to partition P fro testing treatment superiority often have different implications on sample size, power, and significance in both efficacy and comparative effectiveness trial design. Such differences are often overlooked. We provide a theoretical framework for evaluating the statistical properties of different specification of superiority in typical hypothesis testing. This can help investigators to select proper hypotheses for treatment comparison inclinical trial design. Copyright © 2014 Elsevier Inc. All rights reserved.

  3. Multilevel discretized random field models with 'spin' correlations for the simulation of environmental spatial data

    NASA Astrophysics Data System (ADS)

    Žukovič, Milan; Hristopulos, Dionissios T.

    2009-02-01

    A current problem of practical significance is how to analyze large, spatially distributed, environmental data sets. The problem is more challenging for variables that follow non-Gaussian distributions. We show by means of numerical simulations that the spatial correlations between variables can be captured by interactions between 'spins'. The spins represent multilevel discretizations of environmental variables with respect to a number of pre-defined thresholds. The spatial dependence between the 'spins' is imposed by means of short-range interactions. We present two approaches, inspired by the Ising and Potts models, that generate conditional simulations of spatially distributed variables from samples with missing data. Currently, the sampling and simulation points are assumed to be at the nodes of a regular grid. The conditional simulations of the 'spin system' are forced to respect locally the sample values and the system statistics globally. The second constraint is enforced by minimizing a cost function representing the deviation between normalized correlation energies of the simulated and the sample distributions. In the approach based on the Nc-state Potts model, each point is assigned to one of Nc classes. The interactions involve all the points simultaneously. In the Ising model approach, a sequential simulation scheme is used: the discretization at each simulation level is binomial (i.e., ± 1). Information propagates from lower to higher levels as the simulation proceeds. We compare the two approaches in terms of their ability to reproduce the target statistics (e.g., the histogram and the variogram of the sample distribution), to predict data at unsampled locations, as well as in terms of their computational complexity. The comparison is based on a non-Gaussian data set (derived from a digital elevation model of the Walker Lake area, Nevada, USA). We discuss the impact of relevant simulation parameters, such as the domain size, the number of discretization levels, and the initial conditions.

  4. Three-dimensional image contrast using biospeckle

    NASA Astrophysics Data System (ADS)

    Godinho, Robson Pierangeli; Braga, Roberto A., Jr.

    2010-09-01

    The biospeckle laser (BSL) has been applied in many areas of knowledge and a variety of approaches has been presented to address the best results in biological and non-biological samples, in fast or slow activities, or else in defined flow of materials or in random activities. The methodologies accounted in the literature consider the apparatus used in the image assembling and the way the collected data is processed. The image processing steps presents in turn a variety of procedures with first or second order statistics analysis, and as well with different sizes of data collected. One way to access the biospeckle in defined flow, such as in capillary blood flow in alive animals, was the adoption of the image contrast technique which uses only one image from the illuminated sample. That approach presents some problems related to the resolution of the image, which is reduced during the image contrast processing. In order to help the visualization of the low resolution image formed by the contrast technique, this work presents the three-dimensional procedure as a reliable alternative to enhance the final image. The work based on a parallel processing, with the generation of a virtual map of amplitudes, and maintaining the quasi-online characteristic of the contrast technique. Therefore, it was possible to generate in the same display the observed material, the image contrast result and in addiction the three-dimensional image with adjustable options of rotation. The platform also offers to the user the possibility to access the 3D image offline.

  5. Latent Classes of Symptoms related to Clinically Depressed Mood in Adolescents.

    PubMed

    Blom, Eva Henje; Forsman, Mats; Yang, Tony T; Serlachius, Eva; Larsson, Jan-Olov

    2014-01-01

    The diagnosis of major depressive disorder (MDD), according to the Diagnostic and Statistical Manual of Mental Disorders , is based only on adult symptomatology of depression and not adapted for age and gender. This may contribute to the low diagnostic specificity and validity of adolescent MDD. In this study, we investigated whether latent classes based on symptoms associated with depressed mood could be identified in a sample of adolescents seeking psychiatric care, regardless of traditionally defined diagnostic categories. Self-reports of the Strengths and Difficulties Questionnaire and the Development and Well-Being Assessment were collected consecutively from all new patients between the ages of 13 and 17 years at two psychiatric outpatient clinics in Stockholm, Sweden. Those who reported depressed mood at intake yielded a sample of 21 boys and 156 girls. Latent class analyses were performed for all screening items and for the depression-specific items of the Development and Well-Being Assessment. The symptoms that were reported in association with depressed mood differentiated the adolescents into two classes. One class had moderate emotional severity scores on the Strengths and Difficulties Questionnaire and mainly symptoms that were congruent with the Diagnostic and Statistical Manual of Mental Disorders criteria for MDD. The other class had higher emotional severity scores and similar symptoms to those reported in the first class. However, in addition, this group demonstrated more diverse symptomatology, including vegetative symptoms, suicidal ideation, anxiety, conduct problems, body dysmorphic symptoms, and deliberate vomiting. The classes predicted functional impairment in that the members of the second class showed more functional impairment. The relatively small sample size limited the generalizability of the results of this study, and the amount of items included in the analysis was restricted by the rules of latent class analysis. No conclusions about gender differences between the classes could be could be drawn as a result of the low number of boys included in the study. Two distinct classes were identified among adolescents with depressed mood. The class with highest emotional symptom severity score and the most functional impairment had a more diverse symptomatology that included symptoms that were not congruent with the traditional diagnostic criteria of MDD. However, this additional symptomatology is clinically important to consider. As a result, the clinical usefulness of the Diagnostic and Statistical Manual of Mental Disorders during the diagnostic process of adolescent depression is questioned.

  6. A cautionary note on substituting spatial subunits for repeated temporal sampling in studies of site occupancy

    USGS Publications Warehouse

    Kendall, William L.; White, Gary C.

    2009-01-01

    1. Assessing the probability that a given site is occupied by a species of interest is important to resource managers, as well as metapopulation or landscape ecologists. Managers require accurate estimates of the state of the system, in order to make informed decisions. Models that yield estimates of occupancy, while accounting for imperfect detection, have proven useful by removing a potentially important source of bias. To account for detection probability, multiple independent searches per site for the species are required, under the assumption that the species is available for detection during each search of an occupied site. 2. We demonstrate that when multiple samples per site are defined by searching different locations within a site, absence of the species from a subset of these spatial subunits induces estimation bias when locations are exhaustively assessed or sampled without replacement. 3. We further demonstrate that this bias can be removed by choosing sampling locations with replacement, or if the species is highly mobile over a short period of time. 4. Resampling an existing data set does not mitigate bias due to exhaustive assessment of locations or sampling without replacement. 5. Synthesis and applications. Selecting sampling locations for presence/absence surveys with replacement is practical in most cases. Such an adjustment to field methods will prevent one source of bias, and therefore produce more robust statistical inferences about species occupancy. This will in turn permit managers to make resource decisions based on better knowledge of the state of the system.

  7. Periodontal Research: Basics and beyond – Part II (Ethical issues, sampling, outcome measures and bias)

    PubMed Central

    Avula, Haritha

    2013-01-01

    A good research beginning refers to formulating a well-defined research question, developing a hypothesis and choosing an appropriate study design. The first part of the review series has discussed these issues in depth and this paper intends to throw light on other issues pertaining to the implementation of research. These include the various ethical norms and standards in human experimentation, the eligibility criteria for the participants, sampling methods and sample size calculation, various outcome measures that need to be defined and the biases that can be introduced in research. PMID:24174747

  8. Reduction of Complications of Local Anaesthesia in Dental Healthcare Setups by Application of the Six Sigma Methodology: A Statistical Quality Improvement Technique

    PubMed Central

    Khatoon, Farheen

    2015-01-01

    Background Health care faces challenges due to complications, inefficiencies and other concerns that threaten the safety of patients. Aim The purpose of his study was to identify causes of complications encountered after administration of local anaesthesia for dental and oral surgical procedures and to reduce the incidence of complications by introduction of six sigma methodology. Materials and Methods DMAIC (Define, Measure, Analyse, Improve and Control) process of Six Sigma was taken into consideration to reduce the incidence of complications encountered after administration of local anaesthesia injections for dental and oral surgical procedures using failure mode and effect analysis. Pareto analysis was taken into consideration to analyse the most recurring complications. Paired z-sample test using Minitab Statistical Inference and Fisher’s exact test was used to statistically analyse the obtained data. The p-value <0.05 was considered as significant value. Results Total 54 systemic and 62 local complications occurred during three months of analyse and measure phase. Syncope, failure of anaesthesia, trismus, auto mordeduras and pain at injection site was found to be most recurring complications. Cumulative defective percentage was 7.99 in case of pre-improved data and decreased to 4.58 in the control phase. Estimate for difference was 0.0341228 and 95% lower bound for difference was 0.0193966. p-value was found to be highly significant with p= 0.000. Conclusion The application of six sigma improvement methodology in healthcare tends to deliver consistently better results to the patients as well as hospitals and results in better patient compliance as well as satisfaction. PMID:26816989

  9. Rankings Methodology Hurts Public Institutions

    ERIC Educational Resources Information Center

    Van Der Werf, Martin

    2007-01-01

    In the 1980s, when the "U.S. News & World Report" rankings of colleges were based solely on reputation, the nation's public universities were well represented at the top. However, as soon as the magazine began including its "measures of excellence," statistics intended to define quality, public universities nearly disappeared from the top. As the…

  10. HyperCard Monitor System.

    ERIC Educational Resources Information Center

    Harris, Julian; Maurer, Hermann

    An investigation into high level event monitoring within the scope of a well-known multimedia application, HyperCard--a program on the Macintosh computer, is carried out. A monitoring system is defined as a system which automatically monitors usage of some activity and gathers statistics based on what is has observed. Monitor systems can give the…

  11. Symptoms versus Impairment: The Case for Respecting "DSM-IV"'s Criterion D

    ERIC Educational Resources Information Center

    Gordon, Michael; Antshel, Kevin; Faraone, Stephen; Barkley, Russell; Lewandowski, Larry; Hudziak, James J.; Biederman, Joseph; Cunningham, Charles

    2006-01-01

    Diagnosing ADHD based primarily on symptom reports assumes that the number/frequency of symptoms is tied closely to the impairment imposed on an individual's functioning. That presumed linkage encourages diagnosis more by "Diagnostic and Statistical Manual of Mental Disorders" (4th ed.) style symptom lists than well-defined,…

  12. Sampling of prenatal and postnatal offspring from individual rat dams enhances animal use without compromising development

    NASA Technical Reports Server (NTRS)

    Alberts, J. R.; Burden, H. W.; Hawes, N.; Ronca, A. E.

    1996-01-01

    To assess prenatal and postnatal developmental status in the offspring of a group of animals, it is typical to examine fetuses from some of the dams as well as infants born to the remaining dams. Statistical limitations often arise, particularly when the animals are rare or especially precious, because all offspring of the dam represent only a single statistical observation; littermates are not independent observations (biologically or statistically). We describe a study in which pregnant laboratory rats were laparotomized on day 7 of gestation (GD7) to ascertain the number and distribution of uterine implantation sites and were subjected to a simulated experience on a 10-day space shuttle flight. After the simulated landing on GD18, rats were unilaterally hysterectomized, thus providing a sample of fetuses from 10 independent uteruses, followed by successful vaginal delivery on GD22, yielding postnatal samples from 10 uteruses. A broad profile of maternal and offspring morphologic and physiologic measures indicated that these novel sampling procedures did not compromise maternal well-being and maintained normal offspring development and function. Measures included maternal organ weights and hormone concentrations, offspring body size, growth, organ weights, sexual differentiation, and catecholamine concentrations.

  13. Trends in groundwater quality in principal aquifers of the United States, 1988-2012

    USGS Publications Warehouse

    Lindsey, Bruce D.; Rupert, Michael G.

    2014-01-01

    The U.S. Geological Survey (USGS) National Water-Quality Assessment (NAWQA) Program analyzed trends in groundwater quality throughout the nation for the sampling period of 1988-2012. Trends were determined for networks (sets of wells routinely monitored by the USGS) for a subset of constituents by statistical analysis of paired water-quality measurements collected on a near-decadal time scale. The data set for chloride, dissolved solids, and nitrate consisted of 1,511 wells in 67 networks, whereas the data set for methyl tert-butyl ether (MTBE) consisted of 1, 013 wells in 46 networks. The 25 principal aquifers represented by these networks account for about 75 percent of withdrawals of groundwater used for drinking-water supply for the nation. Statistically significant changes in chloride, dissolved-solids, or nitrate concentrations were found in many well networks over a decadal period. Concentrations increased significantly in 48 percent of networks for chloride, 42 percent of networks for dissolved solids, and 21 percent of networks for nitrate. Chloride, dissolved solids, and nitrate concentrations decreased significantly in 3, 3, and 10 percent of the networks, respectively. The magnitude of change in concentrations was typically small in most networks; however, the magnitude of change in networks with statistically significant increases was typically much larger than the magnitude of change in networks with statistically significant decreases. The largest increases of chloride concentrations were in urban areas in the northeastern and north central United States. The largest increases of nitrate concentrations were in networks in agricultural areas. Statistical analysis showed 42 or the 46 networks had no statistically significant changes in MTBE concentrations. The four networks with statistically significant changes in MTBE concentrations were in the northeastern United States, where MTBE was widely used. Two networks had increasing concentrations, and two networks had decreasing concentrations. Production and use of MTBE peaked in about 2000 and has been effectively banned in many areas since about 2006. The two networks that had increasing concentrations were sampled for the second time close to the peak of MTBE production, whereas the two networks that had decreasing concentrations were sampled for the second time 10 years after the peak of MTBE production.

  14. Consideraciones para la estimacion de abundancia de poblaciones de mamiferos. [Considerations for the estimation of abundance of mammal populations.

    USGS Publications Warehouse

    Walker, R.S.; Novare, A.J.; Nichols, J.D.

    2000-01-01

    Estimation of abundance of mammal populations is essential for monitoring programs and for many ecological investigations. The first step for any study of variation in mammal abundance over space or time is to define the objectives of the study and how and why abundance data are to be used. The data used to estimate abundance are count statistics in the form of counts of animals or their signs. There are two major sources of uncertainty that must be considered in the design of the study: spatial variation and the relationship between abundance and the count statistic. Spatial variation in the distribution of animals or signs may be taken into account with appropriate spatial sampling. Count statistics may be viewed as random variables, with the expected value of the count statistic equal to the true abundance of the population multiplied by a coefficient p. With direct counts, p represents the probability of detection or capture of individuals, and with indirect counts it represents the rate of production of the signs as well as their probability of detection. Comparisons of abundance using count statistics from different times or places assume that the p are the same for all times or places being compared (p= pi). In spite of considerable evidence that this assumption rarely holds true, it is commonly made in studies of mammal abundance, as when the minimum number alive or indices based on sign counts are used to compare abundance in different habitats or times. Alternatives to relying on this assumption are to calibrate the index used by testing the assumption of p= pi, or to incorporate the estimation of p into the study design.

  15. Emergent irreversibility and entanglement spectrum statistics

    NASA Astrophysics Data System (ADS)

    Mucciolo, Eduardo; Chamon, Claudio; Hamma, Alioscia

    2014-03-01

    We study the problem of irreversibility when the dynamical evolution of a many-body system is described by a stochastic quantum circuit. Such evolution is more general than Hamitonian, and since energy levels are not well defined, the well-established connection between the statistical fluctuations of the energy spectrum and irreversibility cannot be made. We show that the entanglement spectrum provides a more general connection. Irreversibility is marked by a failure of a disentangling algorithm and is preceded by the appearance of Wigner-Dyson statistical fluctuations in the entanglement spectrum. This analysis can be done at the wavefunction level and offers a new route to study quantum chaos and quantum integrability. We acknowledge financial support from the U.S. National Science Foundation through grants CCF 1116590 and CCF 1117241, from the National Basic Research Program of China through grants 2011CBA00300 and 2011CBA00301, and from the National Natural Science Fo.

  16. The effects of sampling frequency on the climate statistics of the European Centre for Medium-Range Weather Forecasts

    NASA Astrophysics Data System (ADS)

    Phillips, Thomas J.; Gates, W. Lawrence; Arpe, Klaus

    1992-12-01

    The effects of sampling frequency on the first- and second-moment statistics of selected European Centre for Medium-Range Weather Forecasts (ECMWF) model variables are investigated in a simulation of "perpetual July" with a diurnal cycle included and with surface and atmospheric fields saved at hourly intervals. The shortest characteristic time scales (as determined by the e-folding time of lagged autocorrelation functions) are those of ground heat fluxes and temperatures, precipitation and runoff, convective processes, cloud properties, and atmospheric vertical motion, while the longest time scales are exhibited by soil temperature and moisture, surface pressure, and atmospheric specific humidity, temperature, and wind. The time scales of surface heat and momentum fluxes and of convective processes are substantially shorter over land than over oceans. An appropriate sampling frequency for each model variable is obtained by comparing the estimates of first- and second-moment statistics determined at intervals ranging from 2 to 24 hours with the "best" estimates obtained from hourly sampling. Relatively accurate estimation of first- and second-moment climate statistics (10% errors in means, 20% errors in variances) can be achieved by sampling a model variable at intervals that usually are longer than the bandwidth of its time series but that often are shorter than its characteristic time scale. For the surface variables, sampling at intervals that are nonintegral divisors of a 24-hour day yields relatively more accurate time-mean statistics because of a reduction in errors associated with aliasing of the diurnal cycle and higher-frequency harmonics. The superior estimates of first-moment statistics are accompanied by inferior estimates of the variance of the daily means due to the presence of systematic biases, but these probably can be avoided by defining a different measure of low-frequency variability. Estimates of the intradiurnal variance of accumulated precipitation and surface runoff also are strongly impacted by the length of the storage interval. In light of these results, several alternative strategies for storage of the EMWF model variables are recommended.

  17. Cellular network entropy as the energy potential in Waddington's differentiation landscape

    PubMed Central

    Banerji, Christopher R. S.; Miranda-Saavedra, Diego; Severini, Simone; Widschwendter, Martin; Enver, Tariq; Zhou, Joseph X.; Teschendorff, Andrew E.

    2013-01-01

    Differentiation is a key cellular process in normal tissue development that is significantly altered in cancer. Although molecular signatures characterising pluripotency and multipotency exist, there is, as yet, no single quantitative mark of a cellular sample's position in the global differentiation hierarchy. Here we adopt a systems view and consider the sample's network entropy, a measure of signaling pathway promiscuity, computable from a sample's genome-wide expression profile. We demonstrate that network entropy provides a quantitative, in-silico, readout of the average undifferentiated state of the profiled cells, recapitulating the known hierarchy of pluripotent, multipotent and differentiated cell types. Network entropy further exhibits dynamic changes in time course differentiation data, and in line with a sample's differentiation stage. In disease, network entropy predicts a higher level of cellular plasticity in cancer stem cell populations compared to ordinary cancer cells. Importantly, network entropy also allows identification of key differentiation pathways. Our results are consistent with the view that pluripotency is a statistical property defined at the cellular population level, correlating with intra-sample heterogeneity, and driven by the degree of signaling promiscuity in cells. In summary, network entropy provides a quantitative measure of a cell's undifferentiated state, defining its elevation in Waddington's landscape. PMID:24154593

  18. Overweight is associated with low hemoglobin levels in adolescent girls.

    PubMed

    Bagni, Ursula Viana; Luiz, Ronir Raggio; Veiga, Gloria Valeria da

    2013-01-01

    To verify the prevalence of iron deficiency anemia according to sexual maturation stages and its association with overweight as well as excessive body fat in adolescents. A school-based cross-sectional study was performed. Anemia was assessed by measuring the hemoglobin level (Hb). Nutritional status was defined by sex and age specific body mass index (BMI) cutoffs, and body fat (BF) was determined by bioelectrical impedance. Sexual maturation was assessed by breasts/genitalia and pubic hair development stages. Statistical analyses considered the effect of cluster sampling design (classes) and sampling expansion corrected by relative weight. Odds ratio and general linear modeling were used to assess the associations, regarding the value of p < 0.05 for statistical significance. Public schools in the Metropolitan area of Rio de Janeiro, Brazil. Probabilistic sample of 707 teenagers between 11.0 and 19.9 years old. The prevalence of anemia among the adolescents was 22.8% (95%CI 16.7-30.2%), higher among girls than among boys (30.9% vs. 10.9%; p < 0.01). The chance of developing anemia did not change with the nutritional status according BMI or BF percentage, however, overweight girls presented lower Hb levels than those who were not overweight (12.2 g/dL vs. 12.8 g/dL, p < 0.01). In boys this association was not observed. Sexual maturation did not change the association of Hb and anemia with overweight and excessive body fat. The reduction of Hb levels points at overweight as a risk factor for the development of iron deficiency among adolescents. © 2012 Asian Oceanian Association for the Study of Obesity. Published by Elsevier Ltd. All rights reserved.

  19. Comparison of semen parameters in samples collected by masturbation at a clinic and at home.

    PubMed

    Elzanaty, Saad; Malm, Johan

    2008-06-01

    To investigate differences in semen quality between samples collected by masturbation at a clinic and at home. Cross-sectional study. Fertility center. Three hundred seventy-nine men assessed for infertility. None. Semen was analyzed according to World Health Organization guidelines. Seminal markers of epididymal (neutral alpha-glucosidase), prostatic (prostate-specific antigen and zinc), and seminal vesicle (fructose) function were measured. Two patient groups were defined according to sample collection location: at a clinic (n = 273) or at home (n = 106). Compared with clinic-collected semen, home-collected samples had statistically significantly higher values for sperm concentration, total sperm count, rapid progressive motility, and total count of progressive motility. Semen volume, proportion of normal sperm morphology, neutral alpha-glucosidase, prostate-specific antigen, zinc, and fructose did not differ significantly between groups. An abnormal sperm concentration (<20 x 10(6)/mL) was seen in statistically significantly fewer of the samples obtained at home (19/106, 18%) than at the clinic (81/273, 30%), and the same applied to proportions of samples with abnormal (< 25%) rapid progressive motility (68/106 [64%] and 205/273 [75%], respectively). The present results demonstrate superior semen quality in samples collected by masturbation at home compared with at a clinic. This should be taken into consideration in infertility investigations.

  20. Cleanroom certification model

    NASA Technical Reports Server (NTRS)

    Currit, P. A.

    1983-01-01

    The Cleanroom software development methodology is designed to take the gamble out of product releases for both suppliers and receivers of the software. The ingredients of this procedure are a life cycle of executable product increments, representative statistical testing, and a standard estimate of the MTTF (Mean Time To Failure) of the product at the time of its release. A statistical approach to software product testing using randomly selected samples of test cases is considered. A statistical model is defined for the certification process which uses the timing data recorded during test. A reasonableness argument for this model is provided that uses previously published data on software product execution. Also included is a derivation of the certification model estimators and a comparison of the proposed least squares technique with the more commonly used maximum likelihood estimators.

  1. A random-sum Wilcoxon statistic and its application to analysis of ROC and LROC data.

    PubMed

    Tang, Liansheng Larry; Balakrishnan, N

    2011-01-01

    The Wilcoxon-Mann-Whitney statistic is commonly used for a distribution-free comparison of two groups. One requirement for its use is that the sample sizes of the two groups are fixed. This is violated in some of the applications such as medical imaging studies and diagnostic marker studies; in the former, the violation occurs since the number of correctly localized abnormal images is random, while in the latter the violation is due to some subjects not having observable measurements. For this reason, we propose here a random-sum Wilcoxon statistic for comparing two groups in the presence of ties, and derive its variance as well as its asymptotic distribution for large sample sizes. The proposed statistic includes the regular Wilcoxon rank-sum statistic. Finally, we apply the proposed statistic for summarizing location response operating characteristic data from a liver computed tomography study, and also for summarizing diagnostic accuracy of biomarker data.

  2. Limitations of Poisson statistics in describing radioactive decay.

    PubMed

    Sitek, Arkadiusz; Celler, Anna M

    2015-12-01

    The assumption that nuclear decays are governed by Poisson statistics is an approximation. This approximation becomes unjustified when data acquisition times longer than or even comparable with the half-lives of the radioisotope in the sample are considered. In this work, the limits of the Poisson-statistics approximation are investigated. The formalism for the statistics of radioactive decay based on binomial distribution is derived. The theoretical factor describing the deviation of variance of the number of decays predicated by the Poisson distribution from the true variance is defined and investigated for several commonly used radiotracers such as (18)F, (15)O, (82)Rb, (13)N, (99m)Tc, (123)I, and (201)Tl. The variance of the number of decays estimated using the Poisson distribution is significantly different than the true variance for a 5-minute observation time of (11)C, (15)O, (13)N, and (82)Rb. Durations of nuclear medicine studies often are relatively long; they may be even a few times longer than the half-lives of some short-lived radiotracers. Our study shows that in such situations the Poisson statistics is unsuitable and should not be applied to describe the statistics of the number of decays in radioactive samples. However, the above statement does not directly apply to counting statistics at the level of event detection. Low sensitivities of detectors which are used in imaging studies make the Poisson approximation near perfect. Copyright © 2015 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  3. Intrinsic Subtype and Therapeutic Response Among HER2-Positive Breast Tumors from the NCCTG (Alliance) N9831 Trial

    PubMed Central

    Perez, Edith A.; Ballman, Karla V.; Mashadi-Hossein, Afshin; Tenner, Kathleen S.; Kachergus, Jennifer M.; Norton, Nadine; Necela, Brian M.; Carr, Jennifer M.; Ferree, Sean; Perou, Charles M.; Baehner, Frederick; Cheang, Maggie Chon U.

    2017-01-01

    Background: Genomic data from human epidermal growth factor receptor 2–positive (HER2+) tumors were analyzed to assess the association between intrinsic subtype and clinical outcome in a large, well-annotated patient cohort. Methods: Samples from the NCCTG (Alliance) N9831 trial were analyzed using the Prosigna algorithm on the NanoString platform to define intrinsic subtype, risk of recurrence scores, and risk categories for 1392 HER2+ tumors. Subtypes were evaluated for recurrence-free survival (RFS) using Kaplan-Meier and Cox model analysis following adjuvant chemotherapy (n = 484) or chemotherapy plus trastuzumab (n = 908). All statistical tests were two-sided. Results: Patients with HER2+ tumors from N9831 were primarily scored as HER2-enriched (72.1%). These individuals received statistically significant benefit from trastuzumab (hazard ratio [HR] = 0.68, 95% confidence interval [CI] = 0.52 to 0.89, P = .005), as did the patients (291 of 1392) with luminal-type tumors (HR = 0.52, 95% CI = 0.32 to 0.85, P = .01). Patients with basal-like tumors (97 of 1392) did not have statistically significantly better RFS when treated with trastuzumab and chemotherapy compared with chemotherapy alone (HR = 1.06, 95% CI = 0.53 to 2.13, P = .87). Conclusions: The majority of clinically defined HER2-positive tumors were classified as HER2-enriched or luminal using the Prosigna algorithm. Intrinsic subtype alone cannot replace conventional histopathological evaluation of HER2 status because many tumors that are classified as luminal A or luminal B will benefit from adjuvant trastuzumab if that subtype is accompanied by HER2 overexpression. However, among tumors that overexpress HER2, we speculate that assessment of intrinsic subtype may influence treatment, particularly with respect to evaluating alternative therapeutic approaches for that subset of HER2-positive tumors of the basal-like subtype. PMID:27794124

  4. Intrinsic Subtype and Therapeutic Response Among HER2-Positive Breaty st Tumors from the NCCTG (Alliance) N9831 Trial.

    PubMed

    Perez, Edith A; Ballman, Karla V; Mashadi-Hossein, Afshin; Tenner, Kathleen S; Kachergus, Jennifer M; Norton, Nadine; Necela, Brian M; Carr, Jennifer M; Ferree, Sean; Perou, Charles M; Baehner, Frederick; Cheang, Maggie Chon U; Thompson, E Aubrey

    2017-02-01

    Genomic data from human epidermal growth factor receptor 2-positive (HER2+) tumors were analyzed to assess the association between intrinsic subtype and clinical outcome in a large, well-annotated patient cohort. Samples from the NCCTG (Alliance) N9831 trial were analyzed using the Prosigna algorithm on the NanoString platform to define intrinsic subtype, risk of recurrence scores, and risk categories for 1392 HER2+ tumors. Subtypes were evaluated for recurrence-free survival (RFS) using Kaplan-Meier and Cox model analysis following adjuvant chemotherapy (n = 484) or chemotherapy plus trastuzumab (n = 908). All statistical tests were two-sided. Patients with HER2+ tumors from N9831 were primarily scored as HER2-enriched (72.1%). These individuals received statistically significant benefit from trastuzumab (hazard ratio [HR] = 0.68, 95% confidence interval [CI] = 0.52 to 0.89, P = .005), as did the patients (291 of 1392) with luminal-type tumors (HR = 0.52, 95% CI = 0.32 to 0.85, P = .01). Patients with basal-like tumors (97 of 1392) did not have statistically significantly better RFS when treated with trastuzumab and chemotherapy compared with chemotherapy alone (HR = 1.06, 95% CI = 0.53 to 2.13, P = .87). The majority of clinically defined HER2-positive tumors were classified as HER2-enriched or luminal using the Prosigna algorithm. Intrinsic subtype alone cannot replace conventional histopathological evaluation of HER2 status because many tumors that are classified as luminal A or luminal B will benefit from adjuvant trastuzumab if that subtype is accompanied by HER2 overexpression. However, among tumors that overexpress HER2, we speculate that assessment of intrinsic subtype may influence treatment, particularly with respect to evaluating alternative therapeutic approaches for that subset of HER2-positive tumors of the basal-like subtype. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  5. Constructing Sample Space with Combinatorial Reasoning: A Mixed Methods Study

    ERIC Educational Resources Information Center

    McGalliard, William A., III.

    2012-01-01

    Recent curricular developments suggest that students at all levels need to be statistically literate and able to efficiently and accurately make probabilistic decisions. Furthermore, statistical literacy is a requirement to being a well-informed citizen of society. Research also recognizes that the ability to reason probabilistically is supported…

  6. Statistical Rules-of-Thumb.

    ERIC Educational Resources Information Center

    Brewer, James K.

    1988-01-01

    Six best-selling introductory behavioral statistics textbooks that were published in 1982 and two well-known sampling theory textbooks were reviewed to determine the presence of rules-of-thumb--useful principles with wide application that are not intended to be strictly accurate. The relative frequency and type of rules are reported along with a…

  7. Publication bias in situ

    PubMed Central

    Phillips, Carl V

    2004-01-01

    Background Publication bias, as typically defined, refers to the decreased likelihood of studies' results being published when they are near the null, not statistically significant, or otherwise "less interesting." But choices about how to analyze the data and which results to report create a publication bias within the published results, a bias I label "publication bias in situ" (PBIS). Discussion PBIS may create much greater bias in the literature than traditionally defined publication bias (the failure to publish any result from a study). The causes of PBIS are well known, consisting of various decisions about reporting that are influenced by the data. But its impact is not generally appreciated, and very little attention is devoted to it. What attention there is consists largely of rules for statistical analysis that are impractical and do not actually reduce the bias in reported estimates. PBIS cannot be reduced by statistical tools because it is not fundamentally a problem of statistics, but rather of non-statistical choices and plain language interpretations. PBIS should be recognized as a phenomenon worthy of study – it is extremely common and probably has a huge impact on results reported in the literature – and there should be greater systematic efforts to identify and reduce it. The paper presents examples, including results of a recent HIV vaccine trial, that show how easily PBIS can have a large impact on reported results, as well as how there can be no simple answer to it. Summary PBIS is a major problem, worthy of substantially more attention than it receives. There are ways to reduce the bias, but they are very seldom employed because they are largely unrecognized. PMID:15296515

  8. Status of groundwater quality in the Southern, Middle, and Northern Sacramento Valley study units, 2005-08: California GAMA Priority Basin Project

    USGS Publications Warehouse

    Bennett, George L.; Fram, Miranda S.; Belitz, Kenneth

    2011-01-01

    Groundwater quality in the Southern, Middle, and Northern Sacramento Valley study units was investigated as part of the Priority Basin Project of the Groundwater Ambient Monitoring and Assessment (GAMA) Program. The study units are located in California's Central Valley and include parts of Butte, Colusa, Glenn, Placer, Sacramento, Shasta, Solano, Sutter, Tehama, Yolo, and Yuba Counties. The GAMA Priority Basin Project is being conducted by the California State Water Resources Control Board in collaboration with the U.S. Geological Survey and the Lawrence Livermore National Laboratory. The three study units were designated to provide spatially-unbiased assessments of the quality of untreated groundwater in three parts of the Central Valley hydrogeologic province, as well as to provide a statistically consistent basis for comparing water quality regionally and statewide. Samples were collected in 2005 (Southern Sacramento Valley), 2006 (Middle Sacramento Valley), and 2007-08 (Northern Sacramento Valley). The GAMA studies in the Southern, Middle, and Northern Sacramento Valley were designed to provide statistically robust assessments of the quality of untreated groundwater in the primary aquifer systems that are used for drinking-water supply. The assessments are based on water-quality data collected by the USGS from 235 wells in the three study units in 2005-08, and water-quality data from the California Department of Public Health (CDPH) database. The primary aquifer systems (hereinafter, referred to as primary aquifers) assessed in this study are defined by the depth intervals of the wells in the CDPH database for each study unit. The quality of groundwater in shallow or deep water-bearing zones may differ from quality of groundwater in the primary aquifers; shallow groundwater may be more vulnerable to contamination from the surface. The status of the current quality of the groundwater resource was assessed by using data from samples analyzed for volatile organic compounds (VOC), pesticides, and naturally occurring inorganic constituents, such as major ions and trace elements. This status assessment is intended to characterize the quality of groundwater resources within the primary aquifers of the three Sacramento Valley study units, not the treated drinking water delivered to consumers by water purveyors. Relative-concentrations (sample concentrations divided by benchmark concentrations) were used for evaluating groundwater quality for those constituents that have Federal or California regulatory or non-regulatory benchmarks for drinking-water quality. A relative-concentration greater than 1.0 indicates a concentration greater than a benchmark. For organic (volatile organic compounds and pesticides) and special-interest (perchlorate) constituents, relative-concentrations were classified as high (greater than 1.0); moderate (equal to or less than 1.0 and greater than 0.1); or low (equal to or less than 0.1). For inorganic (major ion, trace element, nutrient, and radioactive) constituents, the boundary between low and moderate relative-concentrations was set at 0.5. Aquifer-scale proportions were used in the status assessment for evaluating regional-scale groundwater quality. High aquifer-scale proportion is defined as the percentage of the area of the primary aquifers that have a relative-concentration greater than 1.0 for a particular constituent or class of constituents; percentage is based on an areal rather than a volumetric basis. Moderate and low aquifer-scale proportions were defined as the percentage of the primary aquifers that have moderate and low relative-concentrations, respectively. Two statistical approaches-grid-based, which used one value per grid cell, and spatially-weighted, which used the full dataset-were used to calculate aquifer-scale proportions for individual constituents and classes of constituents. High and moderate aquifer-scale proportions were significantly greater for inorgani

  9. Growth, Characterization and Applications of Beta-Barium Borate and Related Crystals

    DTIC Science & Technology

    1993-10-31

    Crystal symmetry determines the form of the second order polarization tensor. The second order polarizability tensor is defined by the piezoelectric...cold finger. A temperature oscillation technique1 I was used to limit the number of nuclei formed . These experiments typically yielded thin crystal...statistically sampled to determine the optimal seeding orientation. % was reasoned that the large crystal plates were formed from nucleii which had a favorable

  10. Fall 2014 SEI Research Review Probabilistic Analysis of Time Sensitive Systems

    DTIC Science & Technology

    2014-10-28

    Osmosis SMC Tool Osmosis is a tool for Statistical Model Checking (SMC) with Semantic Importance Sampling. • Input model is written in subset of C...ASSERT() statements in model indicate conditions that must hold. • Input probability distributions defined by the user. • Osmosis returns the...on: – Target relative error, or – Set number of simulations Osmosis Main Algorithm 1 http://dreal.cs.cmu.edu/ (?⃑?): Indicator

  11. Surface topography characterization of brass alloys: lead brass (CuZn39Pb3) and lead free brass (CuZn21Si3P)

    NASA Astrophysics Data System (ADS)

    Reddy, Vijeth V.; Vedantha Krishna, Amogh; Schultheiss, Fredrik; Rosén, B.-G.

    2017-06-01

    Manufactured surfaces usually consist of topographical features which include both those put forth by the manufacturing process, and micro-features caused by disturbances during this process. Surface characterization basically involves study of these features which influence the functionality of the surface. This article focuses on characterization of the surface topography of machined lead brass and lead free brass. The adverse effect of lead on human health and the environment has led the manufacturing sector to focus on sustainable manufacturing of lead free brass, as well as how to maintain control of the surface integrity when substituting the lead content in the brass with silicon. The investigation includes defined areal surface parameters measured on the turned samples of lead- and lead free brass using an optical coherence scanning interferometer, CSI. This paper deals with the study of surface topography of turned samples of lead- and lead free brass. It is important to study the topographical characteristics of the brass samples which are the intermediate link between the manufacturing process variables and the functional behaviour of the surface. To numerically evaluate the sample’s surface topography and to validate the measurements for a significant study, a general statistical methodology is implemented. The results indicate higher surface roughness in turned samples of lead brass compared to lead free brass.

  12. Continuous representation of tumor microvessel density and detection of angiogenic hotspots in histological whole-slide images.

    PubMed

    Kather, Jakob Nikolas; Marx, Alexander; Reyes-Aldasoro, Constantino Carlos; Schad, Lothar R; Zöllner, Frank Gerrit; Weis, Cleo-Aron

    2015-08-07

    Blood vessels in solid tumors are not randomly distributed, but are clustered in angiogenic hotspots. Tumor microvessel density (MVD) within these hotspots correlates with patient survival and is widely used both in diagnostic routine and in clinical trials. Still, these hotspots are usually subjectively defined. There is no unbiased, continuous and explicit representation of tumor vessel distribution in histological whole slide images. This shortcoming distorts angiogenesis measurements and may account for ambiguous results in the literature. In the present study, we describe and evaluate a new method that eliminates this bias and makes angiogenesis quantification more objective and more efficient. Our approach involves automatic slide scanning, automatic image analysis and spatial statistical analysis. By comparing a continuous MVD function of the actual sample to random point patterns, we introduce an objective criterion for hotspot detection: An angiogenic hotspot is defined as a clustering of blood vessels that is very unlikely to occur randomly. We evaluate the proposed method in N=11 images of human colorectal carcinoma samples and compare the results to a blinded human observer. For the first time, we demonstrate the existence of statistically significant hotspots in tumor images and provide a tool to accurately detect these hotspots.

  13. Detecting subtle hydrochemical anomalies with multivariate statistics: an example from homogeneous groundwaters in the Great Artesian Basin, Australia

    NASA Astrophysics Data System (ADS)

    O'Shea, Bethany; Jankowski, Jerzy

    2006-12-01

    The major ion composition of Great Artesian Basin groundwater in the lower Namoi River valley is relatively homogeneous in chemical composition. Traditional graphical techniques have been combined with multivariate statistical methods to determine whether subtle differences in the chemical composition of these waters can be delineated. Hierarchical cluster analysis and principal components analysis were successful in delineating minor variations within the groundwaters of the study area that were not visually identified in the graphical techniques applied. Hydrochemical interpretation allowed geochemical processes to be identified in each statistically defined water type and illustrated how these groundwaters differ from one another. Three main geochemical processes were identified in the groundwaters: ion exchange, precipitation, and mixing between waters from different sources. Both statistical methods delineated an anomalous sample suspected of being influenced by magmatic CO2 input. The use of statistical methods to complement traditional graphical techniques for waters appearing homogeneous is emphasized for all investigations of this type. Copyright

  14. ADRB2 and LEPR gene polymorphisms: synergistic effects on the risk of obesity in Japanese.

    PubMed

    Pereira, Tiago V; Mingroni-Netto, Regina C; Yamada, Yoshiji

    2011-07-01

    The objective of the present study was to validate a recently reported synergistic effect between variants located in the leptin receptor (LEPR) gene and in the β-2 adrenergic receptor (ADRB2) gene on the risk of overweight/obesity. We studied a middle-aged/elderly sample of 4,193 nondiabetic Japanese subjects stratified according gender (1,911 women and 2,282 men). The LEPR Gln223Arg (rs1137101) variant as well as both ADRB2 Arg16Gly (rs1042713) and Gln27Glu (rs1042714) polymorphisms were analyzed. The primary outcome was the risk of overweight/obesity defined as BMI ≥25 kg/m(2), whereas secondary outcomes included the risk of a BMI ≥27 kg/m(2) and BMI as a continuous variable. None of the studied polymorphisms showed statistically significant individual effects, regardless of the group or phenotype studied. Haplotype analysis also did not disclose any associations of ADRB2 polymorphisms with BMI. However, dimensionality reduction-based models confirmed significant interactions among the investigated variants for BMI as a continuous variable as well as for the risk of obesity defined as BMI ≥27 kg/m(2). All disclosed interactions were found in men only. Our results provide external validation for a male specific ADRB2-LEPR interaction effect on the risk of overweight/obesity, but indicate that effect sizes associated with these interactions may be smaller in the population studied.

  15. Integrated Assessment and Improvement of the Quality Assurance System for the Cosworth Casting Process

    NASA Astrophysics Data System (ADS)

    Yousif, Dilon

    The purpose of this study was to improve the Quality Assurance (QA) System at the Nemak Windsor Aluminum Plant (WAP). The project used Six Sigma method based on Define, Measure, Analyze, Improve, and Control (DMAIC). Analysis of in process melt at WAP was based on chemical, thermal, and mechanical testing. The control limits for the W319 Al Alloy were statistically recalculated using the composition measured under stable conditions. The "Chemistry Viewer" software was developed for statistical analysis of alloy composition. This software features the Silicon Equivalency (SiBQ) developed by the IRC. The Melt Sampling Device (MSD) was designed and evaluated at WAP to overcome traditional sampling limitations. The Thermal Analysis "Filters" software was developed for cooling curve analysis of the 3XX Al Alloy(s) using IRC techniques. The impact of low melting point impurities on the start of melting was evaluated using the Universal Metallurgical Simulator and Analyzer (UMSA).

  16. Advanced defect classification by smart sampling, based on sub-wavelength anisotropic scatterometry

    NASA Astrophysics Data System (ADS)

    van der Walle, Peter; Kramer, Esther; Ebeling, Rob; Spruit, Helma; Alkemade, Paul; Pereira, Silvania; van der Donck, Jacques; Maas, Diederik

    2018-03-01

    We report on advanced defect classification using TNO's RapidNano particle scanner. RapidNano was originally designed for defect detection on blank substrates. In detection-mode, the RapidNano signal from nine azimuth angles is added for sensitivity. In review-mode signals from individual angles are analyzed to derive additional defect properties. We define the Fourier coefficient parameter space that is useful to study the statistical variation in defect types on a sample. By selecting defects from each defect type for further review by SEM, information on all defects can be obtained efficiently.

  17. A Study to Determine if a Difference Exists Among the Cumulative Incidence of Acute Respiratory Disease Hospital Admissions of Three Groups of Army Basic Trainees as Defined by the Design of Barracks in Which They Are Housed

    DTIC Science & Technology

    1989-08-01

    number) Using chi-square tests of homogeneity, a selected sample of Army Basic Trainees at Ft. Jackso was studied to determine if there was a...Period of training for sample soldiers was January to May 1985. Results of testing for the female trainees indicated no significant difference in incidence...of ARD among three barracks groups. Results of testing for male trainees indicated statistically significant dif -erences of ARD among each of three

  18. US Geological Survey nutrient preservation experiment : experimental design, statistical analysis, and interpretation of analytical results

    USGS Publications Warehouse

    Patton, Charles J.; Gilroy, Edward J.

    1999-01-01

    Data on which this report is based, including nutrient concentrations in synthetic reference samples determined concurrently with those in real samples, are extensive (greater than 20,000 determinations) and have been published separately. In addition to confirming the well-documented instability of nitrite in acidified samples, this study also demonstrates that when biota are removed from samples at collection sites by 0.45-micrometer membrane filtration, subsequent preservation with sulfuric acid or mercury (II) provides no statistically significant improvement in nutrient concentration stability during storage at 4 degrees Celsius for 30 days. Biocide preservation had no statistically significant effect on the 30-day stability of phosphorus concentrations in whole-water splits from any of the 15 stations, but did stabilize Kjeldahl nitrogen concentrations in whole-water splits from three data-collection stations where ammonium accounted for at least half of the measured Kjeldahl nitrogen.

  19. Living Matter Observations with a Novel Hyperspectral Supercontinuum Confocal Microscope for VIS to Near-IR Reflectance Spectroscopy

    PubMed Central

    Bertani, Francesca R.; Ferrari, Luisa; Mussi, Valentina; Botti, Elisabetta; Costanzo, Antonio; Selci, Stefano

    2013-01-01

    A broad range hyper-spectroscopic microscope fed by a supercontinuum laser source and equipped with an almost achromatic optical layout is illustrated with detailed explanations of the design, implementation and data. The real novelty of this instrument, a confocal spectroscopic microscope capable of recording high resolution reflectance data in the VIS-IR spectral range from about 500 nm to 2.5 μm wavelengths, is the possibility of acquiring spectral data at every physical point as defined by lateral coordinates, X and Y, as well as at a depth coordinate, Z, as obtained by the confocal optical sectioning advantage. With this apparatus we collect each single scanning point as a whole spectrum by combining two linear spectral detector arrays, one CCD for the visible range, and one InGaAs infrared array, simultaneously available at the sensor output channel of the home made instrument. This microscope has been developed for biomedical analysis of human skin and other similar applications. Results are shown illustrating the technical performances of the instrument and the capability in extracting information about the composition and the structure of different parts or compartments in biological samples as well as in solid statematter. A complete spectroscopic fingerprinting of samples at microscopic level is shown possible by using statistical analysis on raw data or analytical reflectance models based on Abelés matrix transfer methods. PMID:24233077

  20. Extreme Mean and Its Applications

    NASA Technical Reports Server (NTRS)

    Swaroop, R.; Brownlow, J. D.

    1979-01-01

    Extreme value statistics obtained from normally distributed data are considered. An extreme mean is defined as the mean of p-th probability truncated normal distribution. An unbiased estimate of this extreme mean and its large sample distribution are derived. The distribution of this estimate even for very large samples is found to be nonnormal. Further, as the sample size increases, the variance of the unbiased estimate converges to the Cramer-Rao lower bound. The computer program used to obtain the density and distribution functions of the standardized unbiased estimate, and the confidence intervals of the extreme mean for any data are included for ready application. An example is included to demonstrate the usefulness of extreme mean application.

  1. Evaluation of air quality in a megacity using statistics tools

    NASA Astrophysics Data System (ADS)

    Ventura, Luciana Maria Baptista; de Oliveira Pinto, Fellipe; Soares, Laiza Molezon; Luna, Aderval Severino; Gioda, Adriana

    2018-06-01

    Local physical characteristics (e.g., meteorology and topography) associate to particle concentrations are important to evaluate air quality in a region. Meteorology and topography affect air pollutant dispersions. This study used statistics tools (PCA, HCA, Kruskal-Wallis, Mann-Whitney's test and others) to a better understanding of the relationship between fine particulate matter (PM2.5) levels and seasons, meteorological conditions and air basins. To our knowledge, it is one of the few studies performed in Latin America involving all parameters together. PM2.5 samples were collected in six sampling sites with different emission sources (industrial, vehicular, soil dust) in Rio de Janeiro, Brazil. The PM2.5 daily concentrations ranged from 1 to 61 µg m-3, with averages higher than the annual limit (15 µg m-3) for some of the sites. The results of the statistics evaluation showed that PM2.5 concentrations were not influenced by seasonality. Furthermore, air basins defined previously were not confirmed, because some sites presented similar emission sources. Therefore, new redefinitions of air basins need to be done, once they are important to air quality management.

  2. Evaluation of air quality in a megacity using statistics tools

    NASA Astrophysics Data System (ADS)

    Ventura, Luciana Maria Baptista; de Oliveira Pinto, Fellipe; Soares, Laiza Molezon; Luna, Aderval Severino; Gioda, Adriana

    2017-03-01

    Local physical characteristics (e.g., meteorology and topography) associate to particle concentrations are important to evaluate air quality in a region. Meteorology and topography affect air pollutant dispersions. This study used statistics tools (PCA, HCA, Kruskal-Wallis, Mann-Whitney's test and others) to a better understanding of the relationship between fine particulate matter (PM2.5) levels and seasons, meteorological conditions and air basins. To our knowledge, it is one of the few studies performed in Latin America involving all parameters together. PM2.5 samples were collected in six sampling sites with different emission sources (industrial, vehicular, soil dust) in Rio de Janeiro, Brazil. The PM2.5 daily concentrations ranged from 1 to 61 µg m-3, with averages higher than the annual limit (15 µg m-3) for some of the sites. The results of the statistics evaluation showed that PM2.5 concentrations were not influenced by seasonality. Furthermore, air basins defined previously were not confirmed, because some sites presented similar emission sources. Therefore, new redefinitions of air basins need to be done, once they are important to air quality management.

  3. Identification of Nanoparticle Prototypes and Archetypes.

    PubMed

    Fernandez, Michael; Barnard, Amanda S

    2015-12-22

    High-throughput (HT) computational characterization of nanomaterials is poised to accelerate novel material breakthroughs. The number of possible nanomaterials is increasing exponentially along with their complexity, and so statistical and information technology will play a fundamental role in rationalizing nanomaterials HT data. We demonstrate that multivariate statistical analysis of heterogeneous ensembles can identify the truly significant nanoparticles and their most relevant properties. Virtual samples of diamond nanoparticles and graphene nanoflakes are characterized using clustering and archetypal analysis, where we find that saturated particles are defined by their geometry, while nonsaturated nanoparticles are defined by their carbon chemistry. At the complex hull of the nanostructure spaces, a combination of complex archetypes can efficiency describe a large number of members of the ensembles, whereas the regular shapes that are typically assumed to be representative can only describe a small set of the most regular morphologies. This approach provides a route toward the characterization of computationally intractable virtual nanomaterial spaces, which can aid nanomaterials discovery in the foreseen big data scenario.

  4. Impact of Rating Scale Categories on Reliability and Fit Statistics of the Malay Spiritual Well-Being Scale using Rasch Analysis.

    PubMed

    Daher, Aqil Mohammad; Ahmad, Syed Hassan; Winn, Than; Selamat, Mohd Ikhsan

    2015-01-01

    Few studies have employed the item response theory in examining reliability. We conducted this study to examine the effect of Rating Scale Categories (RSCs) on the reliability and fit statistics of the Malay Spiritual Well-Being Scale, employing the Rasch model. The Malay Spiritual Well-Being Scale (SWBS) with the original six; three and four newly structured RSCs was distributed randomly among three different samples of 50 participants each. The mean age of respondents in the three samples ranged between 36 and 39 years old. The majority was female in all samples, and Islam was the most prevalent religion among the respondents. The predominating race was Malay, followed by Chinese and Indian. The original six RSCs indicated better targeting of 0.99 and smallest model error of 0.24. The Infit Mnsq (mean square) and Zstd (Z standard) of the six RSCs were "1.1"and "-0.1"respectively. The six RSCs achieved the highest person and item reliabilities of 0.86 and 0.85 respectively. These reliabilities yielded the highest person (2.46) and item (2.38) separation indices compared to other the RSCs. The person and item reliability and, to a lesser extent, the fit statistics, were better with the six RSCs compared to the four and three RSCs.

  5. Chemical Structure and Molecular Dimension As Controls on the Inherent Stability of Charcoal in Boreal Forest Soil

    NASA Astrophysics Data System (ADS)

    Hockaday, W. C.; Kane, E. S.; Ohlson, M.; Huang, R.; Von Bargen, J.; Davis, R.

    2014-12-01

    Efforts have been made by various scientific disciplines to study hyporheic zones and characterize their associated processes. One way to approach the study of the hyporheic zone is to define facies, which are elements of a (hydrobio) geologic classification scheme that groups components of a complex system with high variability into a manageable set of discrete classes. In this study, we try to classify the hyporheic zone based on the geology, geochemistry, microbiology, and understand their interactive influences on the integrated biogeochemical distributions and processes. A number of measurements have been taken for 21 freeze core samples along the Columbia River bank in the Hanford 300 Area, and unique datasets have been obtained on biomass, pH, number of microbial taxa, percentage of N/C/H/S, microbial activity parameters, as well as microbial community attributes/modules. In order to gain a complete understanding of the geological control on these variables and processes, the explanatory variables are set to include quantitative gravel/sand/mud/silt/clay percentages, statistical moments of grain size distributions, as well as geological (e.g., Folk-Wentworth) and statistical (e.g., hierarchical) clusters. The dominant factors for major microbial and geochemical variables are identified and summarized using exploratory data analysis approaches (e.g., principal component analysis, hierarchical clustering, factor analysis, multivariate analysis of variance). The feasibility of extending the facies definition and its control of microbial and geochemical properties to larger scales is discussed.

  6. Microbial facies distribution and its geological and geochemical controls at the Hanford 300 area

    NASA Astrophysics Data System (ADS)

    Hou, Z.; Nelson, W.; Stegen, J.; Murray, C. J.; Arntzen, E.

    2015-12-01

    Efforts have been made by various scientific disciplines to study hyporheic zones and characterize their associated processes. One way to approach the study of the hyporheic zone is to define facies, which are elements of a (hydrobio) geologic classification scheme that groups components of a complex system with high variability into a manageable set of discrete classes. In this study, we try to classify the hyporheic zone based on the geology, geochemistry, microbiology, and understand their interactive influences on the integrated biogeochemical distributions and processes. A number of measurements have been taken for 21 freeze core samples along the Columbia River bank in the Hanford 300 Area, and unique datasets have been obtained on biomass, pH, number of microbial taxa, percentage of N/C/H/S, microbial activity parameters, as well as microbial community attributes/modules. In order to gain a complete understanding of the geological control on these variables and processes, the explanatory variables are set to include quantitative gravel/sand/mud/silt/clay percentages, statistical moments of grain size distributions, as well as geological (e.g., Folk-Wentworth) and statistical (e.g., hierarchical) clusters. The dominant factors for major microbial and geochemical variables are identified and summarized using exploratory data analysis approaches (e.g., principal component analysis, hierarchical clustering, factor analysis, multivariate analysis of variance). The feasibility of extending the facies definition and its control of microbial and geochemical properties to larger scales is discussed.

  7. A Bayesian nonparametric method for prediction in EST analysis

    PubMed Central

    Lijoi, Antonio; Mena, Ramsés H; Prünster, Igor

    2007-01-01

    Background Expressed sequence tags (ESTs) analyses are a fundamental tool for gene identification in organisms. Given a preliminary EST sample from a certain library, several statistical prediction problems arise. In particular, it is of interest to estimate how many new genes can be detected in a future EST sample of given size and also to determine the gene discovery rate: these estimates represent the basis for deciding whether to proceed sequencing the library and, in case of a positive decision, a guideline for selecting the size of the new sample. Such information is also useful for establishing sequencing efficiency in experimental design and for measuring the degree of redundancy of an EST library. Results In this work we propose a Bayesian nonparametric approach for tackling statistical problems related to EST surveys. In particular, we provide estimates for: a) the coverage, defined as the proportion of unique genes in the library represented in the given sample of reads; b) the number of new unique genes to be observed in a future sample; c) the discovery rate of new genes as a function of the future sample size. The Bayesian nonparametric model we adopt conveys, in a statistically rigorous way, the available information into prediction. Our proposal has appealing properties over frequentist nonparametric methods, which become unstable when prediction is required for large future samples. EST libraries, previously studied with frequentist methods, are analyzed in detail. Conclusion The Bayesian nonparametric approach we undertake yields valuable tools for gene capture and prediction in EST libraries. The estimators we obtain do not feature the kind of drawbacks associated with frequentist estimators and are reliable for any size of the additional sample. PMID:17868445

  8. 23 CFR 1340.5 - Documentation requirements.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... STATE OBSERVATIONAL SURVEYS OF SEAT BELT USE § 1340.5 Documentation requirements. All sample design, data collection, and estimation procedures used in State surveys conducted in accordance with this part must be well documented. At a minimum, the documentation must: (a) For sample design— (1) Define all...

  9. 23 CFR 1340.5 - Documentation requirements.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... STATE OBSERVATIONAL SURVEYS OF SEAT BELT USE § 1340.5 Documentation requirements. All sample design, data collection, and estimation procedures used in State surveys conducted in accordance with this part must be well documented. At a minimum, the documentation must: (a) For sample design— (1) Define all...

  10. Extremely high genetic diversity in a single tumor points to prevalence of non-Darwinian cell evolution.

    PubMed

    Ling, Shaoping; Hu, Zheng; Yang, Zuyu; Yang, Fang; Li, Yawei; Lin, Pei; Chen, Ke; Dong, Lili; Cao, Lihua; Tao, Yong; Hao, Lingtong; Chen, Qingjian; Gong, Qiang; Wu, Dafei; Li, Wenjie; Zhao, Wenming; Tian, Xiuyun; Hao, Chunyi; Hungate, Eric A; Catenacci, Daniel V T; Hudson, Richard R; Li, Wen-Hsiung; Lu, Xuemei; Wu, Chung-I

    2015-11-24

    The prevailing view that the evolution of cells in a tumor is driven by Darwinian selection has never been rigorously tested. Because selection greatly affects the level of intratumor genetic diversity, it is important to assess whether intratumor evolution follows the Darwinian or the non-Darwinian mode of evolution. To provide the statistical power, many regions in a single tumor need to be sampled and analyzed much more extensively than has been attempted in previous intratumor studies. Here, from a hepatocellular carcinoma (HCC) tumor, we evaluated multiregional samples from the tumor, using either whole-exome sequencing (WES) (n = 23 samples) or genotyping (n = 286) under both the infinite-site and infinite-allele models of population genetics. In addition to the many single-nucleotide variations (SNVs) present in all samples, there were 35 "polymorphic" SNVs among samples. High genetic diversity was evident as the 23 WES samples defined 20 unique cell clones. With all 286 samples genotyped, clonal diversity agreed well with the non-Darwinian model with no evidence of positive Darwinian selection. Under the non-Darwinian model, MALL (the number of coding region mutations in the entire tumor) was estimated to be greater than 100 million in this tumor. DNA sequences reveal local diversities in small patches of cells and validate the estimation. In contrast, the genetic diversity under a Darwinian model would generally be orders of magnitude smaller. Because the level of genetic diversity will have implications on therapeutic resistance, non-Darwinian evolution should be heeded in cancer treatments even for microscopic tumors.

  11. Extremely high genetic diversity in a single tumor points to prevalence of non-Darwinian cell evolution

    PubMed Central

    Ling, Shaoping; Hu, Zheng; Yang, Zuyu; Yang, Fang; Li, Yawei; Lin, Pei; Chen, Ke; Dong, Lili; Cao, Lihua; Tao, Yong; Hao, Lingtong; Chen, Qingjian; Gong, Qiang; Wu, Dafei; Li, Wenjie; Zhao, Wenming; Tian, Xiuyun; Hao, Chunyi; Hungate, Eric A.; Catenacci, Daniel V. T.; Hudson, Richard R.; Li, Wen-Hsiung; Lu, Xuemei; Wu, Chung-I

    2015-01-01

    The prevailing view that the evolution of cells in a tumor is driven by Darwinian selection has never been rigorously tested. Because selection greatly affects the level of intratumor genetic diversity, it is important to assess whether intratumor evolution follows the Darwinian or the non-Darwinian mode of evolution. To provide the statistical power, many regions in a single tumor need to be sampled and analyzed much more extensively than has been attempted in previous intratumor studies. Here, from a hepatocellular carcinoma (HCC) tumor, we evaluated multiregional samples from the tumor, using either whole-exome sequencing (WES) (n = 23 samples) or genotyping (n = 286) under both the infinite-site and infinite-allele models of population genetics. In addition to the many single-nucleotide variations (SNVs) present in all samples, there were 35 “polymorphic” SNVs among samples. High genetic diversity was evident as the 23 WES samples defined 20 unique cell clones. With all 286 samples genotyped, clonal diversity agreed well with the non-Darwinian model with no evidence of positive Darwinian selection. Under the non-Darwinian model, MALL (the number of coding region mutations in the entire tumor) was estimated to be greater than 100 million in this tumor. DNA sequences reveal local diversities in small patches of cells and validate the estimation. In contrast, the genetic diversity under a Darwinian model would generally be orders of magnitude smaller. Because the level of genetic diversity will have implications on therapeutic resistance, non-Darwinian evolution should be heeded in cancer treatments even for microscopic tumors. PMID:26561581

  12. Spin-polarized scanning tunneling microscopy experiments on the rough surface of a polycrystalline NiFe film with a fine magnetic tip sensitive to a well-defined magnetization component

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Matsuyama, H., E-mail: matsu@phys.sci.hokudai.ac.jp; Nara, D.; Kageyama, R.

    We developed a micrometer-sized magnetic tip integrated onto the write head of a hard disk drive for spin-polarized scanning tunneling microscopy (SP-STM) in the modulated tip magnetization mode. Using SP-STM, we measured a well-defined in-plane spin-component of the tunneling current of the rough surface of a polycrystalline NiFe film. The spin asymmetry of the NiFe film was about 1.3% within the bias voltage range of -3 to 1 V. We obtained the local spin component image of the sample surface, switching the magnetic field of the sample to reverse the sample magnetization during scanning. We also obtained a spin imagemore » of the rough surface of a polycrystalline NiFe film evaporated on the recording medium of a hard disk drive.« less

  13. Definition of sampling units begets conclusions in ecology: the case of habitats for plant communities.

    PubMed

    Mörsdorf, Martin A; Ravolainen, Virve T; Støvern, Leif Einar; Yoccoz, Nigel G; Jónsdóttir, Ingibjörg Svala; Bråthen, Kari Anne

    2015-01-01

    In ecology, expert knowledge on habitat characteristics is often used to define sampling units such as study sites. Ecologists are especially prone to such approaches when prior sampling frames are not accessible. Here we ask to what extent can different approaches to the definition of sampling units influence the conclusions that are drawn from an ecological study? We do this by comparing a formal versus a subjective definition of sampling units within a study design which is based on well-articulated objectives and proper methodology. Both approaches are applied to tundra plant communities in mesic and snowbed habitats. For the formal approach, sampling units were first defined for each habitat in concave terrain of suitable slope using GIS. In the field, these units were only accepted as the targeted habitats if additional criteria for vegetation cover were fulfilled. For the subjective approach, sampling units were defined visually in the field, based on typical plant communities of mesic and snowbed habitats. For each approach, we collected information about plant community characteristics within a total of 11 mesic and seven snowbed units distributed between two herding districts of contrasting reindeer density. Results from the two approaches differed significantly in several plant community characteristics in both mesic and snowbed habitats. Furthermore, differences between the two approaches were not consistent because their magnitude and direction differed both between the two habitats and the two reindeer herding districts. Consequently, we could draw different conclusions on how plant diversity and relative abundance of functional groups are differentiated between the two habitats depending on the approach used. We therefore challenge ecologists to formalize the expert knowledge applied to define sampling units through a set of well-articulated rules, rather than applying it subjectively. We see this as instrumental for progress in ecology as only rules based on expert knowledge are transparent and lead to results reproducible by other ecologists.

  14. High risk human papillomavirus in the periodontium : A case control study.

    PubMed

    Shipilova, Anna; Dayakar, Manjunath Mundoor; Gupta, Dinesh

    2017-01-01

    Human papilloma viruses (HPVs) are small DNA viruses that have been identified in periodontal pocket as well as gingival sulcus. High risk HPVs are also associated with a subset of head and neck carcinomas. It is thought that the periodontium could be a reservoir for HPV. 1. Detection of Human Papilloma virus (HPV) in periodontal pocket as well as gingival of patients having localized chronic periodontitis and gingival sulcus of periodontally healthy subjects. 2. Quantitative estimation of E6 and E7 mRNA in subjects showing presence of HPV3. To assess whether periodontal pocket is a reservoir for HPV. This case-control study included 30 subjects with localized chronic Periodontitis (cases) and 30 periodontally healthy subjects (controls). Two samples were taken from cases, one from periodontal pocket and one from gingival sulcus and one sample was taken from controls. Samples were collected in the form of pocket scrapings and gingival sulcus scrapings from cases and controls respectively. These samples were sent in storage media for identification and estimation of E6/E7 mRNA of HPV using in situ hybridization and flow cytometry. Statistical analysis was done by using, mean, percentage and Chi Square test. A statistical package SPSS version 13.0 was used to analyze the data. P value < 0.05 was considered as statistically significant. pocket samples as well as sulcus samples for both cases and controls were found to contain HPV E6/E7 mRNAInterpretation and. Presence of HPV E6/E7 mRNA in periodontium supports the hypothesis that periodontal tissues serve as a reservoir for latent HPV and there may be a synergy between oral cancer, periodontitis and HPV. However prospective studies are required to further explore this link.

  15. Chemical Constituents in Groundwater from Multiple Zones in the Eastern Snake River Plain Aquifer at the Idaho National Laboratory, Idaho, 2005-08

    USGS Publications Warehouse

    Bartholomay, Roy C.; Twining, Brian V.

    2010-01-01

    From 2005 to 2008, the U.S. Geological Survey's Idaho National Laboratory (INL) Project office, in cooperation with the U.S. Department of Energy, collected water-quality samples from multiple water-bearing zones in the eastern Snake River Plain aquifer. Water samples were collected from six monitoring wells completed in about 350-700 feet of the upper part of the aquifer, and the samples were analyzed for major ions, selected trace elements, nutrients, selected radiochemical constituents, and selected stable isotopes. Each well was equipped with a multilevel monitoring system containing four to seven sampling ports that were each isolated by permanent packer systems. The sampling ports were installed in aquifer zones that were highly transmissive and that represented the water chemistry of the top four to five model layers of a steady-state and transient groundwater-flow model. The model's water chemistry and particle-tracking simulations are being used to better define movement of wastewater constituents in the aquifer. The results of the water chemistry analyses indicated that, in each of four separate wells, one zone of water differed markedly from the other zones in the well. In four wells, one zone to as many as five zones contained radiochemical constituents that originated from wastewater disposal at selected laboratory facilities. The multilevel sampling systems are defining the vertical distribution of wastewater constituents in the eastern Snake River Plain aquifer and the concentrations of wastewater constituents in deeper zones in wells Middle 2051, USGS 132, and USGS 103 support the concept of groundwater flow deepening in the southwestern part of the INL.

  16. Modeling motor vehicle crashes using Poisson-gamma models: examining the effects of low sample mean values and small sample size on the estimation of the fixed dispersion parameter.

    PubMed

    Lord, Dominique

    2006-07-01

    There has been considerable research conducted on the development of statistical models for predicting crashes on highway facilities. Despite numerous advancements made for improving the estimation tools of statistical models, the most common probabilistic structure used for modeling motor vehicle crashes remains the traditional Poisson and Poisson-gamma (or Negative Binomial) distribution; when crash data exhibit over-dispersion, the Poisson-gamma model is usually the model of choice most favored by transportation safety modelers. Crash data collected for safety studies often have the unusual attributes of being characterized by low sample mean values. Studies have shown that the goodness-of-fit of statistical models produced from such datasets can be significantly affected. This issue has been defined as the "low mean problem" (LMP). Despite recent developments on methods to circumvent the LMP and test the goodness-of-fit of models developed using such datasets, no work has so far examined how the LMP affects the fixed dispersion parameter of Poisson-gamma models used for modeling motor vehicle crashes. The dispersion parameter plays an important role in many types of safety studies and should, therefore, be reliably estimated. The primary objective of this research project was to verify whether the LMP affects the estimation of the dispersion parameter and, if it is, to determine the magnitude of the problem. The secondary objective consisted of determining the effects of an unreliably estimated dispersion parameter on common analyses performed in highway safety studies. To accomplish the objectives of the study, a series of Poisson-gamma distributions were simulated using different values describing the mean, the dispersion parameter, and the sample size. Three estimators commonly used by transportation safety modelers for estimating the dispersion parameter of Poisson-gamma models were evaluated: the method of moments, the weighted regression, and the maximum likelihood method. In an attempt to complement the outcome of the simulation study, Poisson-gamma models were fitted to crash data collected in Toronto, Ont. characterized by a low sample mean and small sample size. The study shows that a low sample mean combined with a small sample size can seriously affect the estimation of the dispersion parameter, no matter which estimator is used within the estimation process. The probability the dispersion parameter becomes unreliably estimated increases significantly as the sample mean and sample size decrease. Consequently, the results show that an unreliably estimated dispersion parameter can significantly undermine empirical Bayes (EB) estimates as well as the estimation of confidence intervals for the gamma mean and predicted response. The paper ends with recommendations about minimizing the likelihood of producing Poisson-gamma models with an unreliable dispersion parameter for modeling motor vehicle crashes.

  17. STATISTICAL ANALYSIS OF TANK 18F FLOOR SAMPLE RESULTS

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Harris, S.

    2010-09-02

    Representative sampling has been completed for characterization of the residual material on the floor of Tank 18F as per the statistical sampling plan developed by Shine [1]. Samples from eight locations have been obtained from the tank floor and two of the samples were archived as a contingency. Six samples, referred to in this report as the current scrape samples, have been submitted to and analyzed by SRNL [2]. This report contains the statistical analysis of the floor sample analytical results to determine if further data are needed to reduce uncertainty. Included are comparisons with the prior Mantis samples resultsmore » [3] to determine if they can be pooled with the current scrape samples to estimate the upper 95% confidence limits (UCL{sub 95%}) for concentration. Statistical analysis revealed that the Mantis and current scrape sample results are not compatible. Therefore, the Mantis sample results were not used to support the quantification of analytes in the residual material. Significant spatial variability among the current sample results was not found. Constituent concentrations were similar between the North and South hemispheres as well as between the inner and outer regions of the tank floor. The current scrape sample results from all six samples fall within their 3-sigma limits. In view of the results from numerous statistical tests, the data were pooled from all six current scrape samples. As such, an adequate sample size was provided for quantification of the residual material on the floor of Tank 18F. The uncertainty is quantified in this report by an upper 95% confidence limit (UCL{sub 95%}) on each analyte concentration. The uncertainty in analyte concentration was calculated as a function of the number of samples, the average, and the standard deviation of the analytical results. The UCL{sub 95%} was based entirely on the six current scrape sample results (each averaged across three analytical determinations).« less

  18. STATISTICAL ANALYSIS OF TANK 19F FLOOR SAMPLE RESULTS

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Harris, S.

    2010-09-02

    Representative sampling has been completed for characterization of the residual material on the floor of Tank 19F as per the statistical sampling plan developed by Harris and Shine. Samples from eight locations have been obtained from the tank floor and two of the samples were archived as a contingency. Six samples, referred to in this report as the current scrape samples, have been submitted to and analyzed by SRNL. This report contains the statistical analysis of the floor sample analytical results to determine if further data are needed to reduce uncertainty. Included are comparisons with the prior Mantis samples resultsmore » to determine if they can be pooled with the current scrape samples to estimate the upper 95% confidence limits (UCL95%) for concentration. Statistical analysis revealed that the Mantis and current scrape sample results are not compatible. Therefore, the Mantis sample results were not used to support the quantification of analytes in the residual material. Significant spatial variability among the current scrape sample results was not found. Constituent concentrations were similar between the North and South hemispheres as well as between the inner and outer regions of the tank floor. The current scrape sample results from all six samples fall within their 3-sigma limits. In view of the results from numerous statistical tests, the data were pooled from all six current scrape samples. As such, an adequate sample size was provided for quantification of the residual material on the floor of Tank 19F. The uncertainty is quantified in this report by an UCL95% on each analyte concentration. The uncertainty in analyte concentration was calculated as a function of the number of samples, the average, and the standard deviation of the analytical results. The UCL95% was based entirely on the six current scrape sample results (each averaged across three analytical determinations).« less

  19. Satellite Sampling and Retrieval Errors in Regional Monthly Rain Estimates from TMI AMSR-E, SSM/I, AMSU-B and the TRMM PR

    NASA Technical Reports Server (NTRS)

    Fisher, Brad; Wolff, David B.

    2010-01-01

    Passive and active microwave rain sensors onboard earth-orbiting satellites estimate monthly rainfall from the instantaneous rain statistics collected during satellite overpasses. It is well known that climate-scale rain estimates from meteorological satellites incur sampling errors resulting from the process of discrete temporal sampling and statistical averaging. Sampling and retrieval errors ultimately become entangled in the estimation of the mean monthly rain rate. The sampling component of the error budget effectively introduces statistical noise into climate-scale rain estimates that obscure the error component associated with the instantaneous rain retrieval. Estimating the accuracy of the retrievals on monthly scales therefore necessitates a decomposition of the total error budget into sampling and retrieval error quantities. This paper presents results from a statistical evaluation of the sampling and retrieval errors for five different space-borne rain sensors on board nine orbiting satellites. Using an error decomposition methodology developed by one of the authors, sampling and retrieval errors were estimated at 0.25 resolution within 150 km of ground-based weather radars located at Kwajalein, Marshall Islands and Melbourne, Florida. Error and bias statistics were calculated according to the land, ocean and coast classifications of the surface terrain mask developed for the Goddard Profiling (GPROF) rain algorithm. Variations in the comparative error statistics are attributed to various factors related to differences in the swath geometry of each rain sensor, the orbital and instrument characteristics of the satellite and the regional climatology. The most significant result from this study found that each of the satellites incurred negative longterm oceanic retrieval biases of 10 to 30%.

  20. Visual classification of very fine-grained sediments: Evaluation through univariate and multivariate statistics

    USGS Publications Warehouse

    Hohn, M. Ed; Nuhfer, E.B.; Vinopal, R.J.; Klanderman, D.S.

    1980-01-01

    Classifying very fine-grained rocks through fabric elements provides information about depositional environments, but is subject to the biases of visual taxonomy. To evaluate the statistical significance of an empirical classification of very fine-grained rocks, samples from Devonian shales in four cored wells in West Virginia and Virginia were measured for 15 variables: quartz, illite, pyrite and expandable clays determined by X-ray diffraction; total sulfur, organic content, inorganic carbon, matrix density, bulk density, porosity, silt, as well as density, sonic travel time, resistivity, and ??-ray response measured from well logs. The four lithologic types comprised: (1) sharply banded shale, (2) thinly laminated shale, (3) lenticularly laminated shale, and (4) nonbanded shale. Univariate and multivariate analyses of variance showed that the lithologic classification reflects significant differences for the variables measured, difference that can be detected independently of stratigraphic effects. Little-known statistical methods found useful in this work included: the multivariate analysis of variance with more than one effect, simultaneous plotting of samples and variables on canonical variates, and the use of parametric ANOVA and MANOVA on ranked data. ?? 1980 Plenum Publishing Corporation.

  1. Extreme value statistics and finite-size scaling at the ecological extinction/laminar-turbulence transition

    NASA Astrophysics Data System (ADS)

    Shih, Hong-Yan; Goldenfeld, Nigel

    Experiments on transitional turbulence in pipe flow seem to show that turbulence is a transient metastable state since the measured mean lifetime of turbulence puffs does not diverge asymptotically at a critical Reynolds number. Yet measurements reveal that the lifetime scales with Reynolds number in a super-exponential way reminiscent of extreme value statistics, and simulations and experiments in Couette and channel flow exhibit directed percolation type scaling phenomena near a well-defined transition. This universality class arises from the interplay between small-scale turbulence and a large-scale collective zonal flow, which exhibit predator-prey behavior. Why is asymptotically divergent behavior not observed? Using directed percolation and a stochastic individual level model of predator-prey dynamics related to transitional turbulence, we investigate the relation between extreme value statistics and power law critical behavior, and show that the paradox is resolved by carefully defining what is measured in the experiments. We theoretically derive the super-exponential scaling law, and using finite-size scaling, show how the same data can give both super-exponential behavior and power-law critical scaling.

  2. Probability of Unmixed Young Groundwater (defined using chlorofluorocarbon-11 concentrations and tritium activities) in the Eagle River Watershed Valley-Fill Aquifer, Eagle County, North-Central Colorado, 2006-2007

    USGS Publications Warehouse

    Rupert, Michael G.; Plummer, Niel

    2009-01-01

    This raster data set delineates the predicted probability of unmixed young groundwater (defined using chlorofluorocarbon-11 concentrations and tritium activities) in groundwater in the Eagle River watershed valley-fill aquifer, Eagle County, North-Central Colorado, 2006-2007. This data set was developed by a cooperative project between the U.S. Geological Survey, Eagle County, the Eagle River Water and Sanitation District, the Town of Eagle, the Town of Gypsum, and the Upper Eagle Regional Water Authority. This project was designed to evaluate potential land-development effects on groundwater and surface-water resources so that informed land-use and water management decisions can be made. This groundwater probability map and its associated probability maps were developed as follows: (1) A point data set of wells with groundwater quality and groundwater age data was overlaid with thematic layers of anthropogenic (related to human activities) and hydrogeologic data by using a geographic information system to assign each well values for depth to groundwater, distance to major streams and canals, distance to gypsum beds, precipitation, soils, and well depth. These data then were downloaded to a statistical software package for analysis by logistic regression. (2) Statistical models predicting the probability of elevated nitrate concentrations, the probability of unmixed young water (using chlorofluorocarbon-11 concentrations and tritium activities), and the probability of elevated volatile organic compound concentrations were developed using logistic regression techniques. (3) The statistical models were entered into a GIS and the probability map was constructed.

  3. Efficient bootstrap estimates for tail statistics

    NASA Astrophysics Data System (ADS)

    Breivik, Øyvind; Aarnes, Ole Johan

    2017-03-01

    Bootstrap resamples can be used to investigate the tail of empirical distributions as well as return value estimates from the extremal behaviour of the sample. Specifically, the confidence intervals on return value estimates or bounds on in-sample tail statistics can be obtained using bootstrap techniques. However, non-parametric bootstrapping from the entire sample is expensive. It is shown here that it suffices to bootstrap from a small subset consisting of the highest entries in the sequence to make estimates that are essentially identical to bootstraps from the entire sample. Similarly, bootstrap estimates of confidence intervals of threshold return estimates are found to be well approximated by using a subset consisting of the highest entries. This has practical consequences in fields such as meteorology, oceanography and hydrology where return values are calculated from very large gridded model integrations spanning decades at high temporal resolution or from large ensembles of independent and identically distributed model fields. In such cases the computational savings are substantial.

  4. Y-chromosomal diversity of the Valachs from the Czech Republic: model for isolated population in Central Europe

    PubMed Central

    Ehler, Edvard; Vaněk, Daniel; Stenzl, Vlastimil; Vančata, Václav

    2011-01-01

    Aim To evaluate Y-chromosomal diversity of the Moravian Valachs of the Czech Republic and compare them with a Czech population sample and other samples from Central and South-Eastern Europe, and to evaluate the effects of genetic isolation and sampling. Methods The first sample set of the Valachs consisted of 94 unrelated male donors from the Valach region in northeastern Czech Republic border-area. The second sample set of the Valachs consisted of 79 men who originated from 7 paternal lineages defined by surname. No close relatives were sampled. The third sample set consisted of 273 unrelated men from the whole of the Czech Republic and was used for comparison, as well as published data for other 27 populations. The total number of samples was 3244. Y-short tandem repeat (STR) markers were typed by standard methods using PowerPlex® Y System (Promega) and Yfiler® Amplification Kit (Applied Biosystems) kits. Y-chromosomal haplogroups were estimated from the haplotype information. Haplotype diversity and other intra- and inter-population statistics were computed. Results The Moravian Valachs showed a lower genetic variability of Y-STR markers than other Central European populations, resembling more to the isolated Balkan populations (Aromuns, Csango, Bulgarian, and Macedonian Roma) than the surrounding populations (Czechs, Slovaks, Poles, Saxons). We illustrated the effect of sampling on Valach paternal lineages, which includes reduction of discrimination capacity and variability inside Y-chromosomal haplogroups. Valach modal haplotype belongs to R1a haplogroup and it was not detected in the Czech population. Conclusion The Moravian Valachs display strong substructure and isolation in their Y chromosomal markers. They represent a unique Central European population model for population genetics. PMID:21674832

  5. Status of groundwater quality in the California Desert Region, 2006-2008: California GAMA Priority Basin Project

    USGS Publications Warehouse

    Dawson, Barbara J. Milby; Belitz, Kenneth

    2012-01-01

    Groundwater quality in six areas in the California Desert Region (Owens, Antelope, Mojave, Coachella, Colorado River, and Indian Wells) was investigated as part of the Priority Basin Project of the Groundwater Ambient Monitoring and Assessment (GAMA) Program. The GAMA Priority Basin Project is being conducted by the California State Water Resources Control Board in collaboration with the U.S. Geological Survey (USGS) and the Lawrence Livermore National Laboratory. The six Desert studies were designed to provide a spatially unbiased assessment of the quality of untreated groundwater in parts of the Desert and the Basin and Range hydrogeologic provinces, as well as a statistically consistent basis for comparing groundwater quality to other areas in California and across the Nation. Samples were collected by the USGS from September 2006 through April 2008 from 253 wells in Imperial, Inyo, Kern, Los Angeles, Mono, Riverside, and San Bernardino Counties. Two-hundred wells were selected using a spatially distributed, randomized grid-based method to provide a spatially unbiased representation of the study areas (grid wells), and fifty-three wells were sampled to provide additional insight into groundwater conditions (additional wells). The status of the current quality of the groundwater resource was assessed based on data from samples analyzed for volatile organic compounds (VOCs), pesticides, and inorganic constituents such as major ions and trace elements. Water-quality data from the California Department of Public Health (CDPH) database also were incorporated in the assessment. The status assessment is intended to characterize the quality of untreated groundwater resources within the primary aquifer systems of the Desert Region, not the treated drinking water delivered to consumers by water purveyors. The primary aquifer systems (hereinafter, primary aquifers) in the six Desert areas are defined as that part of the aquifer corresponding to the perforation intervals of wells listed in the CDPH database. Relative-concentrations (sample concentration divided by the benchmark concentration) were used as the primary metric for evaluating groundwater quality for those constituents that have Federal and (or) California benchmarks. A relative-concentration (RC) greater than (>) 1.0 indicates a concentration above a benchmark, and an RC less than or equal to (≤) 1.0 indicates a concentration equal to or below a benchmark. Organic and special-interest constituent RCs were classified as “low” (RC ≤ 0.1), “moderate” (0.1 1.0). Inorganic constituent RCs were classified as “low” (RC ≤ 0.5), “moderate” (0.5 1.0). A lower threshold value RC was used to distinguish between low and moderate RCs for organic constituents because these constituents are generally less prevalent and have smaller RCs than inorganic constituents. Aquifer-scale proportion was used as the primary metric for evaluating regional-scale groundwater quality. High aquifer-scale proportion was defined as the percentage of the area of the primary aquifers with an RC greater than 1.0 for a particular constituent or class of constituents; percentage is based on an areal rather than a volumetric basis. Moderate and low aquifer-scale proportions were defined as the percentage of the primary aquifers with moderate and low RCs, respectively. Two statistical approaches—grid-based and spatially weighted—were used to evaluate aquifer-scale proportions for individual constituents and classes of constituents. Grid-based and spatially weighted estimates were comparable in the Desert Region (within 90 percent confidence intervals). The status assessment determined that one or more inorganic constituents with health-based benchmarks had high RCs in 35.4 percent of the Desert Region’s primary aquifers, moderate RCs in 27.4 percent, and low RCs in 37.2 percent. The inorganic constituents with health-based benchmarks having the largest high aquifer-scale proportions were arsenic (17.8 percent), boron (11.4 percent), fluoride (8.9 percent), gross-alpha radioactivity (6.6 percent), molybdenum (5.7 percent), strontium (3.7 percent), vanadium (3.6 percent), uranium (3.2 percent), and perchlorate (2.4 percent). Inorganic constituents with non-health-based benchmarks were also detected at high RCs in 18.6 percent and at moderate RCs in 16.0 percent of the Desert Region’s primary aquifers. In contrast, organic constituents had high RCs in only 0.3 percent of the Desert Region’s primary aquifers, moderate in 2.0 percent, low in 48.0 percent, and were not detected in 49.7 percent of the primary aquifers in the Desert Region. Of 149 organic constituents analyzed for all six study areas, 42 constituents were detected. Six organic constituents, carbon tetrachloride, chloroform, 1,2-dichloropropane, dieldrin, 1,2-dichloroethane, and tetrachloroethene, were found at moderate RCs in one or more of the grid wells. One constituent, N-nitrosodimethylamine, a special-interest VOC, was detected at a high RC in one well. Thirty-nine organic constituents were detected only at low concentrations. Three organic constituents were frequently detected (in more than 10 percent of samples from grid wells): chloroform, simazine, and deethylatrazine.

  6. Association between Periodontal Condition and Nutritional Status of Brazilian Adolescents: A Population-based Study.

    PubMed

    Cavalcanti, Alessandro L; Ramos, Ianny A; Cardoso, Andreia M R; Fernandes, Liege Helena F; Aragão, Amanda S; Santos, Fábio G; Aguiar, Yêska P C; Carvalho, Danielle F; Medeiros, Carla C M; De S C Soares, Renata; Castro, Ricardo D

    2016-12-01

    Obesity is a serious problem of public health and affects all socio-economic groups, irrespective of age, sex or ethnicity. The aim of this study was to evaluate the association between periodontal condition and nutritional status of adolescents. This was a cross-sectional study using a probability cluster sampling, and the sample was defined by statistical criterion, consisting of 559 students aged 15-19 yr enrolled in public schools of adolescents of Campina Grande, PB, Brazil in 2012. Socioeconomic characteristics were analyzed, as well as self-reported general and oral health, anthropometric data and periodontal condition (CPI and OHI-S). Descriptive and analytical analysis from bivariate and multivariate Poisson regression analysis with 5% significance level was performed. Of the 559 adolescents, 18.6% were overweight and 98.4% had some form of periodontal changes such as: bleeding (34.3%), calculus (38.8%), shallow pocket (22.9%) and deep pocket (2.3%). There was association between presence of periodontal changes with obesity ( P <0.05; CI 95%: 0.99 [0.98 - 0.99]). The association between presence of periodontal changes and obesity status in adolescents was indicated.

  7. Association between Periodontal Condition and Nutritional Status of Brazilian Adolescents: A Population-based Study

    PubMed Central

    CAVALCANTI, Alessandro L.; RAMOS, Ianny A.; CARDOSO, Andreia M. R.; FERNANDES, Liege Helena F.; ARAGÃO, Amanda S.; SANTOS, Fábio G.; AGUIAR, Yêska P. C.; CARVALHO, Danielle F.; MEDEIROS, Carla C. M.; De S. C. SOARES, Renata; CASTRO, Ricardo D.

    2016-01-01

    Background: Obesity is a serious problem of public health and affects all socio-economic groups, irrespective of age, sex or ethnicity. The aim of this study was to evaluate the association between periodontal condition and nutritional status of adolescents. Methods: This was a cross-sectional study using a probability cluster sampling, and the sample was defined by statistical criterion, consisting of 559 students aged 15–19 yr enrolled in public schools of adolescents of Campina Grande, PB, Brazil in 2012. Socioeconomic characteristics were analyzed, as well as self-reported general and oral health, anthropometric data and periodontal condition (CPI and OHI-S). Descriptive and analytical analysis from bivariate and multivariate Poisson regression analysis with 5% significance level was performed. Results: Of the 559 adolescents, 18.6% were overweight and 98.4% had some form of periodontal changes such as: bleeding (34.3%), calculus (38.8%), shallow pocket (22.9%) and deep pocket (2.3%). There was association between presence of periodontal changes with obesity (P<0.05; CI 95%: 0.99 [0.98 – 0.99]). Conclusion: The association between presence of periodontal changes and obesity status in adolescents was indicated. PMID:28053924

  8. Dynamic laser speckle analyzed considering inhomogeneities in the biological sample

    NASA Astrophysics Data System (ADS)

    Braga, Roberto A.; González-Peña, Rolando J.; Viana, Dimitri Campos; Rivera, Fernando Pujaico

    2017-04-01

    Dynamic laser speckle phenomenon allows a contactless and nondestructive way to monitor biological changes that are quantified by second-order statistics applied in the images in time using a secondary matrix known as time history of the speckle pattern (THSP). To avoid being time consuming, the traditional way to build the THSP restricts the data to a line or column. Our hypothesis is that the spatial restriction of the information could compromise the results, particularly when undesirable and unexpected optical inhomogeneities occur, such as in cell culture media. It tested a spatial random approach to collect the points to form a THSP. Cells in a culture medium and in drying paint, representing homogeneous samples in different levels, were tested, and a comparison with the traditional method was carried out. An alternative random selection based on a Gaussian distribution around a desired position was also presented. The results showed that the traditional protocol presented higher variation than the outcomes using the random method. The higher the inhomogeneity of the activity map, the higher the efficiency of the proposed method using random points. The Gaussian distribution proved to be useful when there was a well-defined area to monitor.

  9. Broad supernatural punishment but not moralizing high gods precede the evolution of political complexity in Austronesia

    PubMed Central

    Watts, Joseph; Greenhill, Simon J.; Atkinson, Quentin D.; Currie, Thomas E.; Bulbulia, Joseph; Gray, Russell D.

    2015-01-01

    Supernatural belief presents an explanatory challenge to evolutionary theorists—it is both costly and prevalent. One influential functional explanation claims that the imagined threat of supernatural punishment can suppress selfishness and enhance cooperation. Specifically, morally concerned supreme deities or ‘moralizing high gods' have been argued to reduce free-riding in large social groups, enabling believers to build the kind of complex societies that define modern humanity. Previous cross-cultural studies claiming to support the MHG hypothesis rely on correlational analyses only and do not correct for the statistical non-independence of sampled cultures. Here we use a Bayesian phylogenetic approach with a sample of 96 Austronesian cultures to test the MHG hypothesis as well as an alternative supernatural punishment hypothesis that allows punishment by a broad range of moralizing agents. We find evidence that broad supernatural punishment drives political complexity, whereas MHGs follow political complexity. We suggest that the concept of MHGs diffused as part of a suite of traits arising from cultural exchange between complex societies. Our results show the power of phylogenetic methods to address long-standing debates about the origins and functions of religion in human society. PMID:25740888

  10. Probing the tides in interacting galaxy pairs

    NASA Technical Reports Server (NTRS)

    Borne, Kirk D.

    1990-01-01

    Detailed spectroscopic and imaging observations of colliding elliptical galaxies revealed unmistakable diagnostic signatures of the tidal interactions. It is possible to compare both the distorted luminosity distributions and the disturbed internal rotation profiles with numerical simulations in order to model the strength of the tidal gravitational field acting within a given pair of galaxies. Using the best-fit numerical model, one can then measure directly the mass of a specific interacting binary system. This technique applies to individual pairs and therefore complements the classical methods of measuring the masses of galaxy pairs in well-defined statistical samples. The 'personalized' modeling of galaxy pairs also permits the derivation of each binary's orbit, spatial orientation, and interaction timescale. Similarly, one can probe the tides in less-detailed observations of disturbed galaxies in order to estimate some of the physical parameters for larger samples of interacting galaxy pairs. These parameters are useful inputs to the more universal problems of (1) the galaxy merger rate, (2) the strength and duration of the driving forces behind tidally stimulated phenomena (e.g., starbursts and maybe quasi steller objects), and (3) the identification of long-lived signatures of interaction/merger events.

  11. Pump-Flow-Probe X-Ray Absorption Spectroscopy as a Tool for Studying Intermediate States of Photocatalytic Systems.

    PubMed

    Smolentsev, Grigory; Guda, Alexander; Zhang, Xiaoyi; Haldrup, Kristoffer; Andreiadis, Eugen; Chavarot-Kerlidou, Murielle; Canton, Sophie E; Nachtegaal, Maarten; Artero, Vincent; Sundstrom, Villy

    2013-08-29

    A new setup for pump-flow-probe X-ray absorption spectroscopy has been implemented at the SuperXAS beamline of the Swiss Light Source. It allows recording X-ray absorption spectra with a time resolution of tens of microseconds and high detection efficiency for samples with sub-mM concentrations. A continuous wave laser is used for the photoexcitation, with the distance between laser and X-ray beams and velocity of liquid flow determining the time delay, while the focusing of both beams and the flow speed define the time resolution. This method is compared with the alternative measurement technique that utilizes a 1 kHz repetition rate laser and multiple X-ray probe pulses. Such an experiment was performed at beamline 11ID-D of the Advanced Photon Source. Advantages, limitations and potential for improvement of the pump-flow-probe setup are discussed by analyzing the photon statistics. Both methods, with Co K-edge probing were applied to the investigation of a cobaloxime-based photo-catalytic reaction. The interplay between optimizing for efficient photoexcitation and time resolution as well as the effect of sample degradation for these two setups are discussed.

  12. Pump-Flow-Probe X-Ray Absorption Spectroscopy as a Tool for Studying Intermediate States of Photocatalytic Systems

    PubMed Central

    Smolentsev, Grigory; Guda, Alexander; Zhang, XIaoyi; Haldrup, Kristoffer; Andreiadis, Eugen; Chavarot-Kerlidou, Murielle; Canton, Sophie E.; Nachtegaal, Maarten; Artero, Vincent; Sundstrom, Villy

    2014-01-01

    A new setup for pump-flow-probe X-ray absorption spectroscopy has been implemented at the SuperXAS beamline of the Swiss Light Source. It allows recording X-ray absorption spectra with a time resolution of tens of microseconds and high detection efficiency for samples with sub-mM concentrations. A continuous wave laser is used for the photoexcitation, with the distance between laser and X-ray beams and velocity of liquid flow determining the time delay, while the focusing of both beams and the flow speed define the time resolution. This method is compared with the alternative measurement technique that utilizes a 1 kHz repetition rate laser and multiple X-ray probe pulses. Such an experiment was performed at beamline 11ID-D of the Advanced Photon Source. Advantages, limitations and potential for improvement of the pump-flow-probe setup are discussed by analyzing the photon statistics. Both methods, with Co K-edge probing were applied to the investigation of a cobaloxime-based photo-catalytic reaction. The interplay between optimizing for efficient photoexcitation and time resolution as well as the effect of sample degradation for these two setups are discussed. PMID:24443663

  13. Tree-space statistics and approximations for large-scale analysis of anatomical trees.

    PubMed

    Feragen, Aasa; Owen, Megan; Petersen, Jens; Wille, Mathilde M W; Thomsen, Laura H; Dirksen, Asger; de Bruijne, Marleen

    2013-01-01

    Statistical analysis of anatomical trees is hard to perform due to differences in the topological structure of the trees. In this paper we define statistical properties of leaf-labeled anatomical trees with geometric edge attributes by considering the anatomical trees as points in the geometric space of leaf-labeled trees. This tree-space is a geodesic metric space where any two trees are connected by a unique shortest path, which corresponds to a tree deformation. However, tree-space is not a manifold, and the usual strategy of performing statistical analysis in a tangent space and projecting onto tree-space is not available. Using tree-space and its shortest paths, a variety of statistical properties, such as mean, principal component, hypothesis testing and linear discriminant analysis can be defined. For some of these properties it is still an open problem how to compute them; others (like the mean) can be computed, but efficient alternatives are helpful in speeding up algorithms that use means iteratively, like hypothesis testing. In this paper, we take advantage of a very large dataset (N = 8016) to obtain computable approximations, under the assumption that the data trees parametrize the relevant parts of tree-space well. Using the developed approximate statistics, we illustrate how the structure and geometry of airway trees vary across a population and show that airway trees with Chronic Obstructive Pulmonary Disease come from a different distribution in tree-space than healthy ones. Software is available from http://image.diku.dk/aasa/software.php.

  14. A finer view of the conditional galaxy luminosity function and magnitude-gap statistics

    NASA Astrophysics Data System (ADS)

    Trevisan, M.; Mamon, G. A.

    2017-10-01

    The gap between first- and second-ranked galaxy magnitudes in groups is often considered a tracer of their merger histories, which in turn may affect galaxy properties, and also serves to test galaxy luminosity functions (LFs). We remeasure the conditional luminosity function (CLF) of the Main Galaxy Sample of the SDSS in an appropriately cleaned subsample of groups from the Yang catalogue. We find that, at low group masses, our best-fitting CLF has steeper satellite high ends, yet higher ratios of characteristic satellite to central luminosities in comparison with the CLF of Yang et al. The observed fractions of groups with large and small magnitude gaps as well as the Tremaine & Richstone statistics are not compatible with either a single Schechter LF or with a Schechter-like satellite plus lognormal central LF. These gap statistics, which naturally depend on the size of the subsamples, and also on the maximum projected radius, Rmax, for defining the second brightest galaxy, can only be reproduced with two-component CLFs if we allow small gap groups to preferentially have two central galaxies, as expected when groups merge. Finally, we find that the trend of higher gap for higher group velocity dispersion, σv, at a given richness, discovered by Hearin et al., is strongly reduced when we consider σv in bins of richness, and virtually disappears when we use group mass instead of σv. This limits the applicability of gaps in refining cosmographic studies based on cluster counts.

  15. Frontiers of Two-Dimensional Correlation Spectroscopy. Part 1. New concepts and noteworthy developments

    NASA Astrophysics Data System (ADS)

    Noda, Isao

    2014-07-01

    A comprehensive survey review of new and noteworthy developments, which are advancing forward the frontiers in the field of 2D correlation spectroscopy during the last four years, is compiled. This review covers books, proceedings, and review articles published on 2D correlation spectroscopy, a number of significant conceptual developments in the field, data pretreatment methods and other pertinent topics, as well as patent and publication trends and citation activities. Developments discussed include projection 2D correlation analysis, concatenated 2D correlation, and correlation under multiple perturbation effects, as well as orthogonal sample design, predicting 2D correlation spectra, manipulating and comparing 2D spectra, correlation strategy based on segmented data blocks, such as moving-window analysis, features like determination of sequential order and enhanced spectral resolution, statistical 2D spectroscopy using covariance and other statistical metrics, hetero-correlation analysis, and sample-sample correlation technique. Data pretreatment operations prior to 2D correlation analysis are discussed, including the correction for physical effects, background and baseline subtraction, selection of reference spectrum, normalization and scaling of data, derivatives spectra and deconvolution technique, and smoothing and noise reduction. Other pertinent topics include chemometrics and statistical considerations, peak position shift phenomena, variable sampling increments, computation and software, display schemes, such as color coded format, slice and power spectra, tabulation, and other schemes.

  16. Speeding Up Non-Parametric Bootstrap Computations for Statistics Based on Sample Moments in Small/Moderate Sample Size Applications

    PubMed Central

    Chaibub Neto, Elias

    2015-01-01

    In this paper we propose a vectorized implementation of the non-parametric bootstrap for statistics based on sample moments. Basically, we adopt the multinomial sampling formulation of the non-parametric bootstrap, and compute bootstrap replications of sample moment statistics by simply weighting the observed data according to multinomial counts instead of evaluating the statistic on a resampled version of the observed data. Using this formulation we can generate a matrix of bootstrap weights and compute the entire vector of bootstrap replications with a few matrix multiplications. Vectorization is particularly important for matrix-oriented programming languages such as R, where matrix/vector calculations tend to be faster than scalar operations implemented in a loop. We illustrate the application of the vectorized implementation in real and simulated data sets, when bootstrapping Pearson’s sample correlation coefficient, and compared its performance against two state-of-the-art R implementations of the non-parametric bootstrap, as well as a straightforward one based on a for loop. Our investigations spanned varying sample sizes and number of bootstrap replications. The vectorized bootstrap compared favorably against the state-of-the-art implementations in all cases tested, and was remarkably/considerably faster for small/moderate sample sizes. The same results were observed in the comparison with the straightforward implementation, except for large sample sizes, where the vectorized bootstrap was slightly slower than the straightforward implementation due to increased time expenditures in the generation of weight matrices via multinomial sampling. PMID:26125965

  17. Efficient statistical tests to compare Youden index: accounting for contingency correlation.

    PubMed

    Chen, Fangyao; Xue, Yuqiang; Tan, Ming T; Chen, Pingyan

    2015-04-30

    Youden index is widely utilized in studies evaluating accuracy of diagnostic tests and performance of predictive, prognostic, or risk models. However, both one and two independent sample tests on Youden index have been derived ignoring the dependence (association) between sensitivity and specificity, resulting in potentially misleading findings. Besides, paired sample test on Youden index is currently unavailable. This article develops efficient statistical inference procedures for one sample, independent, and paired sample tests on Youden index by accounting for contingency correlation, namely associations between sensitivity and specificity and paired samples typically represented in contingency tables. For one and two independent sample tests, the variances are estimated by Delta method, and the statistical inference is based on the central limit theory, which are then verified by bootstrap estimates. For paired samples test, we show that the estimated covariance of the two sensitivities and specificities can be represented as a function of kappa statistic so the test can be readily carried out. We then show the remarkable accuracy of the estimated variance using a constrained optimization approach. Simulation is performed to evaluate the statistical properties of the derived tests. The proposed approaches yield more stable type I errors at the nominal level and substantially higher power (efficiency) than does the original Youden's approach. Therefore, the simple explicit large sample solution performs very well. Because we can readily implement the asymptotic and exact bootstrap computation with common software like R, the method is broadly applicable to the evaluation of diagnostic tests and model performance. Copyright © 2015 John Wiley & Sons, Ltd.

  18. In quest of a systematic framework for unifying and defining nanoscience

    PubMed Central

    2009-01-01

    This article proposes a systematic framework for unifying and defining nanoscience based on historic first principles and step logic that led to a “central paradigm” (i.e., unifying framework) for traditional elemental/small-molecule chemistry. As such, a Nanomaterials classification roadmap is proposed, which divides all nanomatter into Category I: discrete, well-defined and Category II: statistical, undefined nanoparticles. We consider only Category I, well-defined nanoparticles which are >90% monodisperse as a function of Critical Nanoscale Design Parameters (CNDPs) defined according to: (a) size, (b) shape, (c) surface chemistry, (d) flexibility, and (e) elemental composition. Classified as either hard (H) (i.e., inorganic-based) or soft (S) (i.e., organic-based) categories, these nanoparticles were found to manifest pervasive atom mimicry features that included: (1) a dominance of zero-dimensional (0D) core–shell nanoarchitectures, (2) the ability to self-assemble or chemically bond as discrete, quantized nanounits, and (3) exhibited well-defined nanoscale valencies and stoichiometries reminiscent of atom-based elements. These discrete nanoparticle categories are referred to as hard or soft particle nanoelements. Many examples describing chemical bonding/assembly of these nanoelements have been reported in the literature. We refer to these hard:hard (H-n:H-n), soft:soft (S-n:S-n), or hard:soft (H-n:S-n) nanoelement combinations as nanocompounds. Due to their quantized features, many nanoelement and nanocompound categories are reported to exhibit well-defined nanoperiodic property patterns. These periodic property patterns are dependent on their quantized nanofeatures (CNDPs) and dramatically influence intrinsic physicochemical properties (i.e., melting points, reactivity/self-assembly, sterics, and nanoencapsulation), as well as important functional/performance properties (i.e., magnetic, photonic, electronic, and toxicologic properties). We propose this perspective as a modest first step toward more clearly defining synthetic nanochemistry as well as providing a systematic framework for unifying nanoscience. With further progress, one should anticipate the evolution of future nanoperiodic table(s) suitable for predicting important risk/benefit boundaries in the field of nanoscience. Electronic supplementary material The online version of this article (doi:10.1007/s11051-009-9632-z) contains supplementary material, which is available to authorized users. PMID:21170133

  19. Framework for making better predictions by directly estimating variables’ predictivity

    PubMed Central

    Chernoff, Herman; Lo, Shaw-Hwa

    2016-01-01

    We propose approaching prediction from a framework grounded in the theoretical correct prediction rate of a variable set as a parameter of interest. This framework allows us to define a measure of predictivity that enables assessing variable sets for, preferably high, predictivity. We first define the prediction rate for a variable set and consider, and ultimately reject, the naive estimator, a statistic based on the observed sample data, due to its inflated bias for moderate sample size and its sensitivity to noisy useless variables. We demonstrate that the I-score of the PR method of VS yields a relatively unbiased estimate of a parameter that is not sensitive to noisy variables and is a lower bound to the parameter of interest. Thus, the PR method using the I-score provides an effective approach to selecting highly predictive variables. We offer simulations and an application of the I-score on real data to demonstrate the statistic’s predictive performance on sample data. We conjecture that using the partition retention and I-score can aid in finding variable sets with promising prediction rates; however, further research in the avenue of sample-based measures of predictivity is much desired. PMID:27911830

  20. Specificity of Incident Diagnostic Outcomes in Patients at Clinical High Risk for Psychosis

    PubMed Central

    Webb, Jadon R; Addington, Jean; Perkins, Diana O; Bearden, Carrie E; Cadenhead, Kristin S; Cannon, Tyrone D; Cornblatt, Barbara A; Heinssen, Robert K; Seidman, Larry J; Tarbox, Sarah I; Tsuang, Ming; Walker, Elaine; McGlashan, Thomas H; Woods, Scott W

    2015-01-01

    Abstract It is not well established whether the incident outcomes of the clinical high-risk (CHR) syndrome for psychosis are diagnostically specific for psychosis or whether CHR patients also are at elevated risk for a variety of nonpsychotic disorders. We collected 2 samples (NAPLS-1, PREDICT) that contained CHR patients and a control group who responded to CHR recruitment efforts but did not meet CHR criteria on interview (help-seeking comparison patients [HSC]). Incident diagnostic outcomes were defined as the occurrence of a SIPS-defined psychosis or a structured interview diagnosis from 1 of 3 nonpsychotic Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition (DSM-IV) groups (anxiety, bipolar, or nonbipolar mood disorder), when no diagnosis in that group was present at baseline. Logistic regression revealed that the CHR vs HSC effect did not vary significantly across study for any emergent diagnostic outcome; data from the 2 studies were therefore combined. CHR (n = 271) vs HSC (n = 171) emergent outcomes were: psychosis 19.6% vs 1.8%, bipolar disorders 1.1% vs 1.2%, nonbipolar mood disorders 4.4% vs 5.3%, and anxiety disorders 5.2% vs 5.3%. The main effect of CHR vs HSC was statistically significant (OR = 13.8, 95% CI 4.2–45.0, df = 1, P < .001) for emergent psychosis but not for any emergent nonpsychotic disorder. Sensitivity analyses confirmed these findings. Within the CHR group emergent psychosis was significantly more likely than each nonpsychotic DSM-IV emergent disorder, and within the HSC group emergent psychosis was significantly less likely than most emergent nonpsychotic disorders. The CHR syndrome is specific as a marker for research on predictors and mechanisms of developing psychosis. PMID:26272875

  1. Rethinking the Clinically Based Thresholds of TransCelerate BioPharma for Risk-Based Monitoring.

    PubMed

    Zink, Richard C; Dmitrienko, Anastasia; Dmitrienko, Alex

    2018-01-01

    The quality of data from clinical trials has received a great deal of attention in recent years. Of central importance is the need to protect the well-being of study participants and maintain the integrity of final analysis results. However, traditional approaches to assess data quality have come under increased scrutiny as providing little benefit for the substantial cost. Numerous regulatory guidance documents and industry position papers have described risk-based approaches to identify quality and safety issues. In particular, the position paper of TransCelerate BioPharma recommends defining risk thresholds to assess safety and quality risks based on past clinical experience. This exercise can be extremely time-consuming, and the resulting thresholds may only be relevant to a particular therapeutic area, patient or clinical site population. In addition, predefined thresholds cannot account for safety or quality issues where the underlying rate of observing a particular problem may change over the course of a clinical trial, and often do not consider varying patient exposure. In this manuscript, we appropriate rules commonly utilized for funnel plots to define a traffic-light system for risk indicators based on statistical criteria that consider the duration of patient follow-up. Further, we describe how these methods can be adapted to assess changing risk over time. Finally, we illustrate numerous graphical approaches to summarize and communicate risk, and discuss hybrid clinical-statistical approaches to allow for the assessment of risk at sites with low patient enrollment. We illustrate the aforementioned methodologies for a clinical trial in patients with schizophrenia. Funnel plots are a flexible graphical technique that can form the basis for a risk-based strategy to assess data integrity, while considering site sample size, patient exposure, and changing risk across time.

  2. Temporal changes in water quality at a childhood leukemia cluster.

    PubMed

    Seiler, Ralph L

    2004-01-01

    Since 1997, 15 cases of acute lymphocytic leukemia and one case of acute myelocytic leukemia have been diagnosed in children and teenagers who live, or have lived, in an area centered on the town of Fallon, Nevada. The expected rate for the population is about one case every five years. In 2001, 99 domestic and municipal wells and one industrial well were sampled in the Fallon area. Twenty-nine of these wells had been sampled previously in 1989. Statistical comparison of concentrations of major ions and trace elements in those 29 wells between 1989 and 2001 using the nonparametric Wilcoxon signed-rank test indicate water quality did not substantially change over that period; however, short-term changes may have occurred that were not detected. Volatile organic compounds were seldom detected in ground water samples and those that are regulated were consistently found at concentrations less than the maximum contaminant level (MCL). The MCL for gross-alpha radioactivity and arsenic, radon, and uranium concentrations were commonly exceeded, and sometimes were greatly exceeded. Statistical comparisons using the nonparametric Wilcoxon rank-sum test indicate gross-alpha and -beta radioactivity, arsenic, uranium, and radon concentrations in wells used by families having a child with leukemia did not statistically differ from the remainder of the domestic wells sampled during this investigation. Isotopic measurements indicate the uranium was natural and not the result of a 1963 underground nuclear bomb test near Fallon. In arid and semiarid areas where trace-element concentrations can greatly exceed the MCL, household reverse-osmosis units may not reduce their concentrations to safe levels. In parts of the world where radon concentrations are high, water consumed first thing in the morning may be appreciably more radioactive than water consumed a few minutes later after the pressure tank has been emptied because secular equilibrium between radon and its immediate daughter progeny is attained in pressure tanks overnight.

  3. Temporal changes in water quality at a childhood leukemia cluster

    USGS Publications Warehouse

    Seiler, R.L.

    2004-01-01

    Since 1997, 15 cases of acute lymphocytic leukemia and one case of acute myelocytic leukemia have been diagnosed in children and teenagers who live, or have lived, in an area centered on the town of Fallon, Nevada. The expected rate for the population is about one case every five years. In 2001, 99 domestic and municipal wells and one industrial well were sampled in the Fallon area. Twenty-nine of these wells had been sampled previously in 1989. Statistical comparison of concentrations of major ions and trace elements in those 29 wells between 1989 and 2001 using the nonparametric Wilcoxon signed-rank test indicate water quality did not substantially change over that period; however, short-term changes may have occurred that were not detected. Volatile organic compounds were seldom detected in ground water samples and those that are regulated were consistently found at concentrations less than the maximum contaminant level (MCL). The MCL for gross-alpha radioactivity and arsenic, radon, and uranium concentrations were commonly exceeded, and sometimes were greatly exceeded. Statistical comparisons using the nonparametric Wilcoxon rank-sum test indicate gross-alpha and -beta radioactivity, arsenic, uranium, and radon concentrations in wells used by families having a child with leukemia did not statistically differ from the remainder of the domestic wells sampled during this investigation. Isotopic measurements indicate the uranium was natural and not the result of a 1963 underground nuclear bomb test near Fallon. In arid and semiarid areas where trace-element concentrations can greatly exceed the MCL, household reverse-osmosis units may not reduce their concentrations to safe levels. In parts of the world where radon concentrations are high, water consumed first thing in the morning may be appreciably more radioactive than water consumed a few minutes later after the pressure tank has been emptied because secular equilibrium between radon and its immediate daughter progeny is attained in pressure tanks overnight.

  4. Classification and recognition of dynamical models: the role of phase, independent components, kernels and optimal transport.

    PubMed

    Bissacco, Alessandro; Chiuso, Alessandro; Soatto, Stefano

    2007-11-01

    We address the problem of performing decision tasks, and in particular classification and recognition, in the space of dynamical models in order to compare time series of data. Motivated by the application of recognition of human motion in image sequences, we consider a class of models that include linear dynamics, both stable and marginally stable (periodic), both minimum and non-minimum phase, driven by non-Gaussian processes. This requires extending existing learning and system identification algorithms to handle periodic modes and nonminimum phase behavior, while taking into account higher-order statistics of the data. Once a model is identified, we define a kernel-based cord distance between models that includes their dynamics, their initial conditions as well as input distribution. This is made possible by a novel kernel defined between two arbitrary (non-Gaussian) distributions, which is computed by efficiently solving an optimal transport problem. We validate our choice of models, inference algorithm, and distance on the tasks of human motion synthesis (sample paths of the learned models), and recognition (nearest-neighbor classification in the computed distance). However, our work can be applied more broadly where one needs to compare historical data while taking into account periodic trends, non-minimum phase behavior, and non-Gaussian input distributions.

  5. Statistics Refresher for Molecular Imaging Technologists, Part 2: Accuracy of Interpretation, Significance, and Variance.

    PubMed

    Farrell, Mary Beth

    2018-06-01

    This article is the second part of a continuing education series reviewing basic statistics that nuclear medicine and molecular imaging technologists should understand. In this article, the statistics for evaluating interpretation accuracy, significance, and variance are discussed. Throughout the article, actual statistics are pulled from the published literature. We begin by explaining 2 methods for quantifying interpretive accuracy: interreader and intrareader reliability. Agreement among readers can be expressed simply as a percentage. However, the Cohen κ-statistic is a more robust measure of agreement that accounts for chance. The higher the κ-statistic is, the higher is the agreement between readers. When 3 or more readers are being compared, the Fleiss κ-statistic is used. Significance testing determines whether the difference between 2 conditions or interventions is meaningful. Statistical significance is usually expressed using a number called a probability ( P ) value. Calculation of P value is beyond the scope of this review. However, knowing how to interpret P values is important for understanding the scientific literature. Generally, a P value of less than 0.05 is considered significant and indicates that the results of the experiment are due to more than just chance. Variance, standard deviation (SD), confidence interval, and standard error (SE) explain the dispersion of data around a mean of a sample drawn from a population. SD is commonly reported in the literature. A small SD indicates that there is not much variation in the sample data. Many biologic measurements fall into what is referred to as a normal distribution taking the shape of a bell curve. In a normal distribution, 68% of the data will fall within 1 SD, 95% will fall within 2 SDs, and 99.7% will fall within 3 SDs. Confidence interval defines the range of possible values within which the population parameter is likely to lie and gives an idea of the precision of the statistic being measured. A wide confidence interval indicates that if the experiment were repeated multiple times on other samples, the measured statistic would lie within a wide range of possibilities. The confidence interval relies on the SE. © 2018 by the Society of Nuclear Medicine and Molecular Imaging.

  6. Socioeconomic Inequality in Malnutrition in Under-5 Children in Iran: Evidence From the Multiple Indicator Demographic and Health Survey, 2010.

    PubMed

    Almasian Kia, Abdollah; Rezapour, Aziz; Khosravi, Ardeshir; Afzali Abarghouei, Vajiheh

    2017-01-01

    The aim of this study was to assess the socioeconomic inequality in malnutrition in under-5 children in Iran in order to help policymakers reduce such inequality. Data on 8443 under-5 children were extracted from the Iran Multiple Indicator Demographic and Health Survey. The wealth index was used as proxy for socioeconomic status. Socioeconomic inequality in stunting, underweight, and wasting was calculated using the concentration index. The concentration index was calculated for the whole sample, as well as for subcategories defined in terms of categories such as area of residence (urban and rural) and the sex of children. Stunting was observed to be more prevalent than underweight or wasting. The results of the concentration index at the national level, as well as in rural and urban areas and in terms of children's sex, showed that inequality in stunting and underweight was statistically significant and that children in the lower quintiles were more malnourished. The wasting index was not sensitive to socioeconomic status, and its concentration index value was not statistically significant. This study showed that it can be misleading to assess the mean levels of malnutrition at the national level without knowledge of the distribution of malnutrition among socioeconomic groups. Significant socioeconomic inequalities in stunting and underweight were observed at the national level and in both urban and rural areas. Regarding the influence of nutrition on the health and economic well-being of preschool-aged children, it is necessary for the government to focus on taking targeted measures to reduce malnutrition and to focus on poorer groups within society who bear a greater burden of malnutrition.

  7. Socioeconomic Inequality in Malnutrition in Under-5 Children in Iran: Evidence From the Multiple Indicator Demographic and Health Survey, 2010

    PubMed Central

    2017-01-01

    Objectives The aim of this study was to assess the socioeconomic inequality in malnutrition in under-5 children in Iran in order to help policymakers reduce such inequality. Methods Data on 8443 under-5 children were extracted from the Iran Multiple Indicator Demographic and Health Survey. The wealth index was used as proxy for socioeconomic status. Socioeconomic inequality in stunting, underweight, and wasting was calculated using the concentration index. The concentration index was calculated for the whole sample, as well as for subcategories defined in terms of categories such as area of residence (urban and rural) and the sex of children. Results Stunting was observed to be more prevalent than underweight or wasting. The results of the concentration index at the national level, as well as in rural and urban areas and in terms of children’s sex, showed that inequality in stunting and underweight was statistically significant and that children in the lower quintiles were more malnourished. The wasting index was not sensitive to socioeconomic status, and its concentration index value was not statistically significant. Conclusions This study showed that it can be misleading to assess the mean levels of malnutrition at the national level without knowledge of the distribution of malnutrition among socioeconomic groups. Significant socioeconomic inequalities in stunting and underweight were observed at the national level and in both urban and rural areas. Regarding the influence of nutrition on the health and economic well-being of preschool-aged children, it is necessary for the government to focus on taking targeted measures to reduce malnutrition and to focus on poorer groups within society who bear a greater burden of malnutrition. PMID:28605886

  8. Soil aggregate stability and wind erodible fraction in a semi-arid environment of White Nile State, Sudan

    NASA Astrophysics Data System (ADS)

    Elhaja, Mohamed Eltom; Ibrahim, Ibrahim Saeed; Adam, Hassan Elnour; Csaplovics, Elmar

    2014-11-01

    One of the most important recent issues facing White Nile State, Sudan, as well as Sub Saharan Africa, is the threat of continued land degradation and desertification as a result of climatic factors and human activities. Remote sensing and satellites imageries with multi-temporal and spectral and GIS capability, plays a major role in developing a global and local operational capability for monitoring land degradation and desertification in dry lands, as well as in White Nile State. The process of desertification in form of sand encroachment in White Nile State has increased rapidly, and much effort has been devoted to define and study its causes and impacts. This study depicts the capability afforded by remote sensing and GIS to analyze and map the aggregate stability as indicator for the ability of soil to wind erosion process in White Nile State by using Geo-statistical techniques. Cloud-free subset Landsat; Enhance Thematic Mapper plus (ETM +) scenes covering the study area dated 2008 was selected in order to identify the different features covering the study area as well as to make the soil sampling map. Wet-sieving method was applied to determine the aggregate stability. The geo-statistical methods in EARDAS 9.1 software was used for mapping the aggregate stability. The results showed that the percentage of aggregate stability ranged from (0 to 61%) in the study area, which emphasized the phenomena of sand encroachment from the western part (North Kordofan) to the eastern part (White Nile State), following the wind direction. The study comes out with some valuable recommendations and comments, which could contribute positively in reducing sand encroachments

  9. Using Computer Graphics in Statistics.

    ERIC Educational Resources Information Center

    Kerley, Lyndell M.

    1990-01-01

    Described is software which allows a student to use simulation to produce analytical output as well as graphical results. The results include a frequency histogram of a selected population distribution, a frequency histogram of the distribution of the sample means, and test the normality distributions of the sample means. (KR)

  10. Protein Multiplexed Immunoassay Analysis with R.

    PubMed

    Breen, Edmond J

    2017-01-01

    Plasma samples from 177 control and type 2 diabetes patients collected at three Australian hospitals are screened for 14 analytes using six custom-made multiplex kits across 60 96-well plates. In total 354 samples were collected from the patients, representing one baseline and one end point sample from each patient. R methods and source code for analyzing the analyte fluorescence response obtained from these samples by Luminex Bio-Plex ® xMap multiplexed immunoassay technology are disclosed. Techniques and R procedures for reading Bio-Plex ® result files for statistical analysis and data visualization are also presented. The need for technical replicates and the number of technical replicates are addressed as well as plate layout design strategies. Multinomial regression is used to determine plate to sample covariate balance. Methods for matching clinical covariate information to Bio-Plex ® results and vice versa are given. As well as methods for measuring and inspecting the quality of the fluorescence responses are presented. Both fixed and mixed-effect approaches for immunoassay statistical differential analysis are presented and discussed. A random effect approach to outlier analysis and detection is also shown. The bioinformatics R methodology present here provides a foundation for rigorous and reproducible analysis of the fluorescence response obtained from multiplexed immunoassays.

  11. RANdom SAmple Consensus (RANSAC) algorithm for material-informatics: application to photovoltaic solar cells.

    PubMed

    Kaspi, Omer; Yosipof, Abraham; Senderowitz, Hanoch

    2017-06-06

    An important aspect of chemoinformatics and material-informatics is the usage of machine learning algorithms to build Quantitative Structure Activity Relationship (QSAR) models. The RANdom SAmple Consensus (RANSAC) algorithm is a predictive modeling tool widely used in the image processing field for cleaning datasets from noise. RANSAC could be used as a "one stop shop" algorithm for developing and validating QSAR models, performing outlier removal, descriptors selection, model development and predictions for test set samples using applicability domain. For "future" predictions (i.e., for samples not included in the original test set) RANSAC provides a statistical estimate for the probability of obtaining reliable predictions, i.e., predictions within a pre-defined number of standard deviations from the true values. In this work we describe the first application of RNASAC in material informatics, focusing on the analysis of solar cells. We demonstrate that for three datasets representing different metal oxide (MO) based solar cell libraries RANSAC-derived models select descriptors previously shown to correlate with key photovoltaic properties and lead to good predictive statistics for these properties. These models were subsequently used to predict the properties of virtual solar cells libraries highlighting interesting dependencies of PV properties on MO compositions.

  12. CAN'T MISS--conquer any number task by making important statistics simple. Part 1. Types of variables, mean, median, variance, and standard deviation.

    PubMed

    Hansen, John P

    2003-01-01

    Healthcare quality improvement professionals need to understand and use inferential statistics to interpret sample data from their organizations. In quality improvement and healthcare research studies all the data from a population often are not available, so investigators take samples and make inferences about the population by using inferential statistics. This three-part series will give readers an understanding of the concepts of inferential statistics as well as the specific tools for calculating confidence intervals for samples of data. This article, Part 1, presents basic information about data including a classification system that describes the four major types of variables: continuous quantitative variable, discrete quantitative variable, ordinal categorical variable (including the binomial variable), and nominal categorical variable. A histogram is a graph that displays the frequency distribution for a continuous variable. The article also demonstrates how to calculate the mean, median, standard deviation, and variance for a continuous variable.

  13. Association Between Early Idiopathic Neonatal Jaundice and Urinary Tract Infections

    PubMed Central

    Özcan, Murat; Sarici, S Ümit; Yurdugül, Yüksel; Akpinar, Melis; Altun, Demet; Özcan, Begüm; Serdar, Muhittin A; Sarici, Dilek

    2017-01-01

    Background and purpose: Etiologic role, incidence, demographic, and response-to-treatment characteristics of urinary tract infection (UTI) among neonates, its relationship with significant neonatal hyperbilirubinemia, and abnormalities of the urinary system were studied in a prospective investigation in early (≤10 days) idiopathic neonatal jaundice in which all other etiologic factors of neonatal hyperbilirubinemia were ruled out. Patients and methods: Urine samples for microscopic and bacteriologic examination were obtained with bladder catheterization from 155 newborns with early neonatal jaundice. Newborns with a negative urine culture and with a positive urine culture were defined as group I and group II, respectively, and the 2 groups were compared with each other. Results: The incidence of UTI in whole of the study group was 16.7%. Serum total and direct bilirubin levels were statistically significantly higher in group II when compared with group I (P = .005 and P = .001, respectively). Decrease in serum total bilirubin level at the 24th hour of phototherapy was statistically significantly higher in group I compared with group II (P = .022). Conclusions: Urinary tract infection should be investigated in the etiologic evaluation of newborns with significant hyperbilirubinemia. The possibility of UTI should be considered in jaundiced newborns who do not respond to phototherapy well or have a prolonged duration of phototherapy treatment. PMID:28469520

  14. msap: a tool for the statistical analysis of methylation-sensitive amplified polymorphism data.

    PubMed

    Pérez-Figueroa, A

    2013-05-01

    In this study msap, an R package which analyses methylation-sensitive amplified polymorphism (MSAP or MS-AFLP) data is presented. The program provides a deep analysis of epigenetic variation starting from a binary data matrix indicating the banding pattern between the isoesquizomeric endonucleases HpaII and MspI, with differential sensitivity to cytosine methylation. After comparing the restriction fragments, the program determines if each fragment is susceptible to methylation (representative of epigenetic variation) or if there is no evidence of methylation (representative of genetic variation). The package provides, in a user-friendly command line interface, a pipeline of different analyses of the variation (genetic and epigenetic) among user-defined groups of samples, as well as the classification of the methylation occurrences in those groups. Statistical testing provides support to the analyses. A comprehensive report of the analyses and several useful plots could help researchers to assess the epigenetic and genetic variation in their MSAP experiments. msap is downloadable from CRAN (http://cran.r-project.org/) and its own webpage (http://msap.r-forge.R-project.org/). The package is intended to be easy to use even for those people unfamiliar with the R command line environment. Advanced users may take advantage of the available source code to adapt msap to more complex analyses. © 2013 Blackwell Publishing Ltd.

  15. Robustness of movement models: can models bridge the gap between temporal scales of data sets and behavioural processes?

    PubMed

    Schlägel, Ulrike E; Lewis, Mark A

    2016-12-01

    Discrete-time random walks and their extensions are common tools for analyzing animal movement data. In these analyses, resolution of temporal discretization is a critical feature. Ideally, a model both mirrors the relevant temporal scale of the biological process of interest and matches the data sampling rate. Challenges arise when resolution of data is too coarse due to technological constraints, or when we wish to extrapolate results or compare results obtained from data with different resolutions. Drawing loosely on the concept of robustness in statistics, we propose a rigorous mathematical framework for studying movement models' robustness against changes in temporal resolution. In this framework, we define varying levels of robustness as formal model properties, focusing on random walk models with spatially-explicit component. With the new framework, we can investigate whether models can validly be applied to data across varying temporal resolutions and how we can account for these different resolutions in statistical inference results. We apply the new framework to movement-based resource selection models, demonstrating both analytical and numerical calculations, as well as a Monte Carlo simulation approach. While exact robustness is rare, the concept of approximate robustness provides a promising new direction for analyzing movement models.

  16. Evaluating the best time to intervene acute liver failure in rat models induced by d-galactosamine.

    PubMed

    Éboli, Lígia Patrícia de Carvalho Batista; Netto, Alcides Augusto Salzedas; Azevedo, Ramiro Antero de; Lanzoni, Valéria Pereira; Paula, Tatiana Sugayama de; Goldenberg, Alberto; Gonzalez, Adriano Miziara

    2016-12-01

    To describe an animal model for acute liver failure by intraperitoneal d-galactosamine injections in rats and to define when is the best time to intervene through King's College and Clichy´s criteria evaluation. Sixty-one Wistar female rats were distributed into three groups: group 1 (11 rats received 1.4 g/kg of d-galactosamine intraperitoneally and were observed until they died); group 2 (44 rats received a dose of 1.4 g/kg of d-galactosamine and blood and histological samples were collected for analysis at 12 , 24, 48 , 72 and 120 hours after the injection); and the control group as well (6 rats) . Twelve hours after applying d-galactosamine, AST/ALT, bilirubin, factor V, PT and INR were already altered. The peak was reached at 48 hours. INR > 6.5 was found 12 hours after the injection and factor V < 30% after 24 hours. All the laboratory variables presented statistical differences, except urea (p = 0.758). There were statistical differences among all the histological variables analyzed. King's College and Clichy´s criteria were fulfilled 12 hours after the d-galactosamine injection and this time may represent the best time to intervene in this acute liver failure animal model.

  17. Identification of the country of growth of Sophora flavescens using direct analysis in real time mass spectrometry (DART-MS).

    PubMed

    Fukuda, Eriko; Uesawa, Yoshihiro; Baba, Masaki; Suzuki, Ryuichiro; Fukuda, Tatsuo; Shirataki, Yoshiaki; Okada, Yoshihito

    2014-11-01

    In order to identify the country of growth of Sophora flavescens by chemical fingerprinting, extracts of plants grown in China and Japan were analyzed using direct analysis in real time mass spectrometry (DART)-MS. The peaks characteristic of each country of growth were statistically analyzed using a volcano plot to summarize the relationship between the p-values of a statistical test and the magnitude of the difference in the peak intensities of the samples in the groups. Peaks with ap value < 0.05 in the t-test and a ≥ 2 absolute difference were defined as characteristic. Peaks characteristic of Chinese S. flavescens were found at m/z 439 and 440. In contrast, peaks characteristic of Japanese S. flavescens were found at m/z 313, 423, 437 and 441. The intensity of the selected peaks was similar in Japanese samples, whereas the m/z 439 peak had a significantly higher intensity than the other peaks in Chinese samples. Therefore, differences in selected peak patterns may allow identification of the country of growth of S. flavescens.

  18. Effects of Long Term Thermal Exposure on Chemically Pure (CP) Titanium Grade 2 Room Temperature Tensile Properties and Microstructure

    NASA Technical Reports Server (NTRS)

    Ellis, David L.

    2007-01-01

    Room temperature tensile testing of Chemically Pure (CP) Titanium Grade 2 was conducted for as-received commercially produced sheet and following thermal exposure at 550 and 650 K for times up to 5,000 h. No significant changes in microstructure or failure mechanism were observed. A statistical analysis of the data was performed. Small statistical differences were found, but all properties were well above minimum values for CP Ti Grade 2 as defined by ASTM standards and likely would fall within normal variation of the material.

  19. Thermal equilibrium and statistical thermometers in special relativity.

    PubMed

    Cubero, David; Casado-Pascual, Jesús; Dunkel, Jörn; Talkner, Peter; Hänggi, Peter

    2007-10-26

    There is an intense debate in the recent literature about the correct generalization of Maxwell's velocity distribution in special relativity. The most frequently discussed candidate distributions include the Jüttner function as well as modifications thereof. Here we report results from fully relativistic one-dimensional molecular dynamics simulations that resolve the ambiguity. The numerical evidence unequivocally favors the Jüttner distribution. Moreover, our simulations illustrate that the concept of "thermal equilibrium" extends naturally to special relativity only if a many-particle system is spatially confined. They make evident that "temperature" can be statistically defined and measured in an observer frame independent way.

  20. Non-abelian anyons and topological quantum information processing in 1D wire networks

    NASA Astrophysics Data System (ADS)

    Alicea, Jason

    2012-02-01

    Topological quantum computation provides an elegant solution to decoherence, circumventing this infamous problem at the hardware level. The most basic requirement in this approach is the ability to stabilize and manipulate particles exhibiting non-Abelian exchange statistics -- Majorana fermions being the simplest example. Curiously, Majorana fermions have been predicted to arise both in 2D systems, where non-Abelian statistics is well established, and in 1D, where exchange statistics of any type is ill-defined. An important question then arises: do Majorana fermions in 1D hold the same technological promise as their 2D counterparts? In this talk I will answer this question in the affirmative, describing how one can indeed manipulate and harness the non-Abelian statistics of Majoranas in a remarkably simple fashion using networks formed by quantum wires or topological insulator edges.

  1. Comparison of dialysis membrane diffusion samplers and two purging methods in bedrock wells

    USGS Publications Warehouse

    Imbrigiotta, T.E.; Ehlke, T.A.; Lacombe, P.J.; Dale, J.M.; ,

    2002-01-01

    Collection of ground-water samples from bedrock wells using low-flow purging techniques is problematic because of the random spacing, variable hydraulic conductivity, and variable contamination of contributing fractures in each well's open interval. To test alternatives to this purging method, a field comparison of three ground-water-sampling techniques was conducted on wells in fractured bedrock at a site contaminated primarily with volatile organic compounds. Constituent concentrations in samples collected with a diffusion sampler constructed from dialysis membrane material were compared to those in samples collected from the same wells with a standard low-flow purging technique and a hybrid (high-flow/low-flow) purging technique. Concentrations of trichloroethene, cis-1,2-dichloroethene, vinyl chloride, calcium, chloride, and alkalinity agreed well among samples collected with all three techniques in 9 of the 10 wells tested. Iron concentrations varied more than those of the other parameters, but their pattern of variation was not consistent. Overall, the results of nonparametric analysis of variance testing on the nine wells sampled twice showed no statistically significant difference at the 95-percent confidence level among the concentrations of volatile organic compounds or inorganic constituents recovered by use of any of the three sampling techniques.

  2. 3-dimensional examination of the adult mouse subventricular zone reveals lineage-specific microdomains.

    PubMed

    Azim, Kasum; Fiorelli, Roberto; Zweifel, Stefan; Hurtado-Chong, Anahi; Yoshikawa, Kazuaki; Slomianka, Lutz; Raineteau, Olivier

    2012-01-01

    Recent studies suggest that the subventricular zone (SVZ) of the lateral ventricle is populated by heterogeneous populations of stem and progenitor cells that, depending on their exact location, are biased to acquire specific neuronal fates. This newly described heterogeneity of SVZ stem and progenitor cells underlines the necessity to develop methods for the accurate quantification of SVZ stem and progenitor subpopulations. In this study, we provide 3-dimensional topographical maps of slow cycling "stem" cells and progenitors based on their unique cell cycle properties. These maps revealed that both cell populations are present throughout the lateral ventricle wall as well as in discrete regions of the dorsal wall. Immunodetection of transcription factors expressed in defined progenitor populations further reveals that divergent lineages have clear regional enrichments in the rostro-caudal as well as in the dorso-ventral span of the lateral ventricle. Thus, progenitors expressing Tbr2 and Dlx2 were confined to dorsal and dorso-lateral regions of the lateral ventricle, respectively, while Mash1+ progenitors were more homogeneously distributed. All cell populations were enriched in the rostral-most region of the lateral ventricle. This diversity and uneven distribution greatly impede the accurate quantification of SVZ progenitor populations. This is illustrated by measuring the coefficient of error of estimates obtained by using increasing section sampling interval. Based on our empirical data, we provide such estimates for all progenitor populations investigated in this study. These can be used in future studies as guidelines to judge if the precision obtained with a sampling scheme is sufficient to detect statistically significant differences between experimental groups if a biological effect is present. Altogether, our study underlines the need to consider the SVZ of the lateral ventricle as a complex 3D structure and define methods to accurately assess neural stem cells or progenitor diversity and population sizes in physiological or experimental paradigms.

  3. Determining Differences in Social Cognition between High-Functioning Autistic Disorder and Other Pervasive Developmental Disorders Using New Advanced "Mind-Reading" Tasks

    ERIC Educational Resources Information Center

    Kuroda, Miho; Wakabayashi, Akio; Uchiyama, Tokio; Yoshida, Yuko; Koyama, Tomonori; Kamio, Yoko

    2011-01-01

    Deficits in understanding the mental state of others ("mind-reading") have been well documented in individuals with pervasive developmental disorders (PDD). However, it is unclear whether this deficit in social cognition differs between the subgroups of PDD defined by the Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition, Text…

  4. Analysis of ground-water data for selected wells near Holloman Air Force Base, New Mexico, 1950-95

    USGS Publications Warehouse

    Huff, G.F.

    1996-01-01

    Ground-water-level, ground-water-withdrawal, and ground- water-quality data were evaluated for trends. Holloman Air Force Base is located in the west-central part of Otero County, New Mexico. Ground-water-data analyses include assembly and inspection of U.S. Geological Survey and Holloman Air Force Base data, including ground-water-level data for public-supply and observation wells and withdrawal and water-quality data for public-supply wells in the area. Well Douglas 4 shows a statistically significant decreasing trend in water levels for 1972-86 and a statistically significant increasing trend in water levels for 1986-90. Water levels in wells San Andres 5 and San Andres 6 show statistically significant decreasing trends for 1972-93 and 1981-89, respectively. A mixture of statistically significant increasing trends, statistically significant decreasing trends, and lack of statistically significant trends over periods ranging from the early 1970's to the early 1990's are indicated for the Boles wells and wells near the Boles wells. Well Boles 5 shows a statistically significant increasing trend in water levels for 1981-90. Well Boles 5 and well 17S.09E.25.343 show no statistically significant trends in water levels for 1990-93 and 1988-93, respectively. For 1986-93, well Frenchy 1 shows a statistically significant decreasing trend in water levels. Ground-water withdrawal from the San Andres and Douglas wells regularly exceeded estimated ground-water recharge from San Andres Canyon for 1963-87. For 1951-57 and 1960-86, ground-water withdrawal from the Boles wells regularly exceeded total estimated ground-water recharge from Mule, Arrow, and Lead Canyons. Ground-water withdrawal from the San Andres and Douglas wells and from the Boles wells nearly equaled estimated ground- water recharge for 1989-93 and 1986-93, respectively. For 1987- 93, ground-water withdrawal from the Escondido well regularly exceeded estimated ground-water recharge from Escondido Canyon, and ground-water withdrawal from the Frenchy wells regularly exceeded total estimated ground-water recharge from Dog and Deadman Canyons. Water-quality samples were collected from selected Douglas, San Andres, and Boles public-supply wells from December 1994 to February 1995. Concentrations of dissolved nitrate show the most consistent increases between current and historical data. Current concentrations of dissolved nitrate are greater than historical concentrations in 7 of 10 wells.

  5. Statistical techniques for sampling and monitoring natural resources

    Treesearch

    Hans T. Schreuder; Richard Ernst; Hugo Ramirez-Maldonado

    2004-01-01

    We present the statistical theory of inventory and monitoring from a probabilistic point of view. We start with the basics and show the interrelationships between designs and estimators illustrating the methods with a small artificial population as well as with a mapped realistic population. For such applications, useful open source software is given in Appendix 4....

  6. Animating Statistics: A New Kind of Applet for Exploring Probability Distributions

    ERIC Educational Resources Information Center

    Kahle, David

    2014-01-01

    In this article, I introduce a novel applet ("module") for exploring probability distributions, their samples, and various related statistical concepts. The module is primarily designed to be used by the instructor in the introductory course, but it can be used far beyond it as well. It is a free, cross-platform, stand-alone interactive…

  7. Validation of a single-platform method for hematopoietic CD34+ stem cells enumeration according to accreditation procedure.

    PubMed

    Massin, Frédéric; Huili, Cai; Decot, Véronique; Stoltz, Jean-François; Bensoussan, Danièle; Latger-Cannard, Véronique

    2015-01-01

    Stem cells for autologous and allogenic transplantation are obtained from several sources including bone marrow, peripheral blood or cord blood. Accurate enumeration of viable CD34+ hematopoietic stem cells (HSC) is routinely used in clinical settings, especially to monitor progenitor cell mobilization and apheresis. The number of viable CD34+ HSC has also been shown to be the most critical factor in haematopoietic engraftment. The International Society for Cellular Therapy actually recommends the use of single-platform flow cytometry system using 7-AAD as a viability dye. In a way to move routine analysis from a BD FACSCaliburTM instrument to a BD FACSCantoTM II, according to ISO 15189 standard guidelines, we define laboratory performance data of the BDTM Stem Cell Enumeration (SCE) kit on a CE-IVD system including a BD FACSCanto II flow cytometer and the BD FACSCantoTM Clinical Software. InterQCTM software, a real time internet laboratory QC management system developed by VitroTM and distributed by Becton DickinsonTM, was also tested to monitor daily QC data, to define the internal laboratory statistics and to compare them to external laboratories. Precision was evaluated with BDTM Stem Cell Control (high and low) results and the InterQC software, an internet laboratory QC management system by Vitro. This last one drew Levey-Jennings curves and generated numeral statistical parameters allowing detection of potential changes in the system performances as well as interlaboratory comparisons. Repeatability, linearity and lower limits of detection were obtained with routine samples from different origins. Agreement evaluation between BD FACSCanto II system versus BD FACSCalibur system was tested on fresh peripheral blood, freeze-thawed apheresis, fresh bone marrow and fresh cord blood samples. Instrument's measure and staining repeatability clearly evidenced acceptable variability on the different samples tested. Intra- and inter-laboratory CV in CD34+ cell absolute count are consistent and reproducible. Linearity analysis, established between 2 and 329 cells/μl showed a linear relation between expected counts and measured counts (R2=0.97). Linear regression and Bland-Altman representations showed an excellent correlation on samples from different sources between the two systems and allowed the transfer of routine analysis from BD FACSCalibur to BD FACSCanto II. The BD SCE kit provides an accurate measure of the CD34 HSC, and can be used in daily routine to optimize the enumeration of hematopoietic CD34+ stem cells by flow cytometry. Moreover, the InterQC system seems to be a very useful tool for laboratory daily quality monitoring and thus for accreditation.

  8. Practical continuous-variable quantum key distribution without finite sampling bandwidth effects.

    PubMed

    Li, Huasheng; Wang, Chao; Huang, Peng; Huang, Duan; Wang, Tao; Zeng, Guihua

    2016-09-05

    In a practical continuous-variable quantum key distribution system, finite sampling bandwidth of the employed analog-to-digital converter at the receiver's side may lead to inaccurate results of pulse peak sampling. Then, errors in the parameters estimation resulted. Subsequently, the system performance decreases and security loopholes are exposed to eavesdroppers. In this paper, we propose a novel data acquisition scheme which consists of two parts, i.e., a dynamic delay adjusting module and a statistical power feedback-control algorithm. The proposed scheme may improve dramatically the data acquisition precision of pulse peak sampling and remove the finite sampling bandwidth effects. Moreover, the optimal peak sampling position of a pulse signal can be dynamically calibrated through monitoring the change of the statistical power of the sampled data in the proposed scheme. This helps to resist against some practical attacks, such as the well-known local oscillator calibration attack.

  9. Variation of mercury in fish from Massachusetts lakes based on ecoregion and lake trophic status

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rose, J.; Hutcheson, M.; West, C.R.

    1995-12-31

    Twenty-four of the state`s least-impacted waterbodies were sampled for sediment, water, physical characteristics and 3 species of fish to determine the extent of, and patterns of variation in, mercury contamination. Sampling effort was apportioned among three different ecological subregions of the state, as defined by EPA, and among lakes of differing trophic status. The authors sought to partition the variance to discover if these broadly defined concepts are suitable predictors of mercury levels in fish. Mean fish mercury was 0.14 ppm wet weight in samples of 168 of the bottom feeding brown bullheads (Ameriurus nebulosus) (range = 0.01--0.79 ppm); 0.3more » ppm in 199 of the omnivorous yellow perch (Perca flavescens) (range = 0.01--0.75 ppm); and 0.4 ppm in samples of 152 of the predaceous largemouth bass (Micropterus salmoides) (range = 0.05--1.1 ppm). Multivariate statistics are employed to determine how mercury concentrations in fish correlate with sediment chemistry, water chemistry, fish trophic status, fish size and age, lake and watershed size, the presence and extent of wetlands in the watershed, and physical characteristics of the lake. The survey design complements ongoing efforts begun in 1983 to test fish in a variety of waters, from which emanated fish advisories for impacted rivers and lakes. The study defines a baseline for fish contamination in Massachusetts lakes and ponds that serves as a template for public health decisions regarding fish consumption.« less

  10. A statistical evaluation of spectral fingerprinting methods using analysis of variance and principal component analysis

    USDA-ARS?s Scientific Manuscript database

    Six methods were compared with respect to spectral fingerprinting of a well-characterized series of broccoli samples. Spectral fingerprints were acquired for finely-powdered solid samples using Fourier transform-infrared (IR) and Fourier transform-near infrared (NIR) spectrometry and for aqueous met...

  11. Comparing identified and statistically significant lipids and polar metabolites in 15-year old serum and dried blood spot samples for longitudinal studies: Comparing lipids and metabolites in serum and DBS samples

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kyle, Jennifer E.; Casey, Cameron P.; Stratton, Kelly G.

    The use of dried blood spots (DBS) has many advantages over traditional plasma and serum samples such as smaller blood volume required, storage at room temperature, and ability for sampling in remote locations. However, understanding the robustness of different analytes in DBS samples is essential, especially in older samples collected for longitudinal studies. Here we analyzed DBS samples collected in 2000-2001 and stored at room temperature and compared them to matched serum samples stored at -80°C to determine if they could be effectively used as specific time points in a longitudinal study following metabolic disease. Four hundred small molecules weremore » identified in both the serum and DBS samples using gas chromatograph-mass spectrometry (GC-MS), liquid chromatography-MS (LC-MS) and LC-ion mobility spectrometry-MS (LC-IMS-MS). The identified polar metabolites overlapped well between the sample types, though only one statistically significant polar metabolite in a case-control study was conserved, indicating degradation occurs in the DBS samples affecting quantitation. Differences in the lipid identifications indicated that some oxidation occurs in the DBS samples. However, thirty-six statistically significant lipids correlated in both sample types indicating that lipid quantitation was more stable across the sample types.« less

  12. Multiscale pore structure and constitutive models of fine-grained rocks

    NASA Astrophysics Data System (ADS)

    Heath, J. E.; Dewers, T. A.; Shields, E. A.; Yoon, H.; Milliken, K. L.

    2017-12-01

    A foundational concept of continuum poromechanics is the representative elementary volume or REV: an amount of material large enough that pore- or grain-scale fluctuations in relevant properties are dissipated to a definable mean, but smaller than length scales of heterogeneity. We determine 2D-equivalent representative elementary areas (REAs) of pore areal fraction of three major types of mudrocks by applying multi-beam scanning electron microscopy (mSEM) to obtain terapixel image mosaics. Image analysis obtains pore areal fraction and pore size and shape as a function of progressively larger measurement areas. Using backscattering imaging and mSEM data, pores are identified by the components within which they occur, such as in organics or the clastic matrix. We correlate pore areal fraction with nano-indentation, micropillar compression, and axysimmetic testing at multiple length scales on a terrigenous-argillaceous mudrock sample. The combined data set is used to: investigate representative elementary volumes (and areas for the 2D images); determine if scale separation occurs; and determine if transport and mechanical properties at a given length scale can be statistically defined. Clear scale separation occurs between REAs and observable heterogeneity in two of the samples. A highly-laminated sample exhibits fine-scale heterogeneity and an overlapping in scales, in which case typical continuum assumptions on statistical variability may break down. Sandia National Laboratories is a multimission laboratory managed and operated by National Technology and Engineering Solutions of Sandia LLC, a wholly owned subsidiary of Honeywell International Inc. for the U.S. Department of Energy's National Nuclear Security Administration under contract DE-NA0003525.

  13. Statistical Analyses of Brain Surfaces Using Gaussian Random Fields on 2-D Manifolds

    PubMed Central

    Staib, Lawrence H.; Xu, Dongrong; Zhu, Hongtu; Peterson, Bradley S.

    2008-01-01

    Interest in the morphometric analysis of the brain and its subregions has recently intensified because growth or degeneration of the brain in health or illness affects not only the volume but also the shape of cortical and subcortical brain regions, and new image processing techniques permit detection of small and highly localized perturbations in shape or localized volume, with remarkable precision. An appropriate statistical representation of the shape of a brain region is essential, however, for detecting, localizing, and interpreting variability in its surface contour and for identifying differences in volume of the underlying tissue that produce that variability across individuals and groups of individuals. Our statistical representation of the shape of a brain region is defined by a reference region for that region and by a Gaussian random field (GRF) that is defined across the entire surface of the region. We first select a reference region from a set of segmented brain images of healthy individuals. The GRF is then estimated as the signed Euclidean distances between points on the surface of the reference region and the corresponding points on the corresponding region in images of brains that have been coregistered to the reference. Correspondences between points on these surfaces are defined through deformations of each region of a brain into the coordinate space of the reference region using the principles of fluid dynamics. The warped, coregistered region of each subject is then unwarped into its native space, simultaneously bringing into that space the map of corresponding points that was established when the surfaces of the subject and reference regions were tightly coregistered. The proposed statistical description of the shape of surface contours makes no assumptions, other than smoothness, about the shape of the region or its GRF. The description also allows for the detection and localization of statistically significant differences in the shapes of the surfaces across groups of subjects at both a fine and coarse scale. We demonstrate the effectiveness of these statistical methods by applying them to study differences in shape of the amygdala and hippocampus in a large sample of normal subjects and in subjects with attention deficit/hyperactivity disorder (ADHD). PMID:17243583

  14. Correlating tephras and cryptotephras using glass compositional analyses and numerical and statistical methods: Review and evaluation

    NASA Astrophysics Data System (ADS)

    Lowe, David J.; Pearce, Nicholas J. G.; Jorgensen, Murray A.; Kuehn, Stephen C.; Tryon, Christian A.; Hayward, Chris L.

    2017-11-01

    We define tephras and cryptotephras and their components (mainly ash-sized particles of glass ± crystals in distal deposits) and summarize the basis of tephrochronology as a chronostratigraphic correlational and dating tool for palaeoenvironmental, geological, and archaeological research. We then document and appraise recent advances in analytical methods used to determine the major, minor, and trace elements of individual glass shards from tephra or cryptotephra deposits to aid their correlation and application. Protocols developed recently for the electron probe microanalysis of major elements in individual glass shards help to improve data quality and standardize reporting procedures. A narrow electron beam (diameter ∼3-5 μm) can now be used to analyze smaller glass shards than previously attainable. Reliable analyses of 'microshards' (defined here as glass shards <32 μm in diameter) using narrow beams are useful for fine-grained samples from distal or ultra-distal geographic locations, and for vesicular or microlite-rich glass shards or small melt inclusions. Caveats apply, however, in the microprobe analysis of very small microshards (≤∼5 μm in diameter), where particle geometry becomes important, and of microlite-rich glass shards where the potential problem of secondary fluorescence across phase boundaries needs to be recognised. Trace element analyses of individual glass shards using laser ablation inductively coupled plasma-mass spectrometry (LA-ICP-MS), with crater diameters of 20 μm and 10 μm, are now effectively routine, giving detection limits well below 1 ppm. Smaller ablation craters (<10 μm) can be subject to significant element fractionation during analysis, but the systematic relationship of such fractionation with glass composition suggests that analyses for some elements at these resolutions may be quantifiable. In undertaking analyses, either by microprobe or LA-ICP-MS, reference material data acquired using the same procedure, and preferably from the same analytical session, should be presented alongside new analytical data. In part 2 of the review, we describe, critically assess, and recommend ways in which tephras or cryptotephras can be correlated (in conjunction with other information) using numerical or statistical analyses of compositional data. Statistical methods provide a less subjective means of dealing with analytical data pertaining to tephra components (usually glass or crystals/phenocrysts) than heuristic alternatives. They enable a better understanding of relationships among the data from multiple viewpoints to be developed and help quantify the degree of uncertainty in establishing correlations. In common with other scientific hypothesis testing, it is easier to infer using such analysis that two or more tephras are different rather than the same. Adding stratigraphic, chronological, spatial, or palaeoenvironmental data (i.e. multiple criteria) is usually necessary and allows for more robust correlations to be made. A two-stage approach is useful, the first focussed on differences in the mean composition of samples, or their range, which can be visualised graphically via scatterplot matrices or bivariate plots coupled with the use of statistical tools such as distance measures, similarity coefficients, hierarchical cluster analysis (informed by distance measures or similarity or cophenetic coefficients), and principal components analysis (PCA). Some statistical methods (cluster analysis, discriminant analysis) are referred to as 'machine learning' in the computing literature. The second stage examines sample variance and the degree of compositional similarity so that sample equivalence or otherwise can be established on a statistical basis. This stage may involve discriminant function analysis (DFA), support vector machines (SVMs), canonical variates analysis (CVA), and ANOVA or MANOVA (or its two-sample special case, the Hotelling two-sample T2 test). Randomization tests can be used where distributional assumptions such as multivariate normality underlying parametric tests are doubtful. Compositional data may be transformed and scaled before being subjected to multivariate statistical procedures including calculation of distance matrices, hierarchical cluster analysis, and PCA. Such transformations may make the assumption of multivariate normality more appropriate. A sequential procedure using Mahalanobis distance and the Hotelling two-sample T2 test is illustrated using glass major element data from trachytic to phonolitic Kenyan tephras. All these methods require a broad range of high-quality compositional data which can be used to compare 'unknowns' with reference (training) sets that are sufficiently complete to account for all possible correlatives, including tephras with heterogeneous glasses that contain multiple compositional groups. Currently, incomplete databases are tending to limit correlation efficacy. The development of an open, online global database to facilitate progress towards integrated, high-quality tephrostratigraphic frameworks for different regions is encouraged.

  15. Barefoot running does not affect simple reaction time: an exploratory study

    PubMed Central

    Snow, Nicholas J.; Blair, Jason F.L.; MacDonald, Graham Z.

    2018-01-01

    Background Converging evidence comparing barefoot (BF) and shod (SH) running highlights differences in foot-strike patterns and somatosensory feedback, among others. Anecdotal evidence from SH runners attempting BF running suggests a greater attentional demand may be experienced during BF running. However, little work to date has examined whether there is an attentional cost of BF versus SH running. Objective This exploratory study aimed to examine whether an acute bout of BF running would impact simple reaction time (SRT) compared to SH running, in a sample of runners naïve to BF running. Methods Eight male distance runners completed SRT testing during 10 min of BF or SH treadmill running at 70% maximal aerobic speed (17.9 ± 1.4 km h−1). To test SRT, participants were required to press a hand-held button in response to the flash of a light bulb placed in the center of their visual field. SRT was tested at 1-minute intervals during running. BF and SH conditions were completed in a pseudo-randomized and counterbalanced crossover fashion. SRT was defined as the time elapsed between the light bulb flash and the button press. SRT errors were also recorded and were defined as the number of trials in which a button press was not recorded in response to the light bulb flash. Results Overall, SRT later in the exercise bouts showed a statistically significant increase compared to earlier (p < 0.05). Statistically significant increases in SRT were present at 7 min versus 5 min (0.29 ± 0.02 s vs. 0.27 ± 0.02 s, p < 0.05) and at 9 min versus 2 min (0.29 ± 0.03 s vs. 0.27 ± 0.03 s, p < 0.05). However, BF running did not influence this increase in SRT (p > 0.05) or the number of SRT errors (17.6 ± 6.6 trials vs. 17.0 ± 13.0 trials, p > 0.05). Discussion In a sample of distance runners naïve to BF running, there was no statistically significant difference in SRT or SRT errors during acute bouts of BF and SH running. We interpret these results to mean that BF running does not have a greater attentional cost compared to SH running during a SRT task throughout treadmill running. Literature suggests that stride-to-stride gait modulation during running may occur predominately via mechanisms that preclude conscious perception, thus potentially attenuating effects of increased somatosensory feedback experienced during BF running. Future research should explore the present experimental paradigm in a larger sample using over-ground running trials, as well as employing different tests of attention. PMID:29666760

  16. Certain and possible rules for decision making using rough set theory extended to fuzzy sets

    NASA Technical Reports Server (NTRS)

    Dekorvin, Andre; Shipley, Margaret F.

    1993-01-01

    Uncertainty may be caused by the ambiguity in the terms used to describe a specific situation. It may also be caused by skepticism of rules used to describe a course of action or by missing and/or erroneous data. To deal with uncertainty, techniques other than classical logic need to be developed. Although, statistics may be the best tool available for handling likelihood, it is not always adequate for dealing with knowledge acquisition under uncertainty. Inadequacies caused by estimating probabilities in statistical processes can be alleviated through use of the Dempster-Shafer theory of evidence. Fuzzy set theory is another tool used to deal with uncertainty where ambiguous terms are present. Other methods include rough sets, the theory of endorsements and nonmonotonic logic. J. Grzymala-Busse has defined the concept of lower and upper approximation of a (crisp) set and has used that concept to extract rules from a set of examples. We will define the fuzzy analogs of lower and upper approximations and use these to obtain certain and possible rules from a set of examples where the data is fuzzy. Central to these concepts will be the idea of the degree to which a fuzzy set A is contained in another fuzzy set B, and the degree of intersection of a set A with set B. These concepts will also give meaning to the statement; A implies B. The two meanings will be: (1) if x is certainly in A then it is certainly in B, and (2) if x is possibly in A then it is possibly in B. Next, classification will be looked at and it will be shown that if a classification will be looked at and it will be shown that if a classification is well externally definable then it is well internally definable, and if it is poorly externally definable then it is poorly internally definable, thus generalizing a result of Grzymala-Busse. Finally, some ideas of how to define consensus and group options to form clusters of rules will be given.

  17. Statistical analyses on sandstones: Systematic approach for predicting petrographical and petrophysical properties

    NASA Astrophysics Data System (ADS)

    Stück, H. L.; Siegesmund, S.

    2012-04-01

    Sandstones are a popular natural stone due to their wide occurrence and availability. The different applications for these stones have led to an increase in demand. From the viewpoint of conservation and the natural stone industry, an understanding of the material behaviour of this construction material is very important. Sandstones are a highly heterogeneous material. Based on statistical analyses with a sufficiently large dataset, a systematic approach to predicting the material behaviour should be possible. Since the literature already contains a large volume of data concerning the petrographical and petrophysical properties of sandstones, a large dataset could be compiled for the statistical analyses. The aim of this study is to develop constraints on the material behaviour and especially on the weathering behaviour of sandstones. Approximately 300 samples from historical and presently mined natural sandstones in Germany and ones described worldwide were included in the statistical approach. The mineralogical composition and fabric characteristics were determined from detailed thin section analyses and descriptions in the literature. Particular attention was paid to evaluating the compositional and textural maturity, grain contact respectively contact thickness, type of cement, degree of alteration and the intergranular volume. Statistical methods were used to test for normal distributions and calculating the linear regression of the basic petrophysical properties of density, porosity, water uptake as well as the strength. The sandstones were classified into three different pore size distributions and evaluated with the other petrophysical properties. Weathering behavior like hygric swelling and salt loading tests were also included. To identify similarities between individual sandstones or to define groups of specific sandstone types, principle component analysis, cluster analysis and factor analysis were applied. Our results show that composition and porosity evolution during diagenesis is a very important control on the petrophysical properties of a building stone. The relationship between intergranular volume, cementation and grain contact, can also provide valuable information to predict the strength properties. Since the samples investigated mainly originate from the Triassic German epicontinental basin, arkoses and feldspar-arenites are underrepresented. In general, the sandstones can be grouped as follows: i) quartzites, highly mature with a primary porosity of about 40%, ii) quartzites, highly mature, showing a primary porosity of 40% but with early clay infiltration, iii) sublitharenites-lithic arenites exhibiting a lower primary porosity, higher cementation with quartz and Fe-oxides ferritic and iv) sublitharenites-lithic arenites with a higher content of pseudomatrix. However, in the last two groups the feldspar and lithoclasts can also show considerable alteration. All sandstone groups differ with respect to the pore space and strength data, as well as water uptake properties, which were obtained by linear regression analysis. Similar petrophysical properties are discernible for each type when using principle component analysis. Furthermore, strength as well as the porosity of sandstones shows distinct differences considering their stratigraphic ages and the compositions. The relationship between porosity, strength as well as salt resistance could also be verified. Hygric swelling shows an interrelation to pore size type, porosity and strength but also to the degree of alteration (e.g. lithoclasts, pseudomatrix). To summarize, the different regression analyses and the calculated confidence regions provide a significant tool to classify the petrographical and petrophysical parameters of sandstones. Based on this, the durability and the weathering behavior of the sandstone groups can be constrained. Keywords: sandstones, petrographical & petrophysical properties, predictive approach, statistical investigation

  18. Recent advances of mesoporous materials in sample preparation.

    PubMed

    Zhao, Liang; Qin, Hongqiang; Wu, Ren'an; Zou, Hanfa

    2012-03-09

    Sample preparation has been playing an important role in the analysis of complex samples. Mesoporous materials as the promising adsorbents have gained increasing research interest in sample preparation due to their desirable characteristics of high surface area, large pore volume, tunable mesoporous channels with well defined pore-size distribution, controllable wall composition, as well as modifiable surface properties. The aim of this paper is to review the recent advances of mesoporous materials in sample preparation with emphases on extraction of metal ions, adsorption of organic compounds, size selective enrichment of peptides/proteins, specific capture of post-translational peptides/proteins and enzymatic reactor for protein digestion. Copyright © 2011 Elsevier B.V. All rights reserved.

  19. Understanding sexual orientation and health in Canada: Who are we capturing and who are we missing using the Statistics Canada sexual orientation question?

    PubMed

    Dharma, Christoffer; Bauer, Greta R

    2017-04-20

    Public health research on inequalities in Canada depends heavily on population data sets such as the Canadian Community Health Survey. While sexual orientation has three dimensions - identity, behaviour and attraction - Statistics Canada and public health agencies assess sexual orientation with a single questionnaire item on identity, defined behaviourally. This study aims to evaluate this item, to allow for clearer interpretation of sexual orientation frequencies and inequalities. Through an online convenience sampling of Canadians ≥14 years of age, participants (n = 311) completed the Statistics Canada question and a second set of sexual orientation questions. The single-item question had an 85.8% sensitivity in capturing sexual minorities, broadly defined by their sexual identity, lifetime behaviour and attraction. Kappa statistic for agreement between the single item and sexual identity was 0.89; with past year, lifetime behaviour and attraction were 0.39, 0.48 and 0.57 respectively. The item captured 99.3% of those with a sexual minority identity, 84.2% of those with any lifetime same-sex partners, 98.4% with a past-year same-sex partner, and 97.8% who indicated at least equal attraction to same-sex persons. Findings from Statistics Canada surveys can be best interpreted as applying to those who identify as sexual minorities. Analyses using this measure will underidentify those with same-sex partners or attractions who do not identify as a sexual minority, and should be interpreted accordingly. To understand patterns of sexual minority health in Canada, there is a need to incorporate other dimensions of sexual orientation.

  20. Sand sources and transport pathways for the San Francisco Bay coastal system, based on X-ray diffraction mineralogy

    USGS Publications Warehouse

    Hein, James R.; Mizell, Kira; Barnard, Patrick L.; Barnard, P.L.; Jaffee, B.E.; Schoellhamer, D.H.

    2013-01-01

    The mineralogical compositions of 119 samples collected from throughout the San Francisco Bay coastal system, including bayfloor and seafloor, area beaches, cliff outcrops, and major drainages, were determined using X-ray diffraction (XRD). Comparison of the mineral concentrations and application of statistical cluster analysis of XRD spectra allowed for the determination of provenances and transport pathways. The use of XRD mineral identifications provides semi-quantitative compositions needed for comparisons of beach and offshore sands with potential cliff and river sources, but the innovative cluster analysis of XRD diffraction spectra provides a unique visualization of how groups of samples within the San Francisco Bay coastal system are related so that sand-sized sediment transport pathways can be inferred. The main vector for sediment transport as defined by the XRD analysis is from San Francisco Bay to the outer coast, where the sand then accumulates on the ebb tidal delta and also moves alongshore. This mineralogical link defines a critical pathway because large volumes of sediment have been removed from the Bay over the last century via channel dredging, aggregate mining, and borrow pit mining, with comparable volumes of erosion from the ebb tidal delta over the same period, in addition to high rates of shoreline retreat along the adjacent, open-coast beaches. Therefore, while previously only a temporal relationship was established, the transport pathway defined by mineralogical and geochemical tracers support the link between anthropogenic activities in the Bay and widespread erosion outside the Bay. The XRD results also establish the regional and local importance of sediment derived from cliff erosion, as well as both proximal and distal fluvial sources. This research is an important contribution to a broader provenance study aimed at identifying the driving forces for widespread geomorphic change in a heavily urbanized coastal-estuarine system.

  1. Statistical methods for investigating quiescence and other temporal seismicity patterns

    USGS Publications Warehouse

    Matthews, M.V.; Reasenberg, P.A.

    1988-01-01

    We propose a statistical model and a technique for objective recognition of one of the most commonly cited seismicity patterns:microearthquake quiescence. We use a Poisson process model for seismicity and define a process with quiescence as one with a particular type of piece-wise constant intensity function. From this model, we derive a statistic for testing stationarity against a 'quiescence' alternative. The large-sample null distribution of this statistic is approximated from simulated distributions of appropriate functionals applied to Brownian bridge processes. We point out the restrictiveness of the particular model we propose and of the quiescence idea in general. The fact that there are many point processes which have neither constant nor quiescent rate functions underscores the need to test for and describe nonuniformity thoroughly. We advocate the use of the quiescence test in conjunction with various other tests for nonuniformity and with graphical methods such as density estimation. ideally these methods may promote accurate description of temporal seismicity distributions and useful characterizations of interesting patterns. ?? 1988 Birkha??user Verlag.

  2. The Development of Statistics Textbook Supported with ICT and Portfolio-Based Assessment

    NASA Astrophysics Data System (ADS)

    Hendikawati, Putriaji; Yuni Arini, Florentina

    2016-02-01

    This research was development research that aimed to develop and produce a Statistics textbook model that supported with information and communication technology (ICT) and Portfolio-Based Assessment. This book was designed for students of mathematics at the college to improve students’ ability in mathematical connection and communication. There were three stages in this research i.e. define, design, and develop. The textbooks consisted of 10 chapters which each chapter contains introduction, core materials and include examples and exercises. The textbook developed phase begins with the early stages of designed the book (draft 1) which then validated by experts. Revision of draft 1 produced draft 2 which then limited test for readability test book. Furthermore, revision of draft 2 produced textbook draft 3 which simulated on a small sample to produce a valid model textbook. The data were analysed with descriptive statistics. The analysis showed that the Statistics textbook model that supported with ICT and Portfolio-Based Assessment valid and fill up the criteria of practicality.

  3. A global fit of the MSSM with GAMBIT

    NASA Astrophysics Data System (ADS)

    Athron, Peter; Balázs, Csaba; Bringmann, Torsten; Buckley, Andy; Chrząszcz, Marcin; Conrad, Jan; Cornell, Jonathan M.; Dal, Lars A.; Edsjö, Joakim; Farmer, Ben; Jackson, Paul; Krislock, Abram; Kvellestad, Anders; Mahmoudi, Farvah; Martinez, Gregory D.; Putze, Antje; Raklev, Are; Rogan, Christopher; Saavedra, Aldo; Savage, Christopher; Scott, Pat; Serra, Nicola; Weniger, Christoph; White, Martin

    2017-12-01

    We study the seven-dimensional Minimal Supersymmetric Standard Model (MSSM7) with the new GAMBIT software framework, with all parameters defined at the weak scale. Our analysis significantly extends previous weak-scale, phenomenological MSSM fits, by adding more and newer experimental analyses, improving the accuracy and detail of theoretical predictions, including dominant uncertainties from the Standard Model, the Galactic dark matter halo and the quark content of the nucleon, and employing novel and highly-efficient statistical sampling methods to scan the parameter space. We find regions of the MSSM7 that exhibit co-annihilation of neutralinos with charginos, stops and sbottoms, as well as models that undergo resonant annihilation via both light and heavy Higgs funnels. We find high-likelihood models with light charginos, stops and sbottoms that have the potential to be within the future reach of the LHC. Large parts of our preferred parameter regions will also be accessible to the next generation of direct and indirect dark matter searches, making prospects for discovery in the near future rather good.

  4. Critical issues in ALS case-control studies: the case of the Euro-MOTOR study.

    PubMed

    D'Ovidio, Fabrizio; Rooney, James P K; Visser, Anne E; Vermeulen, Roel C H; Veldink, Jan H; Van Den Berg, Leonard H; Hardiman, Orla; Logroscino, Giancarlo; Chiò, Adriano; Beghi, Ettore

    2017-08-01

    Backround: Political and sociocultural differences between countries can affect the outcome of clinical and epidemiological studies in ALS. Cross-national studies represent the ideal process by which risk factors can be assessed using the same methodology in different geographical areas. A survey of three European countries (The Netherlands, Ireland and Italy) has been conducted in which incident ALS patients and matched controls were recruited in a population-based study based on age, gender and area of residency, under the Euro-MOTOR systems biology programme of research. We have identified strengths and limitations during the trajectory of the Euro-MOTOR study, from the research design to data analysis. We have analysed the implications of factors including cross-national differences in healthcare systems, sample size, types of matching, the definition of exposures and statistical analysis. Addressing critical methodological aspects of the design of the Euro-MOTOR project minimises bias and will facilitate scientific assessment of the independent role of well-defined exposures.

  5. Exploring Human Cognition Using Large Image Databases.

    PubMed

    Griffiths, Thomas L; Abbott, Joshua T; Hsu, Anne S

    2016-07-01

    Most cognitive psychology experiments evaluate models of human cognition using a relatively small, well-controlled set of stimuli. This approach stands in contrast to current work in neuroscience, perception, and computer vision, which have begun to focus on using large databases of natural images. We argue that natural images provide a powerful tool for characterizing the statistical environment in which people operate, for better evaluating psychological theories, and for bringing the insights of cognitive science closer to real applications. We discuss how some of the challenges of using natural images as stimuli in experiments can be addressed through increased sample sizes, using representations from computer vision, and developing new experimental methods. Finally, we illustrate these points by summarizing recent work using large image databases to explore questions about human cognition in four different domains: modeling subjective randomness, defining a quantitative measure of representativeness, identifying prior knowledge used in word learning, and determining the structure of natural categories. Copyright © 2016 Cognitive Science Society, Inc.

  6. Discovery of four recessive developmental disorders using probabilistic genotype and phenotype matching among 4,125 families

    PubMed Central

    Ansari, Morad; Balasubramanian, Meena; Blyth, Moira; Brady, Angela F.; Clayton, Stephen; Cole, Trevor; Deshpande, Charu; Fitzgerald, Tomas W.; Foulds, Nicola; Francis, Richard; Gabriel, George; Gerety, Sebastian S.; Goodship, Judith; Hobson, Emma; Jones, Wendy D.; Joss, Shelagh; King, Daniel; Klena, Nikolai; Kumar, Ajith; Lees, Melissa; Lelliott, Chris; Lord, Jenny; McMullan, Dominic; O'Regan, Mary; Osio, Deborah; Piombo, Virginia; Prigmore, Elena; Rajan, Diana; Rosser, Elisabeth; Sifrim, Alejandro; Smith, Audrey; Swaminathan, Ganesh J.; Turnpenny, Peter; Whitworth, James; Wright, Caroline F.; Firth, Helen V.; Barrett, Jeffrey C.; Lo, Cecilia W.; FitzPatrick, David R.; Hurles, Matthew E.

    2018-01-01

    Discovery of most autosomal recessive disease genes has involved analysis of large, often consanguineous, multiplex families or small cohorts of unrelated individuals with a well-defined clinical condition. Discovery of novel dominant causes of rare, genetically heterogenous developmental disorders has been revolutionized by exome analysis of large cohorts of phenotypically diverse parent-offspring trios 1,2. Here we analysed 4,125 families with diverse, rare, genetically heterogeneous developmental disorders and identified four novel autosomal recessive disorders. These four disorders were identified by integrating Mendelian filtering (identifying probands with rare biallelic putatively damaging variants in the same gene) with statistical assessments of (i) the likelihood of sampling the observed genotypes from the general population, and (ii) the phenotypic similarity of patients with the same recessive candidate gene. This new paradigm promises to catalyse discovery of novel recessive disorders, especially those with less consistent or nonspecific clinical presentations, and those caused predominantly by compound heterozygous genotypes. PMID:26437029

  7. Discovery of four recessive developmental disorders using probabilistic genotype and phenotype matching among 4,125 families.

    PubMed

    Akawi, Nadia; McRae, Jeremy; Ansari, Morad; Balasubramanian, Meena; Blyth, Moira; Brady, Angela F; Clayton, Stephen; Cole, Trevor; Deshpande, Charu; Fitzgerald, Tomas W; Foulds, Nicola; Francis, Richard; Gabriel, George; Gerety, Sebastian S; Goodship, Judith; Hobson, Emma; Jones, Wendy D; Joss, Shelagh; King, Daniel; Klena, Nikolai; Kumar, Ajith; Lees, Melissa; Lelliott, Chris; Lord, Jenny; McMullan, Dominic; O'Regan, Mary; Osio, Deborah; Piombo, Virginia; Prigmore, Elena; Rajan, Diana; Rosser, Elisabeth; Sifrim, Alejandro; Smith, Audrey; Swaminathan, Ganesh J; Turnpenny, Peter; Whitworth, James; Wright, Caroline F; Firth, Helen V; Barrett, Jeffrey C; Lo, Cecilia W; FitzPatrick, David R; Hurles, Matthew E

    2015-11-01

    Discovery of most autosomal recessive disease-associated genes has involved analysis of large, often consanguineous multiplex families or small cohorts of unrelated individuals with a well-defined clinical condition. Discovery of new dominant causes of rare, genetically heterogeneous developmental disorders has been revolutionized by exome analysis of large cohorts of phenotypically diverse parent-offspring trios. Here we analyzed 4,125 families with diverse, rare and genetically heterogeneous developmental disorders and identified four new autosomal recessive disorders. These four disorders were identified by integrating Mendelian filtering (selecting probands with rare, biallelic and putatively damaging variants in the same gene) with statistical assessments of (i) the likelihood of sampling the observed genotypes from the general population and (ii) the phenotypic similarity of patients with recessive variants in the same candidate gene. This new paradigm promises to catalyze the discovery of novel recessive disorders, especially those with less consistent or nonspecific clinical presentations and those caused predominantly by compound heterozygous genotypes.

  8. Probability Distributions for Random Quantum Operations

    NASA Astrophysics Data System (ADS)

    Schultz, Kevin

    Motivated by uncertainty quantification and inference of quantum information systems, in this work we draw connections between the notions of random quantum states and operations in quantum information with probability distributions commonly encountered in the field of orientation statistics. This approach identifies natural sample spaces and probability distributions upon these spaces that can be used in the analysis, simulation, and inference of quantum information systems. The theory of exponential families on Stiefel manifolds provides the appropriate generalization to the classical case. Furthermore, this viewpoint motivates a number of additional questions into the convex geometry of quantum operations relative to both the differential geometry of Stiefel manifolds as well as the information geometry of exponential families defined upon them. In particular, we draw on results from convex geometry to characterize which quantum operations can be represented as the average of a random quantum operation. This project was supported by the Intelligence Advanced Research Projects Activity via Department of Interior National Business Center Contract Number 2012-12050800010.

  9. 40 CFR 257.22 - Ground-water monitoring systems.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... defined in § 257.5(b)) that: (1) Represent the quality of background ground water that has not been affected by leakage from a unit. A determination of background quality may include sampling of wells that... at other wells will provide an indication of background ground-water quality that is as...

  10. 40 CFR 257.22 - Ground-water monitoring systems.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... aquifer (as defined in § 257.5(b)) that: (1) Represent the quality of background ground water that has not been affected by leakage from a unit. A determination of background quality may include sampling of...) Sampling at other wells will provide an indication of background ground-water quality that is as...

  11. 40 CFR 257.22 - Ground-water monitoring systems.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... aquifer (as defined in § 257.5(b)) that: (1) Represent the quality of background ground water that has not been affected by leakage from a unit. A determination of background quality may include sampling of...) Sampling at other wells will provide an indication of background ground-water quality that is as...

  12. Sioux City Riverbank Filtration Study

    NASA Astrophysics Data System (ADS)

    Mach, R.; Condon, J.; Johnson, J.

    2003-04-01

    The City of Sioux City (City) obtains a large percentage of their drinking water supply from both a horizontal collector well system and vertical wells located adjacent to the Missouri River. These wells are set in either the Missouri Alluvium or the Dakota Sandstone aquifer. Several of the collector well laterals extend out beneath the Missouri River, with the laterals being over twenty feet below the river channel bottom. Due to concerns regarding ground water under direct surface water influence, the Iowa Department of Natural Resources (IDNR) required the City to expand their water treatment process to deal with potential surface water contaminant issues. With the extensive cost of these plant upgrades, the City and Olsson Associates (OA) approached the IDNR requesting approval for assessing the degree of natural riverbank filtration for water treatment. If this natural process could be ascertained, the level of treatment from the plant could be reduced. The objective of this study was to quantify the degree of surface water (i.e. Missouri River) filtration due to the underlying Missouri River sediments. Several series of microscopic particulate analysis where conducted, along with tracking of turbidity, temperature, bacteria and a full scale particle count study. Six particle sizes from six sampling points were assessed over a nine-month period that spanned summer, fall and spring weather periods. The project was set up in two phases and utilized industry accepted statistical analyses to identify particle data trends. The first phase consisted of twice daily sample collection from the Missouri River and the collector well system for a one-month period. Statistical analysis of the data indicated reducing the sampling frequency and sampling locations would yield justifiable data while significantly reducing sampling and analysis costs. The IDNR approved this modification, and phase II included sampling and analysis under this reduced plant for an eight-month period. Final statistical analyses of the nine months of data indicate up to a four-log particle reduction occurs through river bank filtration. Consequently, Missouri River sediments within the City's well field are very effective in water filtration. This information was submitted to the IDNR for review and approval. Subsequently, the IDNR approved 4.0 log removal for Giardia and 3.5 log removal for Cryptosporidium through the riverbank and treatment plant. The City and IDNR have agreed on subrogate parameters for monitoring purposes.

  13. [Generalization of the results of clinical studies through the analysis of subgroups].

    PubMed

    Costa, João; Fareleira, Filipa; Ascensão, Raquel; Vaz Carneiro, António

    2012-01-01

    Subgroup analysis in clinical trials are usually performed to define the potential heterogeneity of treatment effect in relation with the baseline risk, physiopathology, practical application of therapy or the under-utilization in clinical practice of effective interventions due to uncertainties of its benefit/risk ratio. When appropriately planned, subgroup analysis are a valid methodology the define benefits in subgroups of patients, thus providing good quality evidence to support clinical decision making. However, in order to be correct, subgroup analysis should be defined a priori, done in small numbers, should be fully reported and, most important, must endure statistical tests for interaction. In this paper we present an example of the treatment of post-menopausal osteoporosis, in which the benefits of an intervention (the higher the fracture risk is, the better the benefit is) with a specific agent (bazedoxifene) was only disclosed after a post-hoc analysis of the initial global trial sample.

  14. CROSS-DISCIPLINARY PHYSICS AND RELATED AREAS OF SCIENCE AND TECHNOLOGY: Statistical interior properties of globular proteins

    NASA Astrophysics Data System (ADS)

    Jiang, Zhou-Ting; Zhang, Lin-Xi; Sun, Ting-Ting; Wu, Tai-Quan

    2009-10-01

    The character of forming long-range contacts affects the three-dimensional structure of globular proteins deeply. As the different ability to form long-range contacts between 20 types of amino acids and 4 categories of globular proteins, the statistical properties are thoroughly discussed in this paper. Two parameters NC and ND are defined to confine the valid residues in detail. The relationship between hydrophobicity scales and valid residue percentage of each amino acid is given in the present work and the linear functions are shown in our statistical results. It is concluded that the hydrophobicity scale defined by chemical derivatives of the amino acids and nonpolar phase of large unilamellar vesicle membranes is the most effective technique to characterise the hydrophobic behavior of amino acid residues. Meanwhile, residue percentage Pi and sequential residue length Li of a certain protein i are calculated under different conditions. The statistical results show that the average value of Pi as well as Li of all-α proteins has a minimum among these 4 classes of globular proteins, indicating that all-α proteins are hardly capable of forming long-range contacts one by one along their linear amino acid sequences. All-β proteins have a higher tendency to construct long-range contacts along their primary sequences related to the secondary configurations, i.e. parallel and anti-parallel configurations of β sheets. The investigation of the interior properties of globular proteins give us the connection between the three-dimensional structure and its primary sequence data or secondary configurations, and help us to understand the structure of protein and its folding process well.

  15. Sexual dimorphism in multiple aspects of 3D facial symmetry and asymmetry defined by spatially dense geometric morphometrics.

    PubMed

    Claes, Peter; Walters, Mark; Shriver, Mark D; Puts, David; Gibson, Greg; Clement, John; Baynam, Gareth; Verbeke, Geert; Vandermeulen, Dirk; Suetens, Paul

    2012-08-01

    Accurate measurement of facial sexual dimorphism is useful to understanding facial anatomy and specifically how faces influence, and have been influenced by, sexual selection. An important facial aspect is the display of bilateral symmetry, invoking the need to investigate aspects of symmetry and asymmetry separately when examining facial shape. Previous studies typically employed landmarks that provided only a sparse facial representation, where different landmark choices could lead to contrasting outcomes. Furthermore, sexual dimorphism is only tested as a difference of sample means, which is statistically the same as a difference in population location only. Within the framework of geometric morphometrics, we partition facial shape, represented in a spatially dense way, into patterns of symmetry and asymmetry, following a two-factor anova design. Subsequently, we investigate sexual dimorphism in symmetry and asymmetry patterns separately, and on multiple aspects, by examining (i) population location differences as well as differences in population variance-covariance; (ii) scale; and (iii) orientation. One important challenge in this approach is the proportionally high number of variables to observations necessitating the implementation of permutational and computationally feasible statistics. In a sample of gender-matched young adults (18-25 years) with self-reported European ancestry, we found greater variation in male faces than in women for all measurements. Statistically significant sexual dimorphism was found for the aspect of location in both symmetry and asymmetry (directional asymmetry), for the aspect of scale only in asymmetry (magnitude of fluctuating asymmetry) and, in contrast, for the aspect of orientation only in symmetry. Interesting interplays with hypotheses in evolutionary and developmental biology were observed, such as the selective nature of the force underpinning sexual dimorphism and the genetic independence of the structural patterns of fluctuating asymmetry. Additionally, insights into growth patterns of the soft tissue envelope of the face and underlying skull structure can also be obtained from the results. © 2012 The Authors. Journal of Anatomy © 2012 Anatomical Society.

  16. Sexual dimorphism in multiple aspects of 3D facial symmetry and asymmetry defined by spatially dense geometric morphometrics

    PubMed Central

    Claes, Peter; Walters, Mark; Shriver, Mark D; Puts, David; Gibson, Greg; Clement, John; Baynam, Gareth; Verbeke, Geert; Vandermeulen, Dirk; Suetens, Paul

    2012-01-01

    Accurate measurement of facial sexual dimorphism is useful to understanding facial anatomy and specifically how faces influence, and have been influenced by, sexual selection. An important facial aspect is the display of bilateral symmetry, invoking the need to investigate aspects of symmetry and asymmetry separately when examining facial shape. Previous studies typically employed landmarks that provided only a sparse facial representation, where different landmark choices could lead to contrasting outcomes. Furthermore, sexual dimorphism is only tested as a difference of sample means, which is statistically the same as a difference in population location only. Within the framework of geometric morphometrics, we partition facial shape, represented in a spatially dense way, into patterns of symmetry and asymmetry, following a two-factor anova design. Subsequently, we investigate sexual dimorphism in symmetry and asymmetry patterns separately, and on multiple aspects, by examining (i) population location differences as well as differences in population variance-covariance; (ii) scale; and (iii) orientation. One important challenge in this approach is the proportionally high number of variables to observations necessitating the implementation of permutational and computationally feasible statistics. In a sample of gender-matched young adults (18–25 years) with self-reported European ancestry, we found greater variation in male faces than in women for all measurements. Statistically significant sexual dimorphism was found for the aspect of location in both symmetry and asymmetry (directional asymmetry), for the aspect of scale only in asymmetry (magnitude of fluctuating asymmetry) and, in contrast, for the aspect of orientation only in symmetry. Interesting interplays with hypotheses in evolutionary and developmental biology were observed, such as the selective nature of the force underpinning sexual dimorphism and the genetic independence of the structural patterns of fluctuating asymmetry. Additionally, insights into growth patterns of the soft tissue envelope of the face and underlying skull structure can also be obtained from the results. PMID:22702244

  17. An Accurate Methodology to detect Leaching of Nickel and Chromium Ions in the Initial Phase of Orthodontic Treatment: An in vivo Study.

    PubMed

    Kumar, R Vinoth; Rajvikram, N; Rajakumar, P; Saravanan, R; Deepak, V Arun; Vijaykumar, V

    2016-03-01

    The aim of this study was to evaluate the release of nickel and chromium ions in human saliva during fixed orthodontic therapy. Ten patients with Angle's Class-I malocclusion with bimaxillary protrusion without any metal restorations or crowns and with all the permanent teeth were selected. Five male patients and five female patients in the age group range of 14 to 23 years were scheduled for orthodontic treatment with first premolar extraction. Saliva samples were collected in three stages: sample 1, before orthodontic treatment; sample 2, after 10 days of bonding sample; and sample 3, after 1 month of bonding. The samples were analyzed for the following metals nickel and chromium using inductively coupled plasma optical emission spectrometry (ICP-OES). The levels of nickel and chromium were statistically significant, while nickel showed a gradual increase in the first 10 days and a decline thereafter. Chromium showed a gradual increase and was statistically significant on the 30th day. There was greatest release of ions during the first 10 days and a gradual decline thereafter. Control group had traces of nickel and chromium. While comparing levels of nickel in saliva, there was a significant rise from baseline to 10th and 30th-day sample, which was statistically significant. While comparing 10th day to that of 30th day, there was no statistical significance. The levels of chromium ion in the saliva were more in 30th day, and when comparing 10th-day sample with 30th day, there was statistical significance. Nickel and chromium levels were well within the permissible levels. However, some hypersensitive individuals may be allergic to this minimal permissible level.

  18. Variance of discharge estimates sampled using acoustic Doppler current profilers from moving boats

    USGS Publications Warehouse

    Garcia, Carlos M.; Tarrab, Leticia; Oberg, Kevin; Szupiany, Ricardo; Cantero, Mariano I.

    2012-01-01

    This paper presents a model for quantifying the random errors (i.e., variance) of acoustic Doppler current profiler (ADCP) discharge measurements from moving boats for different sampling times. The model focuses on the random processes in the sampled flow field and has been developed using statistical methods currently available for uncertainty analysis of velocity time series. Analysis of field data collected using ADCP from moving boats from three natural rivers of varying sizes and flow conditions shows that, even though the estimate of the integral time scale of the actual turbulent flow field is larger than the sampling interval, the integral time scale of the sampled flow field is on the order of the sampling interval. Thus, an equation for computing the variance error in discharge measurements associated with different sampling times, assuming uncorrelated flow fields is appropriate. The approach is used to help define optimal sampling strategies by choosing the exposure time required for ADCPs to accurately measure flow discharge.

  19. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Woodroffe, J. R.; Brito, T. V.; Jordanova, V. K.

    In the standard practice of neutron multiplicity counting , the first three sampled factorial moments of the event triggered neutron count distribution were used to quantify the three main neutron source terms: the spontaneous fissile material effective mass, the relative (α,n) production and the induced fission source responsible for multiplication. Our study compares three methods to quantify the statistical uncertainty of the estimated mass: the bootstrap method, propagation of variance through moments, and statistical analysis of cycle data method. Each of the three methods was implemented on a set of four different NMC measurements, held at the JRC-laboratory in Ispra,more » Italy, sampling four different Pu samples in a standard Plutonium Scrap Multiplicity Counter (PSMC) well counter.« less

  20. Automated sampling assessment for molecular simulations using the effective sample size

    PubMed Central

    Zhang, Xin; Bhatt, Divesh; Zuckerman, Daniel M.

    2010-01-01

    To quantify the progress in the development of algorithms and forcefields used in molecular simulations, a general method for the assessment of the sampling quality is needed. Statistical mechanics principles suggest the populations of physical states characterize equilibrium sampling in a fundamental way. We therefore develop an approach for analyzing the variances in state populations, which quantifies the degree of sampling in terms of the effective sample size (ESS). The ESS estimates the number of statistically independent configurations contained in a simulated ensemble. The method is applicable to both traditional dynamics simulations as well as more modern (e.g., multi–canonical) approaches. Our procedure is tested in a variety of systems from toy models to atomistic protein simulations. We also introduce a simple automated procedure to obtain approximate physical states from dynamic trajectories: this allows sample–size estimation in systems for which physical states are not known in advance. PMID:21221418

  1. Empirical redefinition of comprehensive health and well-being in the older adults of the United States.

    PubMed

    McClintock, Martha K; Dale, William; Laumann, Edward O; Waite, Linda

    2016-05-31

    The World Health Organization (WHO) defines health as a "state of complete physical, mental and social well-being and not merely the absence of disease or infirmity." Despite general acceptance of this comprehensive definition, there has been little rigorous scientific attempt to use it to measure and assess population health. Instead, the dominant model of health is a disease-centered Medical Model (MM), which actively ignores many relevant domains. In contrast to the MM, we approach this issue through a Comprehensive Model (CM) of health consistent with the WHO definition, giving statistically equal consideration to multiple health domains, including medical, physical, psychological, functional, and sensory measures. We apply a data-driven latent class analysis (LCA) to model 54 specific health variables from the National Social Life, Health, and Aging Project (NSHAP), a nationally representative sample of US community-dwelling older adults. We first apply the LCA to the MM, identifying five health classes differentiated primarily by having diabetes and hypertension. The CM identifies a broader range of six health classes, including two "emergent" classes completely obscured by the MM. We find that specific medical diagnoses (cancer and hypertension) and health behaviors (smoking) are far less important than mental health (loneliness), sensory function (hearing), mobility, and bone fractures in defining vulnerable health classes. Although the MM places two-thirds of the US population into "robust health" classes, the CM reveals that one-half belong to less healthy classes, independently associated with higher mortality. This reconceptualization has important implications for medical care delivery, preventive health practices, and resource allocation.

  2. Sex-specific 99th percentiles derived from the AACC Universal Sample Bank for the Roche Gen 5 cTnT assay: Comorbidities and statistical methods influence derivation of reference limits.

    PubMed

    Gunsolus, Ian L; Jaffe, Allan S; Sexter, Anne; Schulz, Karen; Ler, Ranka; Lindgren, Brittany; Saenger, Amy K; Love, Sara A; Apple, Fred S

    2017-12-01

    Our purpose was to determine a) overall and sex-specific 99th percentile upper reference limits (URL) and b) influences of statistical methods and comorbidities on the URLs. Heparin plasma from 838 normal subjects (423 men, 415 women) were obtained from the AACC (Universal Sample Bank). The cobas e602 measured cTnT (Roche Gen 5 assay); limit of detection (LoD), 3ng/L. Hemoglobin A1c (URL 6.5%), NT-proBNP (URL 125ng/L) and eGFR (60mL/min/1.73m 2 ) were measured, along with identification of statin use, to better define normality. 99th percentile URLs were determined by the non-parametric (NP), Harrell-Davis Estimator (HDE) and Robust (R) methods. 355 men and 339 women remained after exclusions. Overall<50% of subjects had measureable concentrations ≥ LoD: 45.6% no exclusion, 43.5% after exclusion; compared to men: 68.1% no exclusion, 65.1% post exclusion; women: 22.7% no exclusion, 20.9% post exclusion. The statistical method used influenced URLs as follows: pre/post exclusion overall, NP 16/16ng/L, HDE 17/17ng/L, R not available; men NP 18/16ng/L, HDE 21/19ng/L, R 16/11ng/L; women NP 13/10ng/L, HDE 14/14ng/L, R not available. We demonstrated that a) the Gen 5 cTnT assay does not meet the IFCC guideline for high-sensitivity assays, b) surrogate biomarkers significantly lowers the URLs and c) statistical methods used impact URLs. Our data suggest lower sex-specific cTnT 99th percentiles than reported in the FDA approved package insert. We emphasize the importance of detailing the criteria used to include and exclude subjects for defining a healthy population and the statistical method used to calculate 99th percentiles and identify outliers. Copyright © 2017 The Canadian Society of Clinical Chemists. Published by Elsevier Inc. All rights reserved.

  3. Metabarcoding Is Powerful yet Still Blind: A Comparative Analysis of Morphological and Molecular Surveys of Seagrass Communities

    PubMed Central

    Cowart, Dominique A.; Pinheiro, Miguel; Mouchel, Olivier; Maguer, Marion; Grall, Jacques; Miné, Jacques; Arnaud-Haond, Sophie

    2015-01-01

    In the context of the sixth wave of extinction, reliable surveys of biodiversity are increasingly needed to infer the cause and consequences of species and community declines, identify early warning indicators of tipping points, and provide reliable impact assessments before engaging in activities with potential environmental hazards. DNA metabarcoding has emerged as having potential to provide speedy assessment of community structure from environmental samples. Here we tested the reliability of metabarcoding by comparing morphological and molecular inventories of invertebrate communities associated with seagrasses through estimates of alpha and beta diversity, as well as the identification of the most abundant taxa. Sediment samples were collected from six Zostera marina seagrass meadows across Brittany, France. Metabarcoding surveys were performed using both mitochondrial (Cytochrome Oxidase I) and nuclear (small subunit 18S ribosomal RNA) markers, and compared to morphological inventories compiled by a long-term benthic monitoring network. A sampling strategy was defined to enhance performance and accuracy of results by preventing the dominance of larger animals, boosting statistical support through replicates, and using two genes to compensate for taxonomic biases. Molecular barcodes proved powerful by revealing a remarkable level of diversity that vastly exceeded the morphological survey, while both surveys identified congruent differentiation of the meadows. However, despite the addition of individual barcodes of common species into taxonomic reference databases, the retrieval of only 36% of these species suggest that the remaining were either not present in the molecular samples or not detected by the molecular screening. This finding exemplifies the necessity of comprehensive and well-curated taxonomic reference libraries and multi-gene surveys. Overall, results offer methodological guidelines and support for metabarcoding as a powerful and repeatable method of characterizing communities, while also presenting suggestions for improvement, including implementation of pilot studies prior to performing full “blind” metabarcoding assessments to optimize sampling and amplification protocols. PMID:25668035

  4. Noninformative prior in the quantum statistical model of pure states

    NASA Astrophysics Data System (ADS)

    Tanaka, Fuyuhiko

    2012-06-01

    In the present paper, we consider a suitable definition of a noninformative prior on the quantum statistical model of pure states. While the full pure-states model is invariant under unitary rotation and admits the Haar measure, restricted models, which we often see in quantum channel estimation and quantum process tomography, have less symmetry and no compelling rationale for any choice. We adopt a game-theoretic approach that is applicable to classical Bayesian statistics and yields a noninformative prior for a general class of probability distributions. We define the quantum detection game and show that there exist noninformative priors for a general class of a pure-states model. Theoretically, it gives one of the ways that we represent ignorance on the given quantum system with partial information. Practically, our method proposes a default distribution on the model in order to use the Bayesian technique in the quantum-state tomography with a small sample.

  5. A framework for joint image-and-shape analysis

    NASA Astrophysics Data System (ADS)

    Gao, Yi; Tannenbaum, Allen; Bouix, Sylvain

    2014-03-01

    Techniques in medical image analysis are many times used for the comparison or regression on the intensities of images. In general, the domain of the image is a given Cartesian grids. Shape analysis, on the other hand, studies the similarities and differences among spatial objects of arbitrary geometry and topology. Usually, there is no function defined on the domain of shapes. Recently, there has been a growing needs for defining and analyzing functions defined on the shape space, and a coupled analysis on both the shapes and the functions defined on them. Following this direction, in this work we present a coupled analysis for both images and shapes. As a result, the statistically significant discrepancies in both the image intensities as well as on the underlying shapes are detected. The method is applied on both brain images for the schizophrenia and heart images for atrial fibrillation patients.

  6. Probabilistic Open Set Recognition

    NASA Astrophysics Data System (ADS)

    Jain, Lalit Prithviraj

    Real-world tasks in computer vision, pattern recognition and machine learning often touch upon the open set recognition problem: multi-class recognition with incomplete knowledge of the world and many unknown inputs. An obvious way to approach such problems is to develop a recognition system that thresholds probabilities to reject unknown classes. Traditional rejection techniques are not about the unknown; they are about the uncertain boundary and rejection around that boundary. Thus traditional techniques only represent the "known unknowns". However, a proper open set recognition algorithm is needed to reduce the risk from the "unknown unknowns". This dissertation examines this concept and finds existing probabilistic multi-class recognition approaches are ineffective for true open set recognition. We hypothesize the cause is due to weak adhoc assumptions combined with closed-world assumptions made by existing calibration techniques. Intuitively, if we could accurately model just the positive data for any known class without overfitting, we could reject the large set of unknown classes even under this assumption of incomplete class knowledge. For this, we formulate the problem as one of modeling positive training data by invoking statistical extreme value theory (EVT) near the decision boundary of positive data with respect to negative data. We provide a new algorithm called the PI-SVM for estimating the unnormalized posterior probability of class inclusion. This dissertation also introduces a new open set recognition model called Compact Abating Probability (CAP), where the probability of class membership decreases in value (abates) as points move from known data toward open space. We show that CAP models improve open set recognition for multiple algorithms. Leveraging the CAP formulation, we go on to describe the novel Weibull-calibrated SVM (W-SVM) algorithm, which combines the useful properties of statistical EVT for score calibration with one-class and binary support vector machines. Building from the success of statistical EVT based recognition methods such as PI-SVM and W-SVM on the open set problem, we present a new general supervised learning algorithm for multi-class classification and multi-class open set recognition called the Extreme Value Local Basis (EVLB). The design of this algorithm is motivated by the observation that extrema from known negative class distributions are the closest negative points to any positive sample during training, and thus should be used to define the parameters of a probabilistic decision model. In the EVLB, the kernel distribution for each positive training sample is estimated via an EVT distribution fit over the distances to the separating hyperplane between positive training sample and closest negative samples, with a subset of the overall positive training data retained to form a probabilistic decision boundary. Using this subset as a frame of reference, the probability of a sample at test time decreases as it moves away from the positive class. Possessing this property, the EVLB is well-suited to open set recognition problems where samples from unknown or novel classes are encountered at test. Our experimental evaluation shows that the EVLB provides a substantial improvement in scalability compared to standard radial basis function kernel machines, as well as P I-SVM and W-SVM, with improved accuracy in many cases. We evaluate our algorithm on open set variations of the standard visual learning benchmarks, as well as with an open subset of classes from Caltech 256 and ImageNet. Our experiments show that PI-SVM, WSVM and EVLB provide significant advances over the previous state-of-the-art solutions for the same tasks.

  7. Correlation Measurement of Lambda-anti-Lambda, Lambda-Lambda and anti-Lambda-anti-Lambda with the ATLAS detector at s=7 TeV

    NASA Astrophysics Data System (ADS)

    Cheng, Hok-Chuen

    This thesis summaries the measurements of correlations between Lambda 0Lambda0, Lambda0Lambda 0, and Lambda0Lambda 0 hyperon pairs produced inclusively at the LHC, which are useful for a better understanding of the quark-antiquark pair production and jet fragmentation and hadronization processes. The analysis is based on hyperon pairs selected using the muon and minimum bias data samples collected at the ATLAS experiment from proton-proton collisions at a center-of-mass energy of 7 TeV in 2010. Excess Lambda0Lambda 0 are observed near the production threshold and are identified to be originated from the parton system in the string model in the MC sample, decaying either directly or through heavy strange resonances such as Sigma0 and Sigma*(1385). Dynamical correlations have been explored through a correlation function defined as the ratio of two-particle to single-particle densities. Positive correlation is observed for Lambda0Lambda0 and anticorrelation is observed for Lambda0Lambda 0 and Lambda0Lambda 0 for Q in [0,2] GeV. The structure replicates similar correlations in pp, pp, and pppp events in PYTHIA generator as predicted by the Lund string fragmentation model. Parameters of the "popcorn" mechanism implemented in the PYTHIA generator are tuned and are found to have little impact on the structure observed. The spin composition of the sample is extracted using a data-driven reference sample built by event mixing. Appropriate corrections have been made to the kinematic distributions in the reference sample by kinematic weighting to make sure that the detector effects are well modeled. A modified Pearson's chi2 test statistics is calculated for the costheta* distribution to determine the best-fitted A-value for data. The results are consistent with zero for both like-type and unlike-type hyperon pairs in Q ∈ [0,10] GeV and Q ∈ [1,10] GeV respectively. The data statistics in the range of Q ∈ [0, 1] GeV is currently too low for the estimation of the emitter size for Fermi-Dirac correlation.

  8. Pesticides in groundwater of the United States: decadal-scale changes, 1993-2011

    USGS Publications Warehouse

    Toccalino, Patricia L.; Gilliom, Robert J.; Lindsey, Bruce D.; Rupert, Michael G.

    2014-01-01

    The national occurrence of 83 pesticide compounds in groundwater of the United States and decadal-scale changes in concentrations for 35 compounds were assessed for the 20-year period from 1993–2011. Samples were collected from 1271 wells in 58 nationally distributed well networks. Networks consisted of shallow (mostly monitoring) wells in agricultural and urban land-use areas and deeper (mostly domestic and public supply) wells in major aquifers in mixed land-use areas. Wells were sampled once during 1993–2001 and once during 2002–2011. Pesticides were frequently detected (53% of all samples), but concentrations seldom exceeded human-health benchmarks (1.8% of all samples). The five most frequently detected pesticide compounds—atrazine, deethylatrazine, simazine, metolachlor, and prometon—each had statistically significant (p < 0.1) changes in concentrations between decades in one or more categories of well networks nationally aggregated by land use. For agricultural networks, concentrations of atrazine, metolachlor, and prometon decreased from the first decade to the second decade. For urban networks, deethylatrazine concentrations increased and prometon concentrations decreased. For major aquifers, concentrations of deethylatrazine and simazine increased. The directions of concentration changes for individual well networks generally were consistent with changes determined from nationally aggregated data. Altogether, 36 of the 58 individual well networks had statistically significant changes in concentrations of one or more pesticides between decades, with the majority of changes attributed to the five most frequently detected pesticide compounds. The magnitudes of median decadal-scale concentration changes were small—ranging from −0.09 to 0.03 µg/L—and were 35- to 230,000-fold less than human-health benchmarks.

  9. The impact of study design and diagnostic approach in a large multi-centre ADHD study. Part 1: ADHD symptom patterns

    PubMed Central

    2011-01-01

    Background The International Multi-centre ADHD Genetics (IMAGE) project with 11 participating centres from 7 European countries and Israel has collected a large behavioural and genetic database for present and future research. Behavioural data were collected from 1068 probands with the combined type of attention deficit/hyperactivity disorder (ADHD-CT) and 1446 'unselected' siblings. The aim was to analyse the IMAGE sample with respect to demographic features (gender, age, family status, and recruiting centres) and psychopathological characteristics (diagnostic subtype, symptom frequencies, age at symptom detection, and comorbidities). A particular focus was on the effects of the study design and the diagnostic procedure on the homogeneity of the sample in terms of symptom-based behavioural data, and potential consequences for further analyses based on these data. Methods Diagnosis was based on the Parental Account of Childhood Symptoms (PACS) interview and the DSM-IV items of the Conners' teacher questionnaire. Demographics of the full sample and the homogeneity of a subsample (all probands) were analysed by using robust statistical procedures which were adjusted for unequal sample sizes and skewed distributions. These procedures included multi-way analyses based on trimmed means and winsorised variances as well as bootstrapping. Results Age and proband/sibling ratios differed between participating centres. There was no significant difference in the distribution of gender between centres. There was a significant interaction between age and centre for number of inattentive, but not number of hyperactive symptoms. Higher ADHD symptom frequencies were reported by parents than teachers. The diagnostic symptoms differed from each other in their frequencies. The face-to-face interview was more sensitive than the questionnaire. The differentiation between ADHD-CT probands and unaffected siblings was mainly due to differences in hyperactive/impulsive symptoms. Conclusions Despite a symptom-based standardized inclusion procedure according to DSM-IV criteria with defined symptom thresholds, centres may differ markedly in probands' ADHD symptom frequencies. Both the diagnostic procedure and the multi-centre design influence the behavioural characteristics of a sample and, thus, may bias statistical analyses, particularly in genetic or neurobehavioral studies. PMID:21473745

  10. A Graphic Survey of Book Publication, 1890-1916. Bulletin, 1917, No. 14

    ERIC Educational Resources Information Center

    Woodward, Fred E.

    1917-01-01

    The rapid increase in the number of books published in the United States, especially during the past decade, has been the subject of much comment. Statistics are collected regularly by the trade papers, and these figures give the number of books for each year in each of several well-defined classes. There are at the present time 24 classes: as…

  11. 2016 Workplace and Gender Relations Survey of Active Duty Members: Frequently Asked Questions

    DTIC Science & Technology

    2017-05-01

    active duty population both at the sample design stage as well as during the statistical weighting process to account for survey non-response and post...used the OPA sampling design , won the 2011 Policy Impact Award from The American Association for Public Opinion Research (AAPOR), which “recognizes

  12. The use of IRMS, (1)H NMR and chemical analysis to characterise Italian and imported Tunisian olive oils.

    PubMed

    Camin, Federica; Pavone, Anita; Bontempo, Luana; Wehrens, Ron; Paolini, Mauro; Faberi, Angelo; Marianella, Rosa Maria; Capitani, Donatella; Vista, Silvia; Mannina, Luisa

    2016-04-01

    Isotope Ratio Mass Spectrometry (IRMS), (1)H Nuclear Magnetic Resonance ((1)H NMR), conventional chemical analysis and chemometric elaboration were used to assess quality and to define and confirm the geographical origin of 177 Italian PDO (Protected Denomination of Origin) olive oils and 86 samples imported from Tunisia. Italian olive oils were richer in squalene and unsaturated fatty acids, whereas Tunisian olive oils showed higher δ(18)O, δ(2)H, linoleic acid, saturated fatty acids β-sitosterol, sn-1 and 3 diglyceride values. Furthermore, all the Tunisian samples imported were of poor quality, with a K232 and/or acidity values above the limits established for extra virgin olive oils. By combining isotopic composition with (1)H NMR data using a multivariate statistical approach, a statistical model able to discriminate olive oil from Italy and those imported from Tunisia was obtained, with an optimal differentiation ability arriving at around 98%. Copyright © 2015 Elsevier Ltd. All rights reserved.

  13. SPICE: exploration and analysis of post-cytometric complex multivariate datasets.

    PubMed

    Roederer, Mario; Nozzi, Joshua L; Nason, Martha C

    2011-02-01

    Polychromatic flow cytometry results in complex, multivariate datasets. To date, tools for the aggregate analysis of these datasets across multiple specimens grouped by different categorical variables, such as demographic information, have not been optimized. Often, the exploration of such datasets is accomplished by visualization of patterns with pie charts or bar charts, without easy access to statistical comparisons of measurements that comprise multiple components. Here we report on algorithms and a graphical interface we developed for these purposes. In particular, we discuss thresholding necessary for accurate representation of data in pie charts, the implications for display and comparison of normalized versus unnormalized data, and the effects of averaging when samples with significant background noise are present. Finally, we define a statistic for the nonparametric comparison of complex distributions to test for difference between groups of samples based on multi-component measurements. While originally developed to support the analysis of T cell functional profiles, these techniques are amenable to a broad range of datatypes. Published 2011 Wiley-Liss, Inc.

  14. Multiplex biomarker approach for determining risk of prostate-specific antigen-defined recurrence of prostate cancer.

    PubMed

    Rhodes, Daniel R; Sanda, Martin G; Otte, Arie P; Chinnaiyan, Arul M; Rubin, Mark A

    2003-05-07

    Molecular signatures in cancer tissue may be useful for diagnosis and are associated with survival. We used results from high-density tissue microarrays (TMAs) to define combinations of candidate biomarkers associated with the rate of prostate cancer progression after radical prostatectomy that could identify patients at high risk for recurrence. Fourteen candidate biomarkers for prostate cancer for which antibodies are available included hepsin, pim-1 kinase, E-cadherin (ECAD; cell adhesion molecule), alpha-methylacyl-coenzyme A racemase, and EZH2 (enhancer of zeste homolog 2, a transcriptional repressor). TMAs containing more than 2000 tumor samples from 259 patients who underwent radical prostatectomy for localized prostate cancer were studied with these antibodies. Immunohistochemistry results were evaluated in conjunction with clinical parameters associated with prostate cancer progression, including tumor stage, Gleason score, and prostate-specific antigen (PSA) level. Recurrence was defined as a postsurgery PSA level of more than 0.2 ng/mL. All statistical tests were two-sided. Moderate or strong expression of EZH2 coupled with at most moderate expression of ECAD (i.e., a positive EZH2:ECAD status) was the biomarker combination that was most strongly associated with the recurrence of prostate cancer. EZH2:ECAD status was statistically significantly associated with prostate cancer recurrence in a training set of 103 patients (relative risk [RR] = 2.52, 95% confidence interval [CI] = 1.09 to 5.81; P =.021), in a validation set of 80 patients (RR = 3.72, 95% CI = 1.27 to 10.91; P =.009), and in the combined set of 183 patients (RR = 2.96, 95% CI = 1.56 to 5.61; P<.001). EZH2:ECAD status was statistically significantly associated with disease recurrence even after adjusting for clinical parameters, such as tumor stage, Gleason score, and PSA level (hazard ratio = 3.19, 95% CI = 1.50 to 6.77; P =.003). EZH2:ECAD status was statistically significantly associated with prostate cancer recurrence after radical prostatectomy and may be useful in defining a cohort of high-risk patients.

  15. Reproducible detection of disease-associated markers from gene expression data.

    PubMed

    Omae, Katsuhiro; Komori, Osamu; Eguchi, Shinto

    2016-08-18

    Detection of disease-associated markers plays a crucial role in gene screening for biological studies. Two-sample test statistics, such as the t-statistic, are widely used to rank genes based on gene expression data. However, the resultant gene ranking is often not reproducible among different data sets. Such irreproducibility may be caused by disease heterogeneity. When we divided data into two subsets, we found that the signs of the two t-statistics were often reversed. Focusing on such instability, we proposed a sign-sum statistic that counts the signs of the t-statistics for all possible subsets. The proposed method excludes genes affected by heterogeneity, thereby improving the reproducibility of gene ranking. We compared the sign-sum statistic with the t-statistic by a theoretical evaluation of the upper confidence limit. Through simulations and applications to real data sets, we show that the sign-sum statistic exhibits superior performance. We derive the sign-sum statistic for getting a robust gene ranking. The sign-sum statistic gives more reproducible ranking than the t-statistic. Using simulated data sets we show that the sign-sum statistic excludes hetero-type genes well. Also for the real data sets, the sign-sum statistic performs well in a viewpoint of ranking reproducibility.

  16. TRAN-STAT: statistics for environmental transuranic studies, July 1978, Number 5

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    This issue is concerned with nonparametric procedures for (1) estimating the central tendency of a population, (2) describing data sets through estimating percentiles, (3) estimating confidence limits for the median and other percentiles, (4) estimating tolerance limits and associated numbers of samples, and (5) tests of significance and associated procedures for a variety of testing situations (counterparts to t-tests and analysis of variance). Some characteristics of several nonparametric tests are illustrated using the NAEG /sup 241/Am aliquot data presented and discussed in the April issue of TRAN-STAT. Some of the statistical terms used here are defined in a glossary. Themore » reference list also includes short descriptions of nonparametric books. 31 references, 3 figures, 1 table.« less

  17. The SDSS-IV MaNGA Sample: Design, Optimization, and Usage Considerations

    NASA Astrophysics Data System (ADS)

    Wake, David A.; Bundy, Kevin; Diamond-Stanic, Aleksandar M.; Yan, Renbin; Blanton, Michael R.; Bershady, Matthew A.; Sánchez-Gallego, José R.; Drory, Niv; Jones, Amy; Kauffmann, Guinevere; Law, David R.; Li, Cheng; MacDonald, Nicholas; Masters, Karen; Thomas, Daniel; Tinker, Jeremy; Weijmans, Anne-Marie; Brownstein, Joel R.

    2017-09-01

    We describe the sample design for the SDSS-IV MaNGA survey and present the final properties of the main samples along with important considerations for using these samples for science. Our target selection criteria were developed while simultaneously optimizing the size distribution of the MaNGA integral field units (IFUs), the IFU allocation strategy, and the target density to produce a survey defined in terms of maximizing signal-to-noise ratio, spatial resolution, and sample size. Our selection strategy makes use of redshift limits that only depend on I-band absolute magnitude (M I ), or, for a small subset of our sample, M I and color (NUV - I). Such a strategy ensures that all galaxies span the same range in angular size irrespective of luminosity and are therefore covered evenly by the adopted range of IFU sizes. We define three samples: the Primary and Secondary samples are selected to have a flat number density with respect to M I and are targeted to have spectroscopic coverage to 1.5 and 2.5 effective radii (R e ), respectively. The Color-Enhanced supplement increases the number of galaxies in the low-density regions of color-magnitude space by extending the redshift limits of the Primary sample in the appropriate color bins. The samples cover the stellar mass range 5× {10}8≤slant {M}* ≤slant 3× {10}11 {M}⊙ {h}-2 and are sampled at median physical resolutions of 1.37 and 2.5 kpc for the Primary and Secondary samples, respectively. We provide weights that will statistically correct for our luminosity and color-dependent selection function and IFU allocation strategy, thus correcting the observed sample to a volume-limited sample.

  18. Small sample mediation testing: misplaced confidence in bootstrapped confidence intervals.

    PubMed

    Koopman, Joel; Howe, Michael; Hollenbeck, John R; Sin, Hock-Peng

    2015-01-01

    Bootstrapping is an analytical tool commonly used in psychology to test the statistical significance of the indirect effect in mediation models. Bootstrapping proponents have particularly advocated for its use for samples of 20-80 cases. This advocacy has been heeded, especially in the Journal of Applied Psychology, as researchers are increasingly utilizing bootstrapping to test mediation with samples in this range. We discuss reasons to be concerned with this escalation, and in a simulation study focused specifically on this range of sample sizes, we demonstrate not only that bootstrapping has insufficient statistical power to provide a rigorous hypothesis test in most conditions but also that bootstrapping has a tendency to exhibit an inflated Type I error rate. We then extend our simulations to investigate an alternative empirical resampling method as well as a Bayesian approach and demonstrate that they exhibit comparable statistical power to bootstrapping in small samples without the associated inflated Type I error. Implications for researchers testing mediation hypotheses in small samples are presented. For researchers wishing to use these methods in their own research, we have provided R syntax in the online supplemental materials. (c) 2015 APA, all rights reserved.

  19. Water-quality characteristics and trends for selected sites at and near the Idaho National Laboratory, Idaho, 1949-2009

    USGS Publications Warehouse

    Bartholomay, Roy C.; Davis, Linda C.; Fisher, Jason C.; Tucker, Betty J.; Raben, Flint A.

    2012-01-01

    The U.S. Geological Survey, in cooperation with the U.S. Department of Energy, analyzed water-quality data collected from 67 aquifer wells and 7 surface-water sites at the Idaho National Laboratory (INL) from 1949 through 2009. The data analyzed included major cations, anions, nutrients, trace elements, and total organic carbon. The analyses were performed to examine water-quality trends that might inform future management decisions about the number of wells to sample at the INL and the type of constituents to monitor. Water-quality trends were determined using (1) the nonparametric Kendall's tau correlation coefficient, p-value, Theil-Sen slope estimator, and summary statistics for uncensored data; and (2) the Kaplan-Meier method for calculating summary statistics, Kendall's tau correlation coefficient, p-value, and Akritas-Theil-Sen slope estimator for robust linear regression for censored data. Statistical analyses for chloride concentrations indicate that groundwater influenced by Big Lost River seepage has decreasing chloride trends or, in some cases, has variable chloride concentration changes that correlate with above-average and below-average periods of recharge. Analyses of trends for chloride in water samples from four sites located along the Big Lost River indicate a decreasing trend or no trend for chloride, and chloride concentrations generally are much lower at these four sites than those in the aquifer. Above-average and below-average periods of recharge also affect concentration trends for sodium, sulfate, nitrate, and a few trace elements in several wells. Analyses of trends for constituents in water from several of the wells that is mostly regionally derived groundwater generally indicate increasing trends for chloride, sodium, sulfate, and nitrate concentrations. These increases are attributed to agricultural or other anthropogenic influences on the aquifer upgradient of the INL. Statistical trends of chemical constituents from several wells near the Naval Reactors Facility may be influenced by wastewater disposal at the facility or by anthropogenic influence from the Little Lost River basin. Groundwater samples from three wells downgradient of the Power Burst Facility area show increasing trends for chloride, nitrate, sodium, and sulfate concentrations. The increases could be caused by wastewater disposal in the Power Burst Facility area. Some groundwater samples in the southwestern part of the INL and southwest of the INL show concentration trends for chloride and sodium that may be influenced by wastewater disposal. Some of the groundwater samples have decreasing trends that could be attributed to the decreasing concentrations in the wastewater from the late 1970s to 2009. The young fraction of groundwater in many of the wells is more than 20 years old, so samples collected in the early 1990s are more representative of groundwater discharged in the 1960s and 1970s, when concentrations in wastewater were much higher. Groundwater sampled in 2009 would be representative of the lower concentrations of chloride and sodium in wastewater discharged in the late 1980s. Analyses of trends for sodium in several groundwater samples from the central and southern part of the eastern Snake River aquifer show increasing trends. In most cases, however, the sodium concentrations are less than background concentrations measured in the aquifer. Many of the wells are open to larger mixed sections of the aquifer, and the increasing trends may indicate that the long history of wastewater disposal in the central part of the INL is increasing sodium concentrations in the groundwater.

  20. Infants' statistical learning: 2- and 5-month-olds' segmentation of continuous visual sequences.

    PubMed

    Slone, Lauren Krogh; Johnson, Scott P

    2015-05-01

    Past research suggests that infants have powerful statistical learning abilities; however, studies of infants' visual statistical learning offer differing accounts of the developmental trajectory of and constraints on this learning. To elucidate this issue, the current study tested the hypothesis that young infants' segmentation of visual sequences depends on redundant statistical cues to segmentation. A sample of 20 2-month-olds and 20 5-month-olds observed a continuous sequence of looming shapes in which unit boundaries were defined by both transitional probability and co-occurrence frequency. Following habituation, only 5-month-olds showed evidence of statistically segmenting the sequence, looking longer to a statistically improbable shape pair than to a probable pair. These results reaffirm the power of statistical learning in infants as young as 5 months but also suggest considerable development of statistical segmentation ability between 2 and 5 months of age. Moreover, the results do not support the idea that infants' ability to segment visual sequences based on transitional probabilities and/or co-occurrence frequencies is functional at the onset of visual experience, as has been suggested previously. Rather, this type of statistical segmentation appears to be constrained by the developmental state of the learner. Factors contributing to the development of statistical segmentation ability during early infancy, including memory and attention, are discussed. Copyright © 2015 Elsevier Inc. All rights reserved.

  1. Increased absenteeism from work among aware and treated hypertensive and hypercholesterolaemic patients.

    PubMed

    Leynen, Françoise; De Backer, Guy; Pelfrene, Edwin; Clays, Els; Kittel, France; Moreau, Michel; Kornitzer, Marcel

    2006-04-01

    The 'labelling hypothesis' was introduced on the basis of the observation that labelling subjects with blood pressure elevation as hypertensive was associated with an increase in sickness absence. In the Belstress I study this hypothesis was analysed in the same way for the possible influence on sick leave of labelling persons with elevated cholesterol as hypercholesterolaemic. The Belstress I cohort concerns a sample of more than 16,000 men and 5,000 women at work in 24 Belgian industries in various sectors. Baseline data were collected by questionnaire and clinical examination. Awareness was defined as answering positively to the question 'did a physician ever tell you that your blood pressure/serum cholesterol was too high?' Sick leave data were independently and objectively recorded during 1 year following the screening. Sick leave was treated in a dichotomous way whereby the event was defined as being in the highest quartile of the annual number of days of sick leave (10 days or more for men and 15 days or more for women) or as being in the highest quartile of the annual number of spells of sick leave (two spells or more for both sexes). Gender-specific logistic regression analyses were performed, with adjustment for a large set of covariates. A positive association was observed between both awareness of hypertension and awareness of hypercholesterolaemia and the various definitions of sick leave, in both sexes and after adjustment for different covariates. When dividing up aware subjects into treated versus untreated, we observed in men the highest sick leave incidence in aware and treated hypertensive patients as well as in aware and treated hypercholesterolaemic patients. In women findings were less consistent, probably due to the smaller sample size. When looking at cumulative effects by examining participants with both hypertension and hypercholesterolaemia and their level of awareness for one or both risk factors, a statistically significant gradient was noticed in men, with the highest sick leave incidence, whatever the definition, in men aware for both risk factors, followed by men aware for one. In women the same trends were observed, but no level of statistical significance was reached. Without being able to test the effect of 'labelling' as such, our study provides support for the association between awareness of two different coronary risk factors and incidence of sick leave. Probably a common mechanism is at the base of these findings. Further research is needed, in order to reduce potential negative effects of screening on human wellbeing as well as on productivity.

  2. Monitoring the Heavens, Today, and Tomorrow

    NASA Technical Reports Server (NTRS)

    Johnson, Nicholas L.

    2006-01-01

    The current Earth satellite population in LEO for all sizes is relatively well-established by a combination of deterministic and statistical means. At higher altitudes, the population of satellites with diameters of less than 1 m is not well defined. Although a few new sensors might become operational in the near- to mid-term, no major improvement in environment characterization is anticipated during this period. With the increasing deployment of micro- and pico-satellites and with the continued growth of the small debris population, a need exists for better space surveillance to support spacecraft design and operations.

  3. Differential gene expression detection and sample classification using penalized linear regression models.

    PubMed

    Wu, Baolin

    2006-02-15

    Differential gene expression detection and sample classification using microarray data have received much research interest recently. Owing to the large number of genes p and small number of samples n (p > n), microarray data analysis poses big challenges for statistical analysis. An obvious problem owing to the 'large p small n' is over-fitting. Just by chance, we are likely to find some non-differentially expressed genes that can classify the samples very well. The idea of shrinkage is to regularize the model parameters to reduce the effects of noise and produce reliable inferences. Shrinkage has been successfully applied in the microarray data analysis. The SAM statistics proposed by Tusher et al. and the 'nearest shrunken centroid' proposed by Tibshirani et al. are ad hoc shrinkage methods. Both methods are simple, intuitive and prove to be useful in empirical studies. Recently Wu proposed the penalized t/F-statistics with shrinkage by formally using the (1) penalized linear regression models for two-class microarray data, showing good performance. In this paper we systematically discussed the use of penalized regression models for analyzing microarray data. We generalize the two-class penalized t/F-statistics proposed by Wu to multi-class microarray data. We formally derive the ad hoc shrunken centroid used by Tibshirani et al. using the (1) penalized regression models. And we show that the penalized linear regression models provide a rigorous and unified statistical framework for sample classification and differential gene expression detection.

  4. Addendum to Sampling and Analysis Plan (SAP) for Assessment of LANL-Derived Residual Radionuclides in Soils within Tract A-16-d for Land Conveyance and Transfer for Sewage Treatment Facility Area

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Whicker, Jeffrey Jay; Gillis, Jessica Mcdonnel; Ruedig, Elizabeth

    This report summarizes the sampling design used, associated statistical assumptions, as well as general guidelines for conducting post-sampling data analysis. Sampling plan components presented here include how many sampling locations to choose and where within the sampling area to collect those samples. The type of medium to sample (i.e., soil, groundwater, etc.) and how to analyze the samples (in-situ, fixed laboratory, etc.) are addressed in other sections of the sampling plan.

  5. Space shuttle solid rocket booster recovery system definition, volume 1

    NASA Technical Reports Server (NTRS)

    1973-01-01

    The performance requirements, preliminary designs, and development program plans for an airborne recovery system for the space shuttle solid rocket booster are discussed. The analyses performed during the study phase of the program are presented. The basic considerations which established the system configuration are defined. A Monte Carlo statistical technique using random sampling of the probability distribution for the critical water impact parameters was used to determine the failure probability of each solid rocket booster component as functions of impact velocity and component strength capability.

  6. Comparative statistical analysis of carcinogenic and non-carcinogenic effects of uranium in groundwater samples from different regions of Punjab, India.

    PubMed

    Saini, Komal; Singh, Parminder; Bajwa, Bikramjit Singh

    2016-12-01

    LED flourimeter has been used for microanalysis of uranium concentration in groundwater samples collected from six districts of South West (SW), West (W) and North East (NE) Punjab, India. Average value of uranium content in water samples of SW Punjab is observed to be higher than WHO, USEPA recommended safe limit of 30µgl -1 as well as AERB proposed limit of 60µgl -1 . Whereas, for W and NE region of Punjab, average level of uranium concentration was within AERB recommended limit of 60µgl -1 . Average value observed in SW Punjab is around 3-4 times the value observed in W Punjab, whereas its value is more than 17 times the average value observed in NE region of Punjab. Statistical analysis of carcinogenic as well as non carcinogenic risks due to uranium have been evaluated for each studied district. Copyright © 2016 Elsevier Ltd. All rights reserved.

  7. The assignment of scores procedure for ordinal categorical data.

    PubMed

    Chen, Han-Ching; Wang, Nae-Sheng

    2014-01-01

    Ordinal data are the most frequently encountered type of data in the social sciences. Many statistical methods can be used to process such data. One common method is to assign scores to the data, convert them into interval data, and further perform statistical analysis. There are several authors who have recently developed assigning score methods to assign scores to ordered categorical data. This paper proposes an approach that defines an assigning score system for an ordinal categorical variable based on underlying continuous latent distribution with interpretation by using three case study examples. The results show that the proposed score system is well for skewed ordinal categorical data.

  8. Well-Being in the Context of Workplace Ethnic Diversity

    ERIC Educational Resources Information Center

    Enchautegui-de-Jesus, Noemi; Hughes, Diane; Johnston, Kristen E.; Oh, Hyun Joo

    2006-01-01

    This research examined the relation between the effects of workplace diversity (defined as the proportion of coworkers of same ethnicity as the respondent) and psychosomatic complaints, psychological well-being, life satisfaction, and job satisfaction. A sample of 648 African American and Latino workers was surveyed in Chicago and New York City. A…

  9. Experimental study of switching in a rho-i(MQW)-eta vertical coupler

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cavailles, J.A.; Erman, M.; Woodbridge, K.

    1989-11-01

    Electrically controlled switching in a vertically arranged directional coupler with GaAs/GaAIAs multiple quantum well waveguides is demonstrated. Coupling lengths and extinction parameters are determined by using a sample processed in such a way that injection conditions are well defined and that the coupler length can be varied continuously.

  10. Improving Generalizations from Experiments Using Propensity Score Subclassification: Assumptions, Properties, and Contexts

    ERIC Educational Resources Information Center

    Tipton, Elizabeth

    2013-01-01

    As a result of the use of random assignment to treatment, randomized experiments typically have high internal validity. However, units are very rarely randomly selected from a well-defined population of interest into an experiment; this results in low external validity. Under nonrandom sampling, this means that the estimate of the sample average…

  11. 42 CFR 485.610 - Condition of participation: Status and location.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... requirements: (i) The CAH is located outside any area that is a Metropolitan Statistical Area, as defined by... Statistical Area, as defined by the Office of Management and Budget, but is being treated as being located in... this section and is located in a county that, in FY 2004, was not part of a Metropolitan Statistical...

  12. Assessment of physicochemical and antioxidant characteristics of Quercus pyrenaica honeydew honeys.

    PubMed

    Shantal Rodríguez Flores, M; Escuredo, Olga; Carmen Seijo, M

    2015-01-01

    Consumers are exhibiting increasing interest in honeydew honey, principally due to its functional properties. Some plants can be sources of honeydew honey, but in north-western Spain, this honey type only comes from Quercus pyrenaica. In the present study, the melissopalynological and physicochemical characteristics and the antioxidant properties of 32 honeydew honey samples are described. Q. pyrenaica honeydew honey was defined by its colour, high pH, phenols and flavonoids. Multivariate statistical techniques were used to analyse the influence of the production year on the honey's physicochemical parameters and polyphenol content. Differences among the honey samples were found, showing that weather affected the physicochemical composition of the honey samples. Optimal conditions for oak growth favoured the production of honeydew honey. Copyright © 2014 Elsevier Ltd. All rights reserved.

  13. Issues in the classification of disease instances with ontologies.

    PubMed

    Burgun, Anita; Bodenreider, Olivier; Jacquelinet, Christian

    2005-01-01

    Ontologies define classes of entities and their interrelations. They are used to organize data according to a theory of the domain. Towards that end, ontologies provide class definitions (i.e., the necessary and sufficient conditions for defining class membership). In medical ontologies, it is often difficult to establish such definitions for diseases. We use three examples (anemia, leukemia and schizophrenia) to illustrate the limitations of ontologies as classification resources. We show that eligibility criteria are often more useful than the Aristotelian definitions traditionally used in ontologies. Examples of eligibility criteria for diseases include complex predicates such as ' x is an instance of the class C when at least n criteria among m are verified' and 'symptoms must last at least one month if not treated, but less than one month, if effectively treated'. References to normality and abnormality are often found in disease definitions, but the operational definition of these references (i.e., the statistical and contextual information necessary to define them) is rarely provided. We conclude that knowledge bases that include probabilistic and statistical knowledge as well as rule-based criteria are more useful than Aristotelian definitions for representing the predicates defined by necessary and sufficient conditions. Rich knowledge bases are needed to clarify the relations between individuals and classes in various studies and applications. However, as ontologies represent relations among classes, they can play a supporting role in disease classification services built primarily on knowledge bases.

  14. Analysis of spatial dynamic of epizootic process of bluetongue and its risk factors.

    PubMed

    Bouchemla, Fayssal; Popova, Olga Mikhailovna; Agoltsov, Valerey Alexandrovich

    2017-10-01

    The study was undertaken to find out the spatial dynamic occurrence and patterns of the global spread of bluetongue (BT) disease for the period from 1996 to 2016, as well as the assessment of the risk of occurrence and its spread in 2017-2018. Outbreaks (serum samples were collected from clinically healthy as well as suspected animals in infected points) were confirmed and reported officially by veterinary departments which represent different geographical regions in the world to World Organization for Animal Health. These reports explained that ELISA and polymerase chain reaction were used to identify the BT disease, taking in the account number of infected, dead animals, and focus of BT infection in all susceptible animals from 1996 to 2016. Once conventional statistical population was defined (an observational study), we had classified data as well as possible to answer to our aim, using descriptive statistics methods, including the test of the relationship between different epizootiological indicators. The spatial dynamic study of BT's occurrence and its spread in the world over the two past decades was presented by different epizootic indicators. The given analysis includes assessment and measurement of risk factors. It was built too, regression models, and allowed to put different forecasts on the different epizootic indicators in the years 2017-2018 by the extrapolation method. We had also determined that, in 2017, BT continues to spread with the total expectancy of 3.4 focus of infection (number of diseased animals in a single unfavorable point) and mortality of about 26 %; these rates tend to decrease in 2018. At abused points by BT, up to 78.4% of animals are mixed (more than one type) and in 21.6% - uniform. By this way, the relative risk of the incidence of appearance-abused points in mixed households has 3.64, which might be considered higher for the BT dissemination. Moreover, between the enzootic index and other epizootiological indicators had revealed an inverse correlation, i.e., to an increase in the level of enzootic index among the cattle population would be formed population less sensitive to BT. Cluster analysis was done, which had demonstrated the zoning of risk levels in the world and the occurrence of the disease intensity in the period 1996-2016 years. Then, assess connection degree of the dynamic of BT tension with geographical and socioeconomic conditions background 0.66 and 0.68, respectively. It is important to define a variety of BT risk factors and assess their influence on BT occurrence. However, the most important is to define the overlapping coinfluence between them that cause serious losses. To have an out of BT territory needs to make an emphasis of co-influence of risk factors on this zone. Was predicted a continue hits of disease in the next year with weight moderation through one year. Far from statists, to assess the given forecast may have a serious variety, taken in account problems of actual climate change in the world.

  15. Statistical variability and confidence intervals for planar dose QA pass rates

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bailey, Daniel W.; Nelms, Benjamin E.; Attwood, Kristopher

    Purpose: The most common metric for comparing measured to calculated dose, such as for pretreatment quality assurance of intensity-modulated photon fields, is a pass rate (%) generated using percent difference (%Diff), distance-to-agreement (DTA), or some combination of the two (e.g., gamma evaluation). For many dosimeters, the grid of analyzed points corresponds to an array with a low areal density of point detectors. In these cases, the pass rates for any given comparison criteria are not absolute but exhibit statistical variability that is a function, in part, on the detector sampling geometry. In this work, the authors analyze the statistics ofmore » various methods commonly used to calculate pass rates and propose methods for establishing confidence intervals for pass rates obtained with low-density arrays. Methods: Dose planes were acquired for 25 prostate and 79 head and neck intensity-modulated fields via diode array and electronic portal imaging device (EPID), and matching calculated dose planes were created via a commercial treatment planning system. Pass rates for each dose plane pair (both centered to the beam central axis) were calculated with several common comparison methods: %Diff/DTA composite analysis and gamma evaluation, using absolute dose comparison with both local and global normalization. Specialized software was designed to selectively sample the measured EPID response (very high data density) down to discrete points to simulate low-density measurements. The software was used to realign the simulated detector grid at many simulated positions with respect to the beam central axis, thereby altering the low-density sampled grid. Simulations were repeated with 100 positional iterations using a 1 detector/cm{sup 2} uniform grid, a 2 detector/cm{sup 2} uniform grid, and similar random detector grids. For each simulation, %/DTA composite pass rates were calculated with various %Diff/DTA criteria and for both local and global %Diff normalization techniques. Results: For the prostate and head/neck cases studied, the pass rates obtained with gamma analysis of high density dose planes were 2%-5% higher than respective %/DTA composite analysis on average (ranging as high as 11%), depending on tolerances and normalization. Meanwhile, the pass rates obtained via local normalization were 2%-12% lower than with global maximum normalization on average (ranging as high as 27%), depending on tolerances and calculation method. Repositioning of simulated low-density sampled grids leads to a distribution of possible pass rates for each measured/calculated dose plane pair. These distributions can be predicted using a binomial distribution in order to establish confidence intervals that depend largely on the sampling density and the observed pass rate (i.e., the degree of difference between measured and calculated dose). These results can be extended to apply to 3D arrays of detectors, as well. Conclusions: Dose plane QA analysis can be greatly affected by choice of calculation metric and user-defined parameters, and so all pass rates should be reported with a complete description of calculation method. Pass rates for low-density arrays are subject to statistical uncertainty (vs. the high-density pass rate), but these sampling errors can be modeled using statistical confidence intervals derived from the sampled pass rate and detector density. Thus, pass rates for low-density array measurements should be accompanied by a confidence interval indicating the uncertainty of each pass rate.« less

  16. Subjective Psychological Well-Being in Families with Blind Children: How Can We Improve It?

    PubMed Central

    Sola-Carmona, Juan J.; Lopez-Liria, Remedios; Padilla-Gongora, David; Daza, María T.; Aguilar-Parra, Jose M.

    2016-01-01

    The aim of this work was to examine family well-being in a sample of Spanish families with blind children. Sixty-one participants reported their perceived economic status, the level of job satisfaction, and state-anxiety symptoms. The participants of our study scored higher on state-anxiety and lower on material well-being than the normative sample, although these differences did not reach statistical significance. They also scored higher on job satisfaction and family satisfaction than the general population. A negative correlation was found between state-anxiety and material well-being (r = - 0.62, p = 0.001) and between state-anxiety and family satisfaction (r = - 0.57, p = 0.001). A positive correlation was found between material well-being and job satisfaction (r = 0.40, p = 0.001), and between material well-being and family satisfaction (r = 0.41, p = 0.001). Higher levels of material well-being, job satisfaction, and family satisfaction were associated with lower levels of anxiety in these families. However, no statistically significant correlation was found between family satisfaction and job satisfaction. Our results suggest that the family experience of having a disabled child is evolving, and this implies achieving greater job and family satisfaction than the normative samples, although anxiety scores continue to be higher and material well-being scores remain lower. On the whole, our results confirm that it is necessary to provide these families with more economic resources, which would have a positive impact on their subjective psychological well-being, decreasing their state-anxiety, and increasing their satisfaction with life. PMID:27092095

  17. The art and science of choosing efficacy endpoints for rare disease clinical trials.

    PubMed

    Cox, Gerald F

    2018-04-01

    An important challenge in rare disease clinical trials is to demonstrate a clinically meaningful and statistically significant response to treatment. Selecting the most appropriate and sensitive efficacy endpoints for a treatment trial is part art and part science. The types of endpoints should align with the stage of development (e.g., proof of concept vs. confirmation of clinical efficacy). The patient characteristics and disease stage should reflect the treatment goal of improving disease manifestations or preventing disease progression. For rare diseases, regulatory approval requires demonstration of clinical benefit, defined as how a patient, feels, functions, or survives, in at least one adequate and well-controlled pivotal study conducted according to Good Clinical Practice. In some cases, full regulatory approval can occur using a validated surrogate biomarker, while accelerated, or provisional, approval can occur using a biomarker that is likely to predict clinical benefit. Rare disease studies are small by necessity and require the use of endpoints with large effect sizes to demonstrate statistical significance. Understanding the quantitative factors that determine effect size and its impact on powering the study with an adequate sample size is key to the successful choice of endpoints. Interpreting the clinical meaningfulness of an observed change in an efficacy endpoint can be justified by statistical methods, regulatory precedence, and clinical context. Heterogeneous diseases that affect multiple organ systems may be better accommodated by endpoints that assess mean change across multiple endpoints within the same patient rather than mean change in an individual endpoint across all patients. © 2018 Wiley Periodicals, Inc.

  18. Hydrogeochemical processes and isotopes analysis. Study case: "La Línea Tunnel", Colombia

    NASA Astrophysics Data System (ADS)

    Piña, Adriana; Donado, Leonardo; Cramer, Thomas

    2017-04-01

    Hydrogeochemical and stable isotopes analyses have been widely used to identify recharge and discharge zones, flowpaths, type, origin and age of water, chemical processes between minerals and groundwater as well as effects caused by anthropogenic or natural pollution. In this paper we analyze the interactions between groundwater and surface water using as laboratory the tunnels located at the La Línea Massif in the Cordillera Central of the Colombian Andes. The massif is formed by two igneous-metamorphic fractured complexes (Cajamarca and Quebradagrande group) plus andesithic porphyry rocks from the tertiary period. There, eight main fault zones related to surface creeks were identified and main inflows inside the tunnels were reported. 60 water samples were collected in surface and inside the tunnel in fault zones in two different years, 2010 and 2015. To classify water samples, a multivariate statistical analysis combining Factor Analysis (FA) with Hierarchical Cluster Analysis (HCA) was performed. Then, analyses of the major chemical elements and water isotopes (18O, 2H and 3H) were used to define the origin of dissolved components and to analyse the evolution in time. Most samples were classified as bicarbonate calcite water or bicarbonate magnesium water type. Isotopic analyses show a characteristic behavior for east and west watershed and each geologic group. According to the FA and HCA, obtained factors and clusters are first related to the location of the samples (surface or tunnel samples) followed by the geology. Surface samples behave according to the Colombian meteoric line as inflows related to permeable faults while less permeable faults show hydrothermal processes. Finally, water evolution in time shows a decrease of pH, conductivity and Mg2+ related to silicate weathering or precipitation/dissolution processes that affect the spacing in fractures and consequently, the hydraulic properties.

  19. OpenMSI Arrayed Analysis Toolkit: Analyzing Spatially Defined Samples Using Mass Spectrometry Imaging

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    de Raad, Markus; de Rond, Tristan; Rübel, Oliver

    Mass spectrometry imaging (MSI) has primarily been applied in localizing biomolecules within biological matrices. Although well-suited, the application of MSI for comparing thousands of spatially defined spotted samples has been limited. One reason for this is a lack of suitable and accessible data processing tools for the analysis of large arrayed MSI sample sets. In this paper, the OpenMSI Arrayed Analysis Toolkit (OMAAT) is a software package that addresses the challenges of analyzing spatially defined samples in MSI data sets. OMAAT is written in Python and is integrated with OpenMSI (http://openmsi.nersc.gov), a platform for storing, sharing, and analyzing MSI data.more » By using a web-based python notebook (Jupyter), OMAAT is accessible to anyone without programming experience yet allows experienced users to leverage all features. OMAAT was evaluated by analyzing an MSI data set of a high-throughput glycoside hydrolase activity screen comprising 384 samples arrayed onto a NIMS surface at a 450 μm spacing, decreasing analysis time >100-fold while maintaining robust spot-finding. The utility of OMAAT was demonstrated for screening metabolic activities of different sized soil particles, including hydrolysis of sugars, revealing a pattern of size dependent activities. Finally, these results introduce OMAAT as an effective toolkit for analyzing spatially defined samples in MSI. OMAAT runs on all major operating systems, and the source code can be obtained from the following GitHub repository: https://github.com/biorack/omaat.« less

  20. The Savannah River Site`s groundwater monitoring program. Third quarter 1990

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1991-05-06

    The Environmental Protection Department/Environmental Monitoring Section (EPD/EMS) administers the Savannah River Site`s (SRS) Groundwater Monitoring Program. During third quarter 1990 (July through September) EPD/EMS conducted routine sampling of monitoring wells and drinking water locations. EPD/EMS established two sets of flagging criteria in 1986 to assist in the management of sample results. The flagging criteria do not define contamination levels; instead they aid personnel in sample scheduling, interpretation of data, and trend identification. The flagging criteria are based on detection limits, background levels in SRS groundwater, and drinking water standards. All analytical results from third quarter 1990 are listed in thismore » report, which is distributed to all site custodians. One or more analytes exceeded Flag 2 in 87 monitoring well series. Analytes exceeded Flat 2 for the first since 1984 in 14 monitoring well series. In addition to groundwater monitoring, EPD/EMS collected drinking water samples from SRS drinking water systems supplied by wells. The drinking water samples were analyzed for radioactive constituents.« less

  1. The Savannah River Site's groundwater monitoring program

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1991-05-06

    The Environmental Protection Department/Environmental Monitoring Section (EPD/EMS) administers the Savannah River Site's (SRS) Groundwater Monitoring Program. During third quarter 1990 (July through September) EPD/EMS conducted routine sampling of monitoring wells and drinking water locations. EPD/EMS established two sets of flagging criteria in 1986 to assist in the management of sample results. The flagging criteria do not define contamination levels; instead they aid personnel in sample scheduling, interpretation of data, and trend identification. The flagging criteria are based on detection limits, background levels in SRS groundwater, and drinking water standards. All analytical results from third quarter 1990 are listed in thismore » report, which is distributed to all site custodians. One or more analytes exceeded Flag 2 in 87 monitoring well series. Analytes exceeded Flat 2 for the first since 1984 in 14 monitoring well series. In addition to groundwater monitoring, EPD/EMS collected drinking water samples from SRS drinking water systems supplied by wells. The drinking water samples were analyzed for radioactive constituents.« less

  2. Sampling Long- versus Short-Range Interactions Defines the Ability of Force Fields To Reproduce the Dynamics of Intrinsically Disordered Proteins.

    PubMed

    Mercadante, Davide; Wagner, Johannes A; Aramburu, Iker V; Lemke, Edward A; Gräter, Frauke

    2017-09-12

    Molecular dynamics (MD) simulations have valuably complemented experiments describing the dynamics of intrinsically disordered proteins (IDPs), particularly since the proposal of models to solve the artificial collapse of IDPs in silico. Such models suggest redefining nonbonded interactions, by either increasing water dispersion forces or adopting the Kirkwood-Buff force field. These approaches yield extended conformers that better comply with experiments, but it is unclear if they all sample the same intrachain dynamics of IDPs. We have tested this by employing MD simulations and single-molecule Förster resonance energy transfer spectroscopy to sample the dimensions of systems with different sequence compositions, namely strong and weak polyelectrolytes. For strong polyelectrolytes in which charge effects dominate, all the proposed solutions equally reproduce the expected ensemble's dimensions. For weak polyelectrolytes, at lower cutoffs, force fields abnormally alter intrachain dynamics, overestimating excluded volume over chain flexibility or reporting no difference between the dynamics of different chains. The TIP4PD water model alone can reproduce experimentally observed changes in extensions (dimensions), but not quantitatively and with only weak statistical significance. Force field limitations are reversed with increased interaction cutoffs, showing that chain dynamics are critically defined by the presence of long-range interactions. Force field analysis aside, our study provides the first insights into how long-range interactions critically define IDP dimensions and raises the question of which length range is crucial to correctly sample the overall dimensions and internal dynamics of the large group of weakly charged yet highly polar IDPs.

  3. Mechanical properties of silicate glasses exposed to a low-Earth orbit

    NASA Technical Reports Server (NTRS)

    Wiedlocher, David E.; Tucker, Dennis S.; Nichols, Ron; Kinser, Donald L.

    1992-01-01

    The effects of a 5.8 year exposure to low earth orbit environment upon the mechanical properties of commercial optical fused silica, low iron soda-lime-silica, Pyrex 7740, Vycor 7913, BK-7, and the glass ceramic Zerodur were examined. Mechanical testing employed the ASTM-F-394 piston on 3-ball method in a liquid nitrogen environment. Samples were exposed on the Long Duration Exposure Facility (LDEF) in two locations. Impacts were observed on all specimens except Vycor. Weibull analysis as well as a standard statistical evaluation were conducted. The Weibull analysis revealed no differences between control samples and the two exposed samples. We thus concluded that radiation components of the Earth orbital environment did not degrade the mechanical strength of the samples examined within the limits of experimental error. The upper bound of strength degradation for meteorite impacted samples based upon statistical analysis and observation was 50 percent.

  4. Statistical analysis of major ion and trace element geochemistry of water, 1986-2006, at seven wells transecting the freshwater/saline-water interface of the Edwards Aquifer, San Antonio, Texas

    USGS Publications Warehouse

    Mahler, Barbara J.

    2008-01-01

    The statistical analyses taken together indicate that the geochemistry at the freshwater-zone wells is more variable than that at the transition-zone wells. The geochemical variability at the freshwater-zone wells might result from dilution of ground water by meteoric water. This is indicated by relatively constant major ion molar ratios; a preponderance of positive correlations between SC, major ions, and trace elements; and a principal components analysis in which the major ions are strongly loaded on the first principal component. Much of the variability at three of the four transition-zone wells might result from the use of different laboratory analytical methods or reporting procedures during the period of sampling. This is reflected by a lack of correlation between SC and major ion concentrations at the transition-zone wells and by a principal components analysis in which the variability is fairly evenly distributed across several principal components. The statistical analyses further indicate that, although the transition-zone wells are less well connected to surficial hydrologic conditions than the freshwater-zone wells, there is some connection but the response time is longer. 

  5. Implementation and Testing of Turbulence Models for the F18-HARV Simulation

    NASA Technical Reports Server (NTRS)

    Yeager, Jessie C.

    1998-01-01

    This report presents three methods of implementing the Dryden power spectral density model for atmospheric turbulence. Included are the equations which define the three methods and computer source code written in Advanced Continuous Simulation Language to implement the equations. Time-history plots and sample statistics of simulated turbulence results from executing the code in a test program are also presented. Power spectral densities were computed for sample sequences of turbulence and are plotted for comparison with the Dryden spectra. The three model implementations were installed in a nonlinear six-degree-of-freedom simulation of the High Alpha Research Vehicle airplane. Aircraft simulation responses to turbulence generated with the three implementations are presented as plots.

  6. Selective transport of palynomorphs in marine turbiditic deposits: An example from the Ascension-Monterey Canyon system offshore central California

    USGS Publications Warehouse

    McGann, Mary

    2017-01-01

    The pollen assemblage of a deep-sea core (15G) collected at lower bathyal depths (3491 m) on a levee of Monterey Canyon off central California was investigated to gain insights into the delivery processes of terrigenous material to submarine fans and the effect this transport has on the palynological record. Thirty-two samples were obtained down the length of the core, 19 from hemipelagic and mixed mud deposits considered to be the background record, and 13 others from displaced flow deposits. The pollen record obtained from the background samples documents variations in the terrestrial flora as it adapted to changing climatic conditions over the last 19,000 cal yrs BP. A Q-mode cluster analysis defined three pollen zones: a Glacial Pollen Zone (ca. 20,000–17,000 cal yr BP), an overlying Transitional Pollen Zone (ca. 17,000–11,500 cal yr BP), and an Interglacial Pollen Zone (ca. 11,500 cal yr BP to present). Another Q-mode cluster analysis, of both the background mud and flow deposits, also defined these three pollen zones, but four of the 13 turbiditic deposits were assigned to pollen zones older than expected by their stratigraphic position. This was due to these samples containing statistically significant fewer palynomorphs than the background muds as well as being enriched (∼10–35% in some cases) in hydraulically-efficient Pinus pollen. A selective bias in the pollen assemblage, such as demonstrated here, may result in incorrect interpretations (e.g., climatic shifts or environmental perturbations) based on the floral record, indicating turbiditic deposits should be avoided in marine palynological studies. Particularly in the case of fine-grained flow deposits that may not be visually distinct, granulometry and grain size frequency distribution curves may not be enough to identify these biased deposits. Determining the relative abundance and source of displaced shallow-water benthic foraminifera entrained in these sediments serves as an excellent additional tool to do so.

  7. Miltenberger blood group typing by real-time polymerase chain reaction (qPCR) melting curve analysis in Thai population.

    PubMed

    Vongsakulyanon, A; Kitpoka, P; Kunakorn, M; Srikhirin, T

    2015-12-01

    To develop reliable and convenient methods for Miltenberger (Mi(a) ) blood group typing. To apply real-time polymerase chain reaction (qPCR) melting curve analysis to Mi(a) blood group typing. The Mi(a) blood group is the collective set of glycophorin hybrids in the MNS blood group system. Mi(a+) blood is common among East Asians and is also found in the Thai population. Incompatible Mi(a) blood transfusions pose the risk of life-threatening haemolysis; therefore, Mi(a) blood group typing is necessary in ethnicities where the Mi(a) blood group is prevalent. One hundred and forty-three blood samples from Thai blood donors were used in the study. The samples included 50 Mi(a+) samples and 93 Mi(a-) samples, which were defined by serology. The samples were typed by Mi(a) typing qPCR, and 50 Mi(a+) samples were sequenced to identify the Mi(a) subtypes. Mi(a) subtyping qPCR was performed to define GP.Mur. Both Mi(a) typing and Mi(a) subtyping were tested on a conventional PCR platform. The results of Mi(a) typing qPCR were all concordant with serology. Sequencing of the 50 Mi(a+) samples revealed 47 GP.Mur samples and 3 GP.Hop or Bun samples. Mi(a) subtyping qPCR was the supplementary test used to further define GP.Mur from other Mi(a) subtypes. Both Mi(a) typing and Mi(a) subtyping performed well using a conventional PCR platform. Mi(a) typing qPCR correctly identified Mi(a) blood groups in a Thai population with the feasibility of Mi(a) subtype discrimination, and Mi(a) subtyping qPCR was able to further define GP.Mur from other Mi(a) subtypes. © 2015 British Blood Transfusion Society.

  8. Ichthyoplankton assemblages in the Gulf of Nicoya and Golfo Dulce embayments, Pacific coast of Costa Rica.

    PubMed

    Molina-Ureña, H

    1996-12-01

    Ichthyoplankton surveys were conducted in December (rainy season), 1993 and February (dry season), 1994, during the RV Victor Hensen German-Costa Rican Expedition to the Gulf of Nicoya and Gulfo Dulce, Costa Rica. Samples from the inner, central, and outer areas of each gulf were collected in oblique tows with a bongo net of 0.6 m mouth diameter, 2.5 m long and 1000-micron mesh. A total of 416 fish larvae of 22 families were sorted out of 14 samples. Stations of both the maximum (11) and the minimum (1) family richness were located in Golfo Dulce. Mean total larval abundances were 124.9 and 197.2 individuals 10 m-2 for the Gulf of Nicoya and Golfo Dulce, respectively, while mean larval densities ranged from 95.3 larvae 10 m-2 in December to 236.7 larvae 10 m-2 in February. However, no statistical differences between gulfs or seasons were detected, due to the high within-group variability. Cluster Analysis, Multi-Dimensional Scaling (MDS), and non-parametric tests showed two well-defined major groups: (1) the Gulf of Nicoya neritic assemblage, represented by Engraulids, Sciaenids, and Gobiids (inner and central stations), and (2) the oceanic assemblage, dominated by Myctophids, Bregmacerotids, Ophiidids, and Trichiurids (outer stations off the Gulf of Nicoya and Golfo Dulce). A third, although less defined group, was an Ophichthid-dominated assemblage (typical in areas nearby coral or rocky reefs). These assemblages closely resemble the clusters based upon adult fish data of the beamtrawl catches of the same cruise. This publication is the first to report on the ichthyoplankton community of Golfo Dulce.

  9. Data-optimized source modeling with the Backwards Liouville Test–Kinetic method

    DOE PAGES

    Woodroffe, J. R.; Brito, T. V.; Jordanova, V. K.; ...

    2017-09-14

    In the standard practice of neutron multiplicity counting , the first three sampled factorial moments of the event triggered neutron count distribution were used to quantify the three main neutron source terms: the spontaneous fissile material effective mass, the relative (α,n) production and the induced fission source responsible for multiplication. Our study compares three methods to quantify the statistical uncertainty of the estimated mass: the bootstrap method, propagation of variance through moments, and statistical analysis of cycle data method. Each of the three methods was implemented on a set of four different NMC measurements, held at the JRC-laboratory in Ispra,more » Italy, sampling four different Pu samples in a standard Plutonium Scrap Multiplicity Counter (PSMC) well counter.« less

  10. Estimating the mass variance in neutron multiplicity counting-A comparison of approaches

    NASA Astrophysics Data System (ADS)

    Dubi, C.; Croft, S.; Favalli, A.; Ocherashvili, A.; Pedersen, B.

    2017-12-01

    In the standard practice of neutron multiplicity counting , the first three sampled factorial moments of the event triggered neutron count distribution are used to quantify the three main neutron source terms: the spontaneous fissile material effective mass, the relative (α , n) production and the induced fission source responsible for multiplication. This study compares three methods to quantify the statistical uncertainty of the estimated mass: the bootstrap method, propagation of variance through moments, and statistical analysis of cycle data method. Each of the three methods was implemented on a set of four different NMC measurements, held at the JRC-laboratory in Ispra, Italy, sampling four different Pu samples in a standard Plutonium Scrap Multiplicity Counter (PSMC) well counter.

  11. Estimating the mass variance in neutron multiplicity counting $-$ A comparison of approaches

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dubi, C.; Croft, S.; Favalli, A.

    In the standard practice of neutron multiplicity counting, the first three sampled factorial moments of the event triggered neutron count distribution are used to quantify the three main neutron source terms: the spontaneous fissile material effective mass, the relative (α,n) production and the induced fission source responsible for multiplication. This study compares three methods to quantify the statistical uncertainty of the estimated mass: the bootstrap method, propagation of variance through moments, and statistical analysis of cycle data method. Each of the three methods was implemented on a set of four different NMC measurements, held at the JRC-laboratory in Ispra, Italy,more » sampling four different Pu samples in a standard Plutonium Scrap Multiplicity Counter (PSMC) well counter.« less

  12. Estimating the mass variance in neutron multiplicity counting $-$ A comparison of approaches

    DOE PAGES

    Dubi, C.; Croft, S.; Favalli, A.; ...

    2017-09-14

    In the standard practice of neutron multiplicity counting, the first three sampled factorial moments of the event triggered neutron count distribution are used to quantify the three main neutron source terms: the spontaneous fissile material effective mass, the relative (α,n) production and the induced fission source responsible for multiplication. This study compares three methods to quantify the statistical uncertainty of the estimated mass: the bootstrap method, propagation of variance through moments, and statistical analysis of cycle data method. Each of the three methods was implemented on a set of four different NMC measurements, held at the JRC-laboratory in Ispra, Italy,more » sampling four different Pu samples in a standard Plutonium Scrap Multiplicity Counter (PSMC) well counter.« less

  13. Groundwater quality in Geauga County, Ohio: status, including detection frequency of methane in water wells, 2009, and changes during 1978-2009

    USGS Publications Warehouse

    Jagucki, Martha L.; Kula, Stephanie P.; Mailot, Brian E.

    2015-01-01

    To evaluate whether constituent concentrations consistently increased or decreased over time, the strength of the association between sampling year (time) and constituent concentration was statistically evaluated for 116 water-quality samples collected by the USGS in 1978, 1980, 1986, 1999, and 2009 from a total of 65 wells across the county (generally domestic wells or wells serving small businesses or churches). Results indicate that many of the constituents that have been analyzed for decades exhibited no consistent temporal trends at a statistically significant level (p-value less than 0.05); fluctuations in concentrations of these constituents represent natural variation in groundwater quality. Dissolved oxygen, calcium, and sulfate concentrations and chloride:bromide ratios increased over time in one or more aquifers, while pH and concentrations of bromide and dissolved organic carbon decreased over time. Detections of total coliform bacteria and nitrate did not become more frequent from 1986 to 2009, even though potential sources of these constituents, such as number of septic systems (linked to population) and percent developed land in the county, increased during this period.

  14. Investigating the Cross-Cultural Validity of "DSM-5" Autism Spectrum Disorder: Evidence from Finnish and UK Samples

    ERIC Educational Resources Information Center

    Mandy, William; Charman, Tony; Puura, Kaija; Skuse, David

    2014-01-01

    The recent "Diagnostic and Statistical Manual of Mental Disorders-Fifth Edition" ("DSM-5") reformulation of autism spectrum disorder has received empirical support from North American and UK samples. Autism spectrum disorder is an increasingly global diagnosis, and research is needed to discover how well it generalises beyond…

  15. Liebowitz Social Anxiety Scale (LSAS): Optimal cut points for remission and response in a German sample.

    PubMed

    von Glischinski, M; Willutzki, U; Stangier, U; Hiller, W; Hoyer, J; Leibing, E; Leichsenring, F; Hirschfeld, G

    2018-02-11

    The Liebowitz Social Anxiety Scale (LSAS) is the most frequently used instrument to assess social anxiety disorder (SAD) in clinical research and practice. Both a self-reported (LSAS-SR) and a clinician-administered (LSAS-CA) version are available. The aim of the present study was to define optimal cut-off (OC) scores for remission and response to treatment for the LSAS in a German sample. Data of N = 311 patients with SAD were used who had completed psychotherapeutic treatment within a multicentre randomized controlled trial. Diagnosis of SAD and reduction in symptom severity according to the Structured Clinical Interview for Diagnostic and Statistical Manual of Mental Disorders, 4th edition, served as gold standard. OCs yielding the best balance between sensitivity and specificity were determined using receiver operating characteristics. The variability of the resulting OCs was estimated by nonparametric bootstrapping. Using diagnosis of SAD (present vs. absent) as a criterion, results for remission indicated cut-off values of 35 for the LSAS-SR and 30 for the LSAS-CA, with acceptable sensitivity (LSAS-SR: .83, LSAS-CA: .88) and specificity (LSAS-SR: .82, LSAS-CA: .87). For detection of response to treatment, assessed by a 1-point reduction in the Structured Clinical Interview for Diagnostic and Statistical Manual of Mental Disorders, 4th edition, rating, a reduction of 28% for the LSAS-SR and 29% for the LSAS-CA yielded the best balance between sensitivity (LSAS-SR: .75, LSAS-CA: .83) and specificity (LSAS-SR: .76, LSAS-CA: .80). To our knowledge, we are the first to define cut points for the LSAS in a German sample. Overall, the cut points for remission and response corroborate previously reported cut points, now building on a broader data basis. Copyright © 2018 John Wiley & Sons, Ltd.

  16. Skill of ship-following large-eddy simulations in reproducing MAGIC observations across the northeast Pacific stratocumulus to cumulus transition region

    DOE PAGES

    McGibbon, J.; Bretherton, C. S.

    2017-03-17

    During the Marine ARM GPCI Investigation of Clouds (MAGIC) in October 2011 to September 2012, a container ship making periodic cruises between Los Angeles, CA, and Honolulu, HI, was instrumented with surface meteorological, aerosol and radiation instruments, a cloud radar and ceilometer, and radiosondes. Here large-eddy simulation (LES) is performed in a ship-following frame of reference for 13 four day transects from the MAGIC field campaign. The goal is to assess if LES can skillfully simulate the broad range of observed cloud characteristics and boundary layer structure across the subtropical stratocumulus to cumulus transition region sampled during different seasons andmore » meteorological conditions. Results from Leg 15A, which sampled a particularly well-defined stratocumulus to cumulus transition, demonstrate the approach. The LES reproduces the observed timing of decoupling and transition from stratocumulus to cumulus and matches the observed evolution of boundary layer structure, cloud fraction, liquid water path, and precipitation statistics remarkably well. Considering the simulations of all 13 cruises, the LES skillfully simulates the mean diurnal variation of key measured quantities, including liquid water path (LWP), cloud fraction, measures of decoupling, and cloud radar-derived precipitation. The daily mean quantities are well represented, and daily mean LWP and cloud fraction show the expected correlation with estimated inversion strength. There is a –0.6 K low bias in LES near-surface air temperature that results in a high bias of 5.6 W m –2 in sensible heat flux (SHF). Altogether, these results build confidence in the ability of LES to represent the northeast Pacific stratocumulus to trade cumulus transition region.« less

  17. A suite of MATLAB-based computational tools for automated analysis of COPAS Biosort data

    PubMed Central

    Morton, Elizabeth; Lamitina, Todd

    2010-01-01

    Complex Object Parametric Analyzer and Sorter (COPAS) devices are large-object, fluorescence-capable flow cytometers used for high-throughput analysis of live model organisms, including Drosophila melanogaster, Caenorhabditis elegans, and zebrafish. The COPAS is especially useful in C. elegans high-throughput genome-wide RNA interference (RNAi) screens that utilize fluorescent reporters. However, analysis of data from such screens is relatively labor-intensive and time-consuming. Currently, there are no computational tools available to facilitate high-throughput analysis of COPAS data. We used MATLAB to develop algorithms (COPAquant, COPAmulti, and COPAcompare) to analyze different types of COPAS data. COPAquant reads single-sample files, filters and extracts values and value ratios for each file, and then returns a summary of the data. COPAmulti reads 96-well autosampling files generated with the ReFLX adapter, performs sample filtering, graphs features across both wells and plates, performs some common statistical measures for hit identification, and outputs results in graphical formats. COPAcompare performs a correlation analysis between replicate 96-well plates. For many parameters, thresholds may be defined through a simple graphical user interface (GUI), allowing our algorithms to meet a variety of screening applications. In a screen for regulators of stress-inducible GFP expression, COPAquant dramatically accelerated data analysis and allowed us to rapidly move from raw data to hit identification. Because the COPAS file structure is standardized and our MATLAB code is freely available, our algorithms should be extremely useful for analysis of COPAS data from multiple platforms and organisms. The MATLAB code is freely available at our web site (www.med.upenn.edu/lamitinalab/downloads.shtml). PMID:20569218

  18. Investigating the cross-cultural validity of DSM-5 autism spectrum disorder: evidence from Finnish and UK samples.

    PubMed

    Mandy, William; Charman, Tony; Puura, Kaija; Skuse, David

    2014-01-01

    The recent Diagnostic and Statistical Manual of Mental Disorders-Fifth Edition (DSM-5) reformulation of autism spectrum disorder has received empirical support from North American and UK samples. Autism spectrum disorder is an increasingly global diagnosis, and research is needed to discover how well it generalises beyond North America and the United Kingdom. We tested the applicability of the DSM-5 model to a sample of Finnish young people with autism spectrum disorder (n = 130) or the broader autism phenotype (n = 110). Confirmatory factor analysis tested the DSM-5 model in Finland and compared the fit of this model between Finnish and UK participants (autism spectrum disorder, n = 488; broader autism phenotype, n = 220). In both countries, autistic symptoms were measured using the Developmental, Diagnostic and Dimensional Interview. Replicating findings from English-speaking samples, the DSM-5 model fitted well in Finnish autism spectrum disorder participants, outperforming a Diagnostic and Statistical Manual of Mental Disorders-Fourth Edition (DSM-IV) model. The DSM-5 model fitted equally well in Finnish and UK autism spectrum disorder samples. Among broader autism phenotype participants, this model fitted well in the United Kingdom but poorly in Finland, suggesting that cross-cultural variability may be greatest for milder autistic characteristics. We encourage researchers with data from other cultures to emulate our methodological approach, to map any cultural variability in the manifestation of autism spectrum disorder and the broader autism phenotype. This would be especially valuable given the ongoing revision of the International Classification of Diseases-11th Edition, the most global of the diagnostic manuals.

  19. Conceptual Model of Clinical Governance Information System for Statistical Indicators by Using UML in Two Sample Hospitals.

    PubMed

    Jeddi, Fatemeh Rangraz; Farzandipoor, Mehrdad; Arabfard, Masoud; Hosseini, Azam Haj Mohammad

    2014-04-01

    The purpose of this study was investigating situation and presenting a conceptual model for clinical governance information system by using UML in two sample hospitals. However, use of information is one of the fundamental components of clinical governance; but unfortunately, it does not pay much attention to information management. A cross sectional study was conducted in October 2012- May 2013. Data were gathered through questionnaires and interviews in two sample hospitals. Face and content validity of the questionnaire has been confirmed by experts. Data were collected from a pilot hospital and reforms were carried out and Final questionnaire was prepared. Data were analyzed by descriptive statistics and SPSS 16 software. With the scenario derived from questionnaires, UML diagrams are presented by using Rational Rose 7 software. The results showed that 32.14 percent Indicators of the hospitals were calculated. Database was not designed and 100 percent of the hospital's clinical governance was required to create a database. Clinical governance unit of hospitals to perform its mission, do not have access to all the needed indicators. Defining of Processes and drawing of models and creating of database are essential for designing of information systems.

  20. Conceptual Model of Clinical Governance Information System for Statistical Indicators by Using UML in Two Sample Hospitals.

    PubMed

    Jeddi, Fatemeh Rangraz; Farzandipoor, Mehrdad; Arabfard, Masoud; Hosseini, Azam Haj Mohammad

    2016-04-01

    The purpose of this study was investigating situation and presenting a conceptual model for clinical governance information system by using UML in two sample hospitals. However, use of information is one of the fundamental components of clinical governance; but unfortunately, it does not pay much attention to information management. A cross sectional study was conducted in October 2012- May 2013. Data were gathered through questionnaires and interviews in two sample hospitals. Face and content validity of the questionnaire has been confirmed by experts. Data were collected from a pilot hospital and reforms were carried out and Final questionnaire was prepared. Data were analyzed by descriptive statistics and SPSS 16 software. With the scenario derived from questionnaires, UML diagrams are presented by using Rational Rose 7 software. The results showed that 32.14 percent Indicators of the hospitals were calculated. Database was not designed and 100 percent of the hospital's clinical governance was required to create a database. Clinical governance unit of hospitals to perform its mission, do not have access to all the needed indicators. Defining of Processes and drawing of models and creating of database are essential for designing of information systems.

  1. Urban-Rural and Regional Variability in the Prevalence of Food Insecurity: the Survey of the Health of Wisconsin

    PubMed Central

    Guerrero, Natalie; Walsh, Matthew C; Malecki, Kristen C; Nieto, F Javier

    2014-01-01

    Background Food insecurity is a public health concern and it is estimated to affect 18 million American households nationally, which can result in chronic nutritional deficiencies and other health risks. The relationships between food insecurity and specific demographic and geographic factors in Wisconsin is not well documented. The goals of this paper are to investigate socio-demographic and geographic features associated with food insecurity in a representative sample of Wisconsin adults. Methods This study used data from the Survey of the Health of Wisconsin (SHOW). SHOW annually collects health-related data on a representative sample of Wisconsin residents. Between 2008-2012, 2,947 participants were enrolled in the SHOW study. The presence of food insecurity was defined based on the participant's affirmative answer to the question “In the last 12 months, have you been concerned about having enough food for you or your family?” Results After adjustment for age, race, and gender, 13.2% (95% Confidence Limit (CI): 10.8%-15.1%) of participants reported food insecurity, 56.7% (95% CI: 50.6%-62.7%) of whom were female. Food insecurity did not statistically differ by state public health region (p=0.30). The adjusted prevalence of food insecurity in the urban core, other urban, and rural areas of Wisconsin was 14.1%, 6.5% and 10.5%, respectively. These differences were not statistically significant (p=0.13). Conclusions The prevalence of food insecurity is substantial, affecting an estimated number of 740,000 Wisconsin residents. The prevalence was similarly high in all urbanicity levels and across all state public health regions in Wisconsin. Food insecurity is a common problem with potentially serious health consequences affecting populations across the entire state. PMID:25211799

  2. Urban-rural and regional variability in the prevalence of food insecurity: the survey of the health of Wisconsin.

    PubMed

    Guerrero, Natalie; Walsh, Matthew C; Malecki, Kristen C; Nieto, F Javier

    2014-08-01

    Food insecurity is a public health concern estimated to affect 18 million American households nationally, which can result in chronic nutritional deficiencies and other health risks. The relationships between food insecurity and specific demographic and geographic factors in Wisconsin are not well documented. The goals of this paper are to investigate sociodemographic and geographic features associated with food insecurity in a representative sample of Wisconsin adults. This study used data from the Survey of the Health of Wisconsin (SHOW). SHOW annually collects health-related data on a representative sample of Wisconsin residents. Between 2008-2012, 2,947 participants were enrolled in the SHOW study. The presence of food insecurity was defined based on the participant's affirmative answer to the question "In the last 12 months, have you been concerned about having enough food for you or your family?" After adjustment for age, race, and gender, 13.2% (95% CI, 10.8%-15.1%) of participants reported food insecurity, 56.7% (95% CI, 50.6%-62.7%) of whom were female. Food insecurity did not statistically differ by region (P = 0.30). The adjusted prevalence of food insecurity in the urban core, other urban, and rural areas was 14.1%, 6.5%, and 10.5%, respectively. These differences were not statistically significant (P = 0.13) and, for urban core and rural areas, persisted even when accounting for level of economic hardship in the community. The prevalence of food insecurity is substantial, affecting an estimated 740,000 or more Wisconsin residents. The prevalence was similarly high in all urbanicity levels and across all state public health regions in Wisconsin. Food insecurity is a common problem with potentially serious health consequences affecting populations across the entire state.

  3. Religiosity and risky sexual behavior in African-American adolescent females.

    PubMed

    McCree, Donna Hubbard; Wingood, Gina M; DiClemente, Ralph; Davies, Susan; Harrington, Katherine F

    2003-07-01

    To examine the association between religiosity (defined by frequency of engaging in religious/spiritual activities) and African-American adolescent females' sexual behaviors, attitudes toward sex, and ability to negotiate safer sex. Between December 1996 and April 1999, 1130 female adolescents were screened for eligibility in a sexually transmitted disease (STD)/human immunodeficiency virus (HIV) prevention trial. Data collection was achieved through a confidential self-administered questionnaire that examined religiosity and a structured interview regarding sexual behavior. Descriptive statistics were used to characterize the sociodemographics of the sample and logistic regression was used to measure the association between religiosity and the outcome variables. In the study sample (n = 522), 64% of the adolescents had higher religiosity scores based on a 4-item scale (alpha =.68). Results indicate that adolescents who had higher religiosity scores were significantly more likely to have higher self-efficacy in communicating with new, as well as steady male partners about sex; about STDs, HIV, and pregnancy prevention; and in refusing an unsafe sexual encounter. These adolescents were also more likely to have initiated sex at a later age, used a condom in the past 6 months, and possess more positive attitudes toward condom use. Results from this study indicate a relationship between religiosity and sexual behaviors, attitudes toward sex, and ability to negotiate safer sex.

  4. Broad supernatural punishment but not moralizing high gods precede the evolution of political complexity in Austronesia.

    PubMed

    Watts, Joseph; Greenhill, Simon J; Atkinson, Quentin D; Currie, Thomas E; Bulbulia, Joseph; Gray, Russell D

    2015-04-07

    Supernatural belief presents an explanatory challenge to evolutionary theorists-it is both costly and prevalent. One influential functional explanation claims that the imagined threat of supernatural punishment can suppress selfishness and enhance cooperation. Specifically, morally concerned supreme deities or 'moralizing high gods' have been argued to reduce free-riding in large social groups, enabling believers to build the kind of complex societies that define modern humanity. Previous cross-cultural studies claiming to support the MHG hypothesis rely on correlational analyses only and do not correct for the statistical non-independence of sampled cultures. Here we use a Bayesian phylogenetic approach with a sample of 96 Austronesian cultures to test the MHG hypothesis as well as an alternative supernatural punishment hypothesis that allows punishment by a broad range of moralizing agents. We find evidence that broad supernatural punishment drives political complexity, whereas MHGs follow political complexity. We suggest that the concept of MHGs diffused as part of a suite of traits arising from cultural exchange between complex societies. Our results show the power of phylogenetic methods to address long-standing debates about the origins and functions of religion in human society. © 2015 The Author(s) Published by the Royal Society. All rights reserved.

  5. An assessment of the variability in performance of wet atmospheric deposition samplers

    USGS Publications Warehouse

    Graham, R.C.; Robertson, J.K.; Obal, John

    1987-01-01

    The variability in performance of two brands of wet/dry atmospheric deposition samplers were compared for 1 year at a sincle site. A total of nine samplers were used. Samples were collected weekly and analyzed for pH, specific conductance, common chemical constituents, and sample volume. Additionally, data on the duration of each sampler opening were recorded using a microdatalogger. These data disprove the common perception that samplers remain open throughout a precipitation event. The sensitivity of sampler sensors within the range tested did not have a defineable impact on sample collection. The nonnormal distribution within the data set necessitated application of the nonparametric Friedman Test to assess comparability of sample chemical composition and volume between and within sampler brands. Statistically significant differences existed for most comparisons, however the test did not permit quantification of their magnitudes. Differences in analyte concentrations between samplers were small. (USGS)

  6. Interpersonal differentiation within depression diagnosis: relating interpersonal subgroups to symptom load and the quality of the early therapeutic alliance.

    PubMed

    Grosse Holtforth, Martin; Altenstein, David; Krieger, Tobias; Flückiger, Christoph; Wright, Aidan G C; Caspar, Franz

    2014-01-01

    We examined interpersonal problems in psychotherapy outpatients with a principal diagnosis of a depressive disorder in routine care (n=361). These patients were compared to a normative non-clinical sample and to outpatients with other principal diagnoses (n=959). Furthermore, these patients were statistically assigned to interpersonally defined subgroups that were compared regarding symptoms and the quality of the early alliance. The sample of depressive patients reported higher levels of interpersonal problems than the normative sample and the sample of outpatients without a principal diagnosis of depression. Latent Class Analysis identified eight distinct interpersonal subgroups, which differed regarding self-reported symptom load and the quality of the early alliance. However, therapists' alliance ratings did not differentiate between the groups. This interpersonal differentiation within the group of patients with a principal diagnosis of depression may add to a personalized psychotherapy based on interpersonal profiles.

  7. Conformational variety for the ansa chain of rifamycins: Comparison of observed crystal structures and molecular dynamics simulations

    NASA Astrophysics Data System (ADS)

    Bacchi, Alessia; Pelizzi, Giancarlo

    1999-07-01

    The antibiotic activity (via inhibition of DNA-dependent RNA polymerase, DDRP) of rifamycins has been correlated to the conformation of the ansa chain, which can be described by means of 17 torsion angles defined along the ansa backbone. It has been shown that favourable or unfavourable conformations of the ansa chain in rifamycin crystals are generally diagnostic of activity or inactivity against isolated DDRP. The principles of structure correlation suggest that the torsional variety observed in rifamycin crystals should mimic the dynamic flexibility of the ansa chain in solution. Twenty-six crystal structures of rifamycins are grouped into two classes (active and non-active). For each class the variance of the 17 ansa backbone torsion angles is analysed. Active compounds show a well-defined common pattern, while non-active molecules are more scattered, mainly due to steric constraints forcing the molecules into unfavourable conformations. The experimental distributions of torsion angles are compared to the torsional freedom of the ansa chain simulated by molecular dynamics calculations performed at different temperatures and conditions on rifamycin S and rifamycin O, which represent a typical active and a typical sterically constrained molecule, respectively. It is shown that the torsional variety found in the crystalline state samples the dynamic behaviour of the ansa chain for active compounds. The methods of circular statistics are illustrated to describe torsion angle distributions.

  8. General constraints on sampling wildlife on FIA plots

    USGS Publications Warehouse

    Bailey, L.L.; Sauer, J.R.; Nichols, J.D.; Geissler, P.H.; McRoberts, Ronald E.; Reams, Gregory A.; Van Deusen, Paul C.; McWilliams, William H.; Cieszewski, Chris J.

    2005-01-01

    This paper reviews the constraints to sampling wildlife populations at FIA points. Wildlife sampling programs must have well-defined goals and provide information adequate to meet those goals. Investigators should choose a State variable based on information needs and the spatial sampling scale. We discuss estimation-based methods for three State variables: species richness, abundance, and patch occupancy. All methods incorporate two essential sources of variation: detectability estimation and spatial variation. FIA sampling imposes specific space and time criteria that may need to be adjusted to meet local wildlife objectives.

  9. A heteroskedastic error covariance matrix estimator using a first-order conditional autoregressive Markov simulation for deriving asympotical efficient estimates from ecological sampled Anopheles arabiensis aquatic habitat covariates

    PubMed Central

    Jacob, Benjamin G; Griffith, Daniel A; Muturi, Ephantus J; Caamano, Erick X; Githure, John I; Novak, Robert J

    2009-01-01

    Background Autoregressive regression coefficients for Anopheles arabiensis aquatic habitat models are usually assessed using global error techniques and are reported as error covariance matrices. A global statistic, however, will summarize error estimates from multiple habitat locations. This makes it difficult to identify where there are clusters of An. arabiensis aquatic habitats of acceptable prediction. It is therefore useful to conduct some form of spatial error analysis to detect clusters of An. arabiensis aquatic habitats based on uncertainty residuals from individual sampled habitats. In this research, a method of error estimation for spatial simulation models was demonstrated using autocorrelation indices and eigenfunction spatial filters to distinguish among the effects of parameter uncertainty on a stochastic simulation of ecological sampled Anopheles aquatic habitat covariates. A test for diagnostic checking error residuals in an An. arabiensis aquatic habitat model may enable intervention efforts targeting productive habitats clusters, based on larval/pupal productivity, by using the asymptotic distribution of parameter estimates from a residual autocovariance matrix. The models considered in this research extends a normal regression analysis previously considered in the literature. Methods Field and remote-sampled data were collected during July 2006 to December 2007 in Karima rice-village complex in Mwea, Kenya. SAS 9.1.4® was used to explore univariate statistics, correlations, distributions, and to generate global autocorrelation statistics from the ecological sampled datasets. A local autocorrelation index was also generated using spatial covariance parameters (i.e., Moran's Indices) in a SAS/GIS® database. The Moran's statistic was decomposed into orthogonal and uncorrelated synthetic map pattern components using a Poisson model with a gamma-distributed mean (i.e. negative binomial regression). The eigenfunction values from the spatial configuration matrices were then used to define expectations for prior distributions using a Markov chain Monte Carlo (MCMC) algorithm. A set of posterior means were defined in WinBUGS 1.4.3®. After the model had converged, samples from the conditional distributions were used to summarize the posterior distribution of the parameters. Thereafter, a spatial residual trend analyses was used to evaluate variance uncertainty propagation in the model using an autocovariance error matrix. Results By specifying coefficient estimates in a Bayesian framework, the covariate number of tillers was found to be a significant predictor, positively associated with An. arabiensis aquatic habitats. The spatial filter models accounted for approximately 19% redundant locational information in the ecological sampled An. arabiensis aquatic habitat data. In the residual error estimation model there was significant positive autocorrelation (i.e., clustering of habitats in geographic space) based on log-transformed larval/pupal data and the sampled covariate depth of habitat. Conclusion An autocorrelation error covariance matrix and a spatial filter analyses can prioritize mosquito control strategies by providing a computationally attractive and feasible description of variance uncertainty estimates for correctly identifying clusters of prolific An. arabiensis aquatic habitats based on larval/pupal productivity. PMID:19772590

  10. Physico-Chemical and Bacterial Evaluation of Public and Packaged Drinking Water in Vikarabad, Telangana, India - Potential Public Health Implications

    PubMed Central

    Rao, Koppula Yadav; Anjum, Mohammad Shakeel; Reddy, Peddireddy Parthasarathi; Monica, Mocherla; Hameed, Irram Abbass

    2016-01-01

    Introduction Humanity highly depends on water and its proper utilization and management. Water has various uses and its use as thirst quenching fluid is the most significant one. Aim To assess physical, chemical, trace metal and bacterial parameters of various public and packaged drinking water samples collected from villages of Vikarabad mandal. Materials and Methods Public and packaged drinking water samples collected were analysed for various parameters using American Public Health Association (APHA 18th edition 1992) guidelines and the results obtained were compared with bureau of Indian standards for drinking water. Statistical Analysis Descriptive statistics and Pearson’s correlations were done. Results Among bottled water samples, magnesium in 1 sample was >30mg/litre, nickel in 2 samples was >0.02mg/litre. Among sachet water samples, copper in 1 sample was >0.05mg/litre, nickel in 2 samples was >0.02mg/litre. Among canned water samples, total hardness in 1 sample was >200mg/litre, magnesium in 3 samples was >30mg/litre. In tap water sample, calcium was >75mg/litre, magnesium was >30mg/litre, nickel was >0.02mg/litre. Among public bore well water samples, pH in 1 sample was >8.5, total dissolved solids in 17 samples was >500mg/litre, total alkalinity in 9 samples was >200mg/litre, total hardness in 20 samples was >200mg/litre, calcium in 14 samples was >75mg/litre, fluoride in 1 sample was >1mg/litre, magnesium in 14 samples was >30mg/litre. Total coliform was absent in bottled water, sachet water, canned water, tap water samples. Total Coliform was present but E. coli was absent in 4 public bore well water samples. The MPN per 100 ml in those 4 samples of public bore well water was 50. Conclusion Physical, chemical, trace metal and bacterial parameters tested in present study showed values greater than acceptable limit for some samples, which can pose serious threat to consumers of that region. PMID:27437248

  11. Nonlinear analysis of pupillary dynamics.

    PubMed

    Onorati, Francesco; Mainardi, Luca Tommaso; Sirca, Fabiola; Russo, Vincenzo; Barbieri, Riccardo

    2016-02-01

    Pupil size reflects autonomic response to different environmental and behavioral stimuli, and its dynamics have been linked to other autonomic correlates such as cardiac and respiratory rhythms. The aim of this study is to assess the nonlinear characteristics of pupil size of 25 normal subjects who participated in a psychophysiological experimental protocol with four experimental conditions, namely “baseline”, “anger”, “joy”, and “sadness”. Nonlinear measures, such as sample entropy, correlation dimension, and largest Lyapunov exponent, were computed on reconstructed signals of spontaneous fluctuations of pupil dilation. Nonparametric statistical tests were performed on surrogate data to verify that the nonlinear measures are an intrinsic characteristic of the signals. We then developed and applied a piecewise linear regression model to detrended fluctuation analysis (DFA). Two joinpoints and three scaling intervals were identified: slope α0, at slow time scales, represents a persistent nonstationary long-range correlation, whereas α1 and α2, at middle and fast time scales, respectively, represent long-range power-law correlations, similarly to DFA applied to heart rate variability signals. Of the computed complexity measures, α0 showed statistically significant differences among experimental conditions (p<0.001). Our results suggest that (a) pupil size at constant light condition is characterized by nonlinear dynamics, (b) three well-defined and distinct long-memory processes exist at different time scales, and (c) autonomic stimulation is partially reflected in nonlinear dynamics. (c) autonomic stimulation is partially reflected in nonlinear dynamics.

  12. Radon-222 concentrations in ground water and soil gas on Indian reservations in Wisconsin

    USGS Publications Warehouse

    DeWild, John F.; Krohelski, James T.

    1995-01-01

    For sites with wells finished in the sand and gravel aquifer, the coefficient of determination (R2) of the regression of concentration of radon-222 in ground water as a function of well depth is 0.003 and the significance level is 0.32, which indicates that there is not a statistically significant relation between radon-222 concentrations in ground water and well depth. The coefficient of determination of the regression of radon-222 in ground water and soil gas is 0.19 and the root mean square error of the regression line is 271 picocuries per liter. Even though the significance level (0.036) indicates a statistical relation, the root mean square error of the regression is so large that the regression equation would not give reliable predictions. Because of an inadequate number of samples, similar statistical analyses could not be performed for sites with wells finished in the crystalline and sedimentary bedrock aquifers.

  13. The magnetic nature of umbra-penumbra boundary in sunspots

    NASA Astrophysics Data System (ADS)

    Jurčák, J.; Rezaei, R.; González, N. Bello; Schlichenmaier, R.; Vomlel, J.

    2018-03-01

    Context. Sunspots are the longest-known manifestation of solar activity, and their magnetic nature has been known for more than a century. Despite this, the boundary between umbrae and penumbrae, the two fundamental sunspot regions, has hitherto been solely defined by an intensity threshold. Aim. Here, we aim at studying the magnetic nature of umbra-penumbra boundaries in sunspots of different sizes, morphologies, evolutionary stages, and phases of the solar cycle. Methods: We used a sample of 88 scans of the Hinode/SOT spectropolarimeter to infer the magnetic field properties in at the umbral boundaries. We defined these umbra-penumbra boundaries by an intensity threshold and performed a statistical analysis of the magnetic field properties on these boundaries. Results: We statistically prove that the umbra-penumbra boundary in stable sunspots is characterised by an invariant value of the vertical magnetic field component: the vertical component of the magnetic field strength does not depend on the umbra size, its morphology, and phase of the solar cycle. With the statistical Bayesian inference, we find that the strength of the vertical magnetic field component is, with a likelihood of 99%, in the range of 1849-1885 G with the most probable value of 1867 G. In contrast, the magnetic field strength and inclination averaged along individual boundaries are found to be dependent on the umbral size: the larger the umbra, the stronger and more horizontal the magnetic field at its boundary. Conclusions: The umbra and penumbra of sunspots are separated by a boundary that has hitherto been defined by an intensity threshold. We now unveil the empirical law of the magnetic nature of the umbra-penumbra boundary in stable sunspots: it is an invariant vertical component of the magnetic field.

  14. Taking a statistical approach

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wild, M.; Rouhani, S.

    1995-02-01

    A typical site investigation entails extensive sampling and monitoring. In the past, sampling plans have been designed on purely ad hoc bases, leading to significant expenditures and, in some cases, collection of redundant information. In many instances, sampling costs exceed the true worth of the collected data. The US Environmental Protection Agency (EPA) therefore has advocated the use of geostatistics to provide a logical framework for sampling and analysis of environmental data. Geostatistical methodology uses statistical techniques for the spatial analysis of a variety of earth-related data. The use of geostatistics was developed by the mining industry to estimate oremore » concentrations. The same procedure is effective in quantifying environmental contaminants in soils for risk assessments. Unlike classical statistical techniques, geostatistics offers procedures to incorporate the underlying spatial structure of the investigated field. Sample points spaced close together tend to be more similar than samples spaced further apart. This can guide sampling strategies and determine complex contaminant distributions. Geostatistic techniques can be used to evaluate site conditions on the basis of regular, irregular, random and even spatially biased samples. In most environmental investigations, it is desirable to concentrate sampling in areas of known or suspected contamination. The rigorous mathematical procedures of geostatistics allow for accurate estimates at unsampled locations, potentially reducing sampling requirements. The use of geostatistics serves as a decision-aiding and planning tool and can significantly reduce short-term site assessment costs, long-term sampling and monitoring needs, as well as lead to more accurate and realistic remedial design criteria.« less

  15. ALK-FISH borderline cases in non-small cell lung cancer: Implications for diagnostics and clinical decision making.

    PubMed

    von Laffert, Maximilian; Stenzinger, Albrecht; Hummel, Michael; Weichert, Wilko; Lenze, Dido; Warth, Arne; Penzel, Roland; Herbst, Hermann; Kellner, Udo; Jurmeister, Philipp; Schirmacher, Peter; Dietel, Manfred; Klauschen, Frederick

    2015-12-01

    Fluorescence in-situ hybridization (FISH) for the detection of ALK-rearrangements in non-small cell lung cancer (NSCLC) is based on at first sight clear cut-off criteria (≥15% of tumor cells) for split signals (SS) and single red signals (SRS). However, NSCLC with SS-counts around the cut-off may cause interpretation problems. Tissue microarrays containing 753 surgically resected NSCLCs were independently tested for ALK-alterations by FISH and immunohistochemistry (IHC). Our analysis focused on samples with SS/SRS in the range between 10% and 20% (ALK-FISH borderline group). To better understand the role of these samples in routine diagnostics, we performed statistical analyses to systematically estimate the probability of ALK-FISH-misclassification (false negative or positive) for different numbers of evaluated tumor cell nuclei (30, 50, 100, and 200). 94.3% (710/753) of the cases were classified as unequivocally (<10% or ≥20%) ALK-FISH-negative (93%; 700/753) or positive (1.3%; 10/753) and showed concordant IHC results. 5.7% (43/753) of the samples showed SS/SRS between 10% and 20% of the tumor cells. Out of these, 7% (3/43; ALK-FISH: 14%, 18% and 20%) were positive by ALK-IHC, while 93% (40/43) had no detectable expression of the ALK-protein. Statistical analysis showed that ALK-FISH misclassifications occur frequently for samples with rearrangements between 10% and 20% if ALK-characterization is based on a sharp cut-off point (15%). If results in this interval are defined as equivocal (borderline), statistical sampling-related ALK-FISH misclassifications will occur in less than 1% of the cases if 100 tumor cells are evaluated. While ALK status can be determined robustly for the majority of NSCLC by FISH our analysis showed that ∼6% of the cases belong to a borderline group for which ALK-FISH evaluation has only limited reliability due to statistical sampling effects. These cases should be considered equivocal and therapy decisions should include additional tests and clinical considerations. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  16. VOCs, pesticides, nitrate, and their mixtures in groundwater used for drinking water in the United States

    USGS Publications Warehouse

    Squillace, P.J.; Scott, J.C.; Moran, M.J.; Nolan, B.T.; Kolpin, D.W.

    2002-01-01

    Samples of untreated groundwater from 1255 domestic drinking-water wells and 242 public supply wells were analyzed as part of the National Water-Quality Assessment Program of the U.S. Geological Survey between 1992 and 1999. Wells were sampled to define the regional quality of the groundwater resource and, thus, were distributed geographically across large aquifers, primarily in rural areas. For each sample, as many as 60 volatile organic compounds (VOCs), 83 pesticides, and nitrate were analyzed. On the basis of previous studies, nitrate concentrations as nitrogen ≥3 mg/L were considered to have an anthropogenic origin. VOCs were detected more frequently (44%) than pesticides (38%) or anthropogenic nitrate (28%). Seventy percent of the samples contained at least one VOC, pesticide, or anthropogenic nitrate; 47% contained at least two compounds; and 33% contained at least three compounds. The combined concentrations of VOCs and pesticides ranged from about 0.001 to 100 μg/L, with a median of 0.02 μg/L. Water from about 12% of the wells contained one or more compounds that exceeded U.S. Environmental Protection Agency drinking-water standards or human health criteria, primarily because of nitrate concentrations exceeding the maximum contaminant level in domestic wells. A mixture is defined as a unique combination of two or more particular compounds, regardless of the presence of other compounds that may occur in the same sample. There were 100 mixtures (significantly associated with agricultural land use) that had a detection frequency between 2% and 19%. There were 302 mixtures (significantly associated with urban land use) that had a detection frequency between 1% and <2%. Only 14 compounds (seven VOCs, six pesticides, and nitrate) contributed over 95% of the detections in these 402 mixtures; however, most samples with these mixtures also contain a variety of other compounds.

  17. Differentiation of women with premenstrual dysphoric disorder, recurrent brief depression, and healthy controls by daily mood rating dynamics.

    PubMed

    Pincus, Steven M; Schmidt, Peter J; Palladino-Negro, Paula; Rubinow, David R

    2008-04-01

    Enhanced statistical characterization of mood-rating data holds the potential to more precisely classify and sub-classify recurrent mood disorders like premenstrual dysphoric disorder (PMDD) and recurrent brief depressive disorder (RBD). We applied several complementary statistical methods to differentiate mood rating dynamics among women with PMDD, RBD, and normal controls (NC). We compared three subgroups of women: NC (n=8); PMDD (n=15); and RBD (n=9) on the basis of daily self-ratings of sadness, study lengths between 50 and 120 days. We analyzed mean levels; overall variability, SD; sequential irregularity, approximate entropy (ApEn); and a quantification of the extent of brief and staccato dynamics, denoted 'Spikiness'. For each of SD, irregularity (ApEn), and Spikiness, we showed highly significant subgroup differences, ANOVA0.001 for each statistic; additionally, many paired subgroup comparisons showed highly significant differences. In contrast, mean levels were indistinct among the subgroups. For SD, normal controls had much smaller levels than the other subgroups, with RBD intermediate. ApEn showed PMDD to be significantly more regular than the other subgroups. Spikiness showed NC and RBD data sets to be much more staccato than their PMDD counterparts, and appears to suitably characterize the defining feature of RBD dynamics. Compound criteria based on these statistical measures discriminated diagnostic subgroups with high sensitivity and specificity. Taken together, the statistical suite provides well-defined specifications of each subgroup. This can facilitate accurate diagnosis, and augment the prediction and evaluation of response to treatment. The statistical methodologies have broad and direct applicability to behavioral studies for many psychiatric disorders, and indeed to similar analyses of associated biological signals across multiple axes.

  18. Statistical Analysis of Adaptive Beam-Forming Methods

    DTIC Science & Technology

    1988-05-01

    minimum amount of computing resources? * What are the tradeoffs being made when a system design selects block averaging over exponential averaging? Will...understood by many signal processing practitioners, however, is how system parameters and the number of sensors effect the distribution of the... system performance improve and if so by how much? b " It is well known that the noise sampled at adjacent sensors is not statistically independent

  19. Statistical universals reveal the structures and functions of human music.

    PubMed

    Savage, Patrick E; Brown, Steven; Sakai, Emi; Currie, Thomas E

    2015-07-21

    Music has been called "the universal language of mankind." Although contemporary theories of music evolution often invoke various musical universals, the existence of such universals has been disputed for decades and has never been empirically demonstrated. Here we combine a music-classification scheme with statistical analyses, including phylogenetic comparative methods, to examine a well-sampled global set of 304 music recordings. Our analyses reveal no absolute universals but strong support for many statistical universals that are consistent across all nine geographic regions sampled. These universals include 18 musical features that are common individually as well as a network of 10 features that are commonly associated with one another. They span not only features related to pitch and rhythm that are often cited as putative universals but also rarely cited domains including performance style and social context. These cross-cultural structural regularities of human music may relate to roles in facilitating group coordination and cohesion, as exemplified by the universal tendency to sing, play percussion instruments, and dance to simple, repetitive music in groups. Our findings highlight the need for scientists studying music evolution to expand the range of musical cultures and musical features under consideration. The statistical universals we identified represent important candidates for future investigation.

  20. Statistical universals reveal the structures and functions of human music

    PubMed Central

    Savage, Patrick E.; Brown, Steven; Sakai, Emi; Currie, Thomas E.

    2015-01-01

    Music has been called “the universal language of mankind.” Although contemporary theories of music evolution often invoke various musical universals, the existence of such universals has been disputed for decades and has never been empirically demonstrated. Here we combine a music-classification scheme with statistical analyses, including phylogenetic comparative methods, to examine a well-sampled global set of 304 music recordings. Our analyses reveal no absolute universals but strong support for many statistical universals that are consistent across all nine geographic regions sampled. These universals include 18 musical features that are common individually as well as a network of 10 features that are commonly associated with one another. They span not only features related to pitch and rhythm that are often cited as putative universals but also rarely cited domains including performance style and social context. These cross-cultural structural regularities of human music may relate to roles in facilitating group coordination and cohesion, as exemplified by the universal tendency to sing, play percussion instruments, and dance to simple, repetitive music in groups. Our findings highlight the need for scientists studying music evolution to expand the range of musical cultures and musical features under consideration. The statistical universals we identified represent important candidates for future investigation. PMID:26124105

  1. Analysis of nutrients, selected inorganic constituents, and trace elements in water from Illinois community-supply wells, 1984-91

    USGS Publications Warehouse

    Warner, Kelly L.

    2000-01-01

    The lower Illinois River Basin (LIRB) study unit is part of the National Water-Quality Assessment program that includes studies of most major aquifer systems in the United States. Retrospective water-quality data from community-supply wells in the LIRB and in the rest of Illinois are grouped by aquifer and depth interval. Concentrations of selected chemical constituents in water samples from community-supply wells within the LIRB vary with aquifer and depth of well. Ranked data for 16 selected trace elements and nutrients are compared by aquifer, depth interval, and between the LIRB and the rest of Illinois using nonparametric statistical analyses. For all wells, median concentrations of nitrate and nitrite (as Nitrogen) are highest in water samples from the Quaternary aquifer at well depths less than 100 ft; ammonia concentrations (as Nitrogen), however, are highest in samples from well depths greater than 200 ft. Chloride and sulfate concentrations are higher in samples from the older bedrock aquifers. Arsenic, lead, sulfate, and zinc concentrations are appreciably different between samples from the LIRB and samples from the rest of Illinois for ground water from the Quaternary aquifer. Arsenic concentration is highest in the deep Quaternary aquifer. Chromium, cyanide, lead, and mercury are not frequently detected in water samples from community-supply wells in Illinois.

  2. CALIPSO Observations of Near-Cloud Aerosol Properties as a Function of Cloud Fraction

    NASA Technical Reports Server (NTRS)

    Yang, Weidong; Marshak, Alexander; Varnai, Tamas; Wood, Robert

    2015-01-01

    This paper uses spaceborne lidar data to study how near-cloud aerosol statistics of attenuated backscatter depend on cloud fraction. The results for a large region around the Azores show that: (1) far-from-cloud aerosol statistics are dominated by samples from scenes with lower cloud fractions, while near-cloud aerosol statistics are dominated by samples from scenes with higher cloud fractions; (2) near-cloud enhancements of attenuated backscatter occur for any cloud fraction but are most pronounced for higher cloud fractions; (3) the difference in the enhancements for different cloud fractions is most significant within 5km from clouds; (4) near-cloud enhancements can be well approximated by logarithmic functions of cloud fraction and distance to clouds. These findings demonstrate that if variability in cloud fraction across the scenes used to composite aerosol statistics are not considered, a sampling artifact will affect these statistics calculated as a function of distance to clouds. For the Azores-region dataset examined here, this artifact occurs mostly within 5 km from clouds, and exaggerates the near-cloud enhancements of lidar backscatter and color ratio by about 30. This shows that for accurate characterization of the changes in aerosol properties with distance to clouds, it is important to account for the impact of changes in cloud fraction.

  3. Fronthaul evolution: From CPRI to Ethernet

    NASA Astrophysics Data System (ADS)

    Gomes, Nathan J.; Chanclou, Philippe; Turnbull, Peter; Magee, Anthony; Jungnickel, Volker

    2015-12-01

    It is proposed that using Ethernet in the fronthaul, between base station baseband unit (BBU) pools and remote radio heads (RRHs), can bring a number of advantages, from use of lower-cost equipment, shared use of infrastructure with fixed access networks, to obtaining statistical multiplexing and optimised performance through probe-based monitoring and software-defined networking. However, a number of challenges exist: ultra-high-bit-rate requirements from the transport of increased bandwidth radio streams for multiple antennas in future mobile networks, and low latency and jitter to meet delay requirements and the demands of joint processing. A new fronthaul functional division is proposed which can alleviate the most demanding bit-rate requirements by transport of baseband signals instead of sampled radio waveforms, and enable statistical multiplexing gains. Delay and synchronisation issues remain to be solved.

  4. Analysis of the Einstein sample of early-type galaxies

    NASA Technical Reports Server (NTRS)

    Eskridge, Paul B.; Fabbiano, Giuseppina

    1993-01-01

    The EINSTEIN galaxy catalog contains x-ray data for 148 early-type (E and SO) galaxies. A detailed analysis of the global properties of this sample are studied. By comparing the x-ray properties with other tracers of the ISM, as well as with observables related to the stellar dynamics and populations of the sample, we expect to determine more clearly the physical relationships that determine the evolution of early-type galaxies. Previous studies with smaller samples have explored the relationships between x-ray luminosity (L(sub x)) and luminosities in other bands. Using our larger sample and the statistical techniques of survival analysis, a number of these earlier analyses were repeated. For our full sample, a strong statistical correlation is found between L(sub X) and L(sub B) (the probability that the null hypothesis is upheld is P less than 10(exp -4) from a variety of rank correlation tests. Regressions with several algorithms yield consistent results.

  5. California GAMA Program: Ground-Water Quality Data in the Northern San Joaquin Basin Study Unit, 2005

    USGS Publications Warehouse

    Bennett, George L.; Belitz, Kenneth; Milby Dawson, Barbara J.

    2006-01-01

    Growing concern over the closure of public-supply wells because of ground-water contamination has led the State Water Board to establish the Ground-Water Ambient Monitoring and Assessment (GAMA) Program. With the aid of the U.S. Geological Survey (USGS) and Lawrence Livermore National Laboratory, the program goals are to enhance understanding and provide a current assessment of ground-water quality in areas where ground water is an important source of drinking water. The Northern San Joaquin Basin GAMA study unit covers an area of approximately 2,079 square miles (mi2) across four hydrologic study areas in the San Joaquin Valley. The four study areas are the California Department of Water Resources (CADWR) defined Tracy subbasin, the CADWR-defined Eastern San Joaquin subbasin, the CADWR-defined Cosumnes subbasin, and the sedimentologically distinct USGS-defined Uplands study area, which includes portions of both the Cosumnes and Eastern San Joaquin subbasins. Seventy ground-water samples were collected from 64 public-supply, irrigation, domestic, and monitoring wells within the Northern San Joaquin Basin GAMA study unit. Thirty-two of these samples were collected in the Eastern San Joaquin Basin study area, 17 in the Tracy Basin study area, 10 in the Cosumnes Basin study area, and 11 in the Uplands Basin study area. Of the 32 samples collected in the Eastern San Joaquin Basin, 6 were collected using a depth-dependent sampling pump. This pump allows for the collection of samples from discrete depths within the pumping well. Two wells were chosen for depth-dependent sampling and three samples were collected at varying depths within each well. Over 350 water-quality field parameters, chemical constituents, and microbial constituents were analyzed and are reported as concentrations and as detection frequencies, by compound classification as well as for individual constituents, for the Northern San Joaquin Basin study unit as a whole and for each individual study area. Results are presented in a descending order based on detection frequencies (most frequently detected compound listed first), or alphabetically when a detection frequency could not be calculated. Only certain wells were measured for all constituents and water-quality parameters. The results of all of the analyses were compared with U.S. Environmental Protection Agency (USEPA) and California Department of Health Services (CADHS) Maximum Contaminant Levels (MCLs), Secondary Maximum Contaminant Levels (SMCLs), USEPA lifetime health advisories (HA-Ls), the risk-specific dose at a cancer risk level equal to 1 in 100,000 or 10E-5 (RSD5), and CADHS notification levels (NLs). When USEPA and CADHS MCLs are the same, detection levels were compared with the USEPA standard; however, in some cases, the CADHS MCL may be lower. In those cases, the data were compared with the CADHS MCL. Constituents listed by CADHS as 'unregulated chemicals for which monitoring is required' were compared with the CADHS 'detection level for the purposes of reporting' (DLR). DLRs unlike MCLs are not health based standards. Instead, they are levels at which current laboratory detection capabilities allow eighty percent of qualified laboratories to achieve measurements within thirty percent of the true concentration. Twenty-three volatile organic compounds (VOCs) and seven gasoline oxygenates were detected in ground-water samples collected in the Northern San Joaquin Basin GAMA study unit. Additionally, 13 tentatively identified compounds were detected. VOCs were most frequently detected in the Eastern San Joaquin Basin study area and least frequently detected in samples collected in the Cosumnes Basin study area. Dichlorodifluoromethane (CFC-12), a CADHS 'unregulated chemical for which monitoring is required,' was detected in two wells at concentrations greater than the DLR. Trihalomethanes were the most frequently detected class of VOC constituents. Chloroform (trichloromethane) was the m

  6. Groundwater quality in the Genesee River Basin, New York, 2010

    USGS Publications Warehouse

    Reddy, James E.

    2012-01-01

    Water samples collected from eight production wells and eight private residential wells in the Genesee River Basin from September through December 2010 were analyzed to characterize the groundwater quality in the basin. Eight of the wells were completed in sand and gravel aquifers, and eight were finished in bedrock aquifers. Three of the 16 wells were sampled in the first Genesee River Basin study during 2005-2006. Water samples from the 2010 study were analyzed for 147 physiochemical properties and constituents that included major ions, nutrients, trace elements, radionuclides, pesticides, volatile organic compounds (VOCs), and indicator bacteria. Results of the water-quality analyses are presented in tabular form for individual wells, and summary statistics for specific constituents are presented by aquifer type. The results are compared with Federal and New York State drinking-water standards, which typically are identical. The results indicate that groundwater generally is of acceptable quality, although concentrations of the following constituents exceeded current or proposed Federal or New York State drinking-water standards at each of the 16 wells sampled: color (one sample), sodium (three samples), sulfate (three samples), total dissolved solids (four samples), aluminum (one sample), arsenic (two samples), copper (one sample), iron (nine samples), manganese (eight samples), radon-222 (nine samples), and total coliform bacteria (six samples). Existing drinking-water standards for pH, chloride, fluoride, nitrate, nitrite, antimony, barium, beryllium, cadmium, chromium, lead, mercury, selenium, silver, thallium, zinc, gross alpha radioactivity, uranium, fecal coliform, Escherichia coli, and heterotrophic bacteria were not exceeded in any of the samples collected. None of the pesticides and VOCs analyzed exceeded existing drinking-water standards.

  7. Groundwater quality in western New York, 2011

    USGS Publications Warehouse

    Reddy, James E.

    2013-01-01

    Water samples collected from 16 production wells and 15 private residential wells in western New York from July through November 2011 were analyzed to characterize the groundwater quality. Fifteen of the wells were finished in sand and gravel aquifers, and 16 were finished in bedrock aquifers. Six of the 31 wells were sampled in a previous western New York study, which was conducted in 2006. Water samples from the 2011 study were analyzed for 147 physiochemical properties and constituents that included major ions, nutrients, trace elements, radionuclides, pesticides, volatile organic compounds (VOCs), and indicator bacteria. Results of the water-quality analyses are presented in tabular form for individual wells, and summary statistics for specific constituents are presented by aquifer type. The results are compared with Federal and New York State drinking-water standards, which typically are identical. The results indicate that groundwater generally is of acceptable quality, although at 30 of the 31 wells sampled, at least one of the following constituents was detected at a concentration that exceeded current or proposed Federal or New York State drinking-water standards: pH (two samples), sodium (eight samples), sulfate (three samples), total dissolved solids (nine samples), aluminum (two samples), arsenic (one sample), iron (ten samples), manganese (twelve samples), radon-222 (sixteen samples), benzene (one sample), and total coliform bacteria (nine samples). Existing drinking-water standards for color, chloride, fluoride, nitrate, nitrite, antimony, barium, beryllium, cadmium, chromium, copper, lead, mercury, selenium, silver, thallium, zinc, gross alpha radioactivity, uranium, fecal coliform, Escherichia coli, and heterotrophic bacteria were not exceeded in any of the samples collected. None of the pesticides analyzed exceeded existing drinking-water standards.

  8. Knowledge level of effect size statistics, confidence intervals and meta-analysis in Spanish academic psychologists.

    PubMed

    Badenes-Ribera, Laura; Frias-Navarro, Dolores; Pascual-Soler, Marcos; Monterde-I-Bort, Héctor

    2016-11-01

    The statistical reform movement and the American Psychological Association (APA) defend the use of estimators of the effect size and its confidence intervals, as well as the interpretation of the clinical significance of the findings. A survey was conducted in which academic psychologists were asked about their behavior in designing and carrying out their studies. The sample was composed of 472 participants (45.8% men). The mean number of years as a university professor was 13.56 years (SD= 9.27). The use of effect-size estimators is becoming generalized, as well as the consideration of meta-analytic studies. However, several inadequate practices still persist. A traditional model of methodological behavior based on statistical significance tests is maintained, based on the predominance of Cohen’s d and the unadjusted R2/η2, which are not immune to outliers or departure from normality and the violations of statistical assumptions, and the under-reporting of confidence intervals of effect-size statistics. The paper concludes with recommendations for improving statistical practice.

  9. Remineralization Property of an Orthodontic Primer Containing a Bioactive Glass with Silver and Zinc

    PubMed Central

    Lee, Seung-Min; Kim, In-Ryoung; Park, Bong-Soo; Ko, Ching-Chang; Son, Woo-Sung; Kim, Yong-Il

    2017-01-01

    White spot lesions (WSLs) are irreversible damages in orthodontic treatment due to excessive etching or demineralization by microorganisms. In this study, we conducted a mechanical and cell viability test to examine the antibacterial properties of 0.2% and 1% bioactive glass (BAG) and silver-doped and zinc-doped BAGs in a primer and evaluated their clinical applicability to prevent WSLs. The microhardness statistically significantly increased in the adhesive-containing BAG, while the other samples showed no statistically significant difference compared with the control group. The shear bond strength of all samples increased compared with that of the control group. The cell viability of the control and sample groups was similar within 24 h, but decreased slightly over 48 h. All samples showed antibacterial properties. Regarding remineralization property, the group containing 0.2% of the samples showed remineralization properties compared with the control group, but was not statistically significant; further, the group containing 1% of the samples showed a significant difference compared with the control group. Among them, the orthodontic bonding primer containing 1% silver-doped BAG showed the highest remineralization property. The new orthodontic bonding primer used in this study showed an antimicrobial effect, chemical remineralization effect, and WSL prevention as well as clinically applicable properties, both physically and biologically. PMID:29088092

  10. General Satisfaction Among Healthcare Workers: Differences Between Employees in Medical and Mental Health Sector

    PubMed Central

    Papathanasiou, Ioanna V.; Kleisiaris, Christos F.; Tsaras, Konstantinos; Fradelos, Evangelos C.; Kourkouta, Lambrini

    2015-01-01

    Background: General satisfaction is a personal experience and sources of satisfaction or dissatisfaction vary between professional groups. General satisfaction is usually related with work settings, work performance and mental health status. Aim: The purpose of this research study was to investigate the level of general satisfaction of health care workers and to examine whether there were any differences among employees of medical and mental health sector. Methods: The sample consisted of employees from the medical and mental health sector, who were all randomly selected. A two-part questionnaire was used to collect data. The first section involved demographic information and the second part was a General Satisfaction Questionnaire (GSQ). The statistical analysis of data was performed using the software package 19.0 for Windows. Descriptive statistics were initially generated for sample characteristics. All data exhibited normal distributions and thus the parametric t-test was used to compare mean scores between the two health sectors. P values < 0.05 were defined as reflecting the acceptable level of statistical significance. Results: 457 healthcare workers completed the questionnaire. The mean age of the sample was 41.8 ± 7.9 years. The Cronbach alpha coefficient for GSQ was 0.79. The total mean score of general satisfaction for the employees in medical sector was 4.5 (5=very satisfied) and for the employees in mental health sector is 4.8. T-test showed that these results are statistical different (t=4.55, p<0.01) and therefore the two groups of healthcare workers feel different general satisfaction. Conclusions: Mental health employees appear to experience higher levels of general satisfaction and mainly they experience higher satisfaction from family roles, life and sexual life, emotional state and relations with patients. PMID:26543410

  11. Indian Ocean radiocarbon: Data from the INDIGO 1, 2, and 3 cruises

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sepanski, R.J.

    1991-01-01

    This document presents {sup 14}C activities (expressed in the internationally adopted {Delta}{sup 14}C scale) from water samples taken at various locations and depths in the Indian and Southern oceans through the Indien Gaz Ocean (INDIGO) project. These data were collected as part of the INDIGO 1, INDIGO 2, and INDIGO 3 cruises, which took place during the years 1985, 1986, and 1987, respectively. These data have been used to estimate the penetration of anthropogenic CO{sub 2} in the Indian and Southern oceans. The document also presents supporting data for potential temperature, salinity, density (sigma-theta), {delta}{sup 13}C, and total CO{sub 2}.more » All radiocarbon measurements have been examined statistically for quality of sample counts and stability of counting efficiency and background. In addition, all data have been reviewed by the Carbon Dioxide Information Analysis Center and assessed for gross accuracy and consistency (absence of obvious outliers and other anomalous values). These data are available free of charge as a numeric data package (NDP) from the Carbon Dioxide Information Analysis Center. The NDP consists of this document and a magnetic tape containing machine-readable files. This document provides sample listing of the Indian Ocean radiocarbon data as they appear on the magnetic tape, as well as a complete listing of these data in tabular form. This document also offers retrieval program listings, furnishes information on sampling methods and data selection, defines limitations and restrictions of the data, and provides reprints of pertinent literature. 13 refs., 4 tabs.« less

  12. Ground-Water Quality Data in the Owens and Indian Wells Valleys Study Unit, 2006: Results from the California GAMA Program

    USGS Publications Warehouse

    Densmore, Jill N.; Fram, Miranda S.; Belitz, Kenneth

    2009-01-01

    Ground-water quality in the approximately 1,630 square-mile Owens and Indian Wells Valleys study unit (OWENS) was investigated in September-December 2006 as part of the Priority Basin Project of Groundwater Ambient Monitoring and Assessment (GAMA) Program. The GAMA Priority Basin Project was developed in response to the Groundwater Quality Monitoring Act of 2001 and is being conducted by the U.S. Geological Survey (USGS) in collaboration with the California State Water Resources Control Board (SWRCB). The Owens and Indian Wells Valleys study was designed to provide a spatially unbiased assessment of raw ground-water quality within OWENS study unit, as well as a statistically consistent basis for comparing water quality throughout California. Samples were collected from 74 wells in Inyo, Kern, Mono, and San Bernardino Counties. Fifty-three of the wells were selected using a spatially distributed, randomized grid-based method to provide statistical representation of the study area (grid wells), and 21 wells were selected to evaluate changes in water chemistry in areas of interest (understanding wells). The ground-water samples were analyzed for a large number of synthetic organic constituents [volatile organic compounds (VOCs), pesticides and pesticide degradates, pharmaceutical compounds, and potential wastewater- indicator compounds], constituents of special interest [perchlorate, N-nitrosodimethylamine (NDMA), and 1,2,3- trichloropropane (1,2,3-TCP)], naturally occurring inorganic constituents [nutrients, major and minor ions, and trace elements], radioactive constituents, and microbial indicators. Naturally occurring isotopes [tritium, and carbon-14, and stable isotopes of hydrogen and oxygen in water], and dissolved noble gases also were measured to help identify the source and age of the sampled ground water. This study evaluated the quality of raw ground water in the aquifer in the OWENS study unit and did not attempt to evaluate the quality of treated water delivered to consumers. Water supplied to consumers typically is treated after withdrawal from the ground, disinfected, and blended with other waters to maintain acceptable water quality. Regulatory thresholds apply to treated water that is served to the consumer, not to raw ground water. However, to provide some context for the results, concentrations of constituents measured in the raw ground water were compared with regulatory and non-regulatory health-based thresholds established by the U.S. Environmental Protection Agency (USEPA) and California Department of Public Health (CDPH) and non-regulatory thresholds established for aesthetic concerns (secondary maximum contamination levels, SMCL-CA) by CDPH. VOCs and pesticides were detected in samples from less than one-third of the grid wells; all detections were below health-based thresholds, and most were less than one-one hundredth of threshold values. All detections of perchlorate and nutrients in samples from OWENS were below health-based thresholds. Most detections of trace elements in ground-water samples from OWENS wells were below health-based thresholds. In samples from the 53 grid wells, three constituents were detected at concentrations above USEPA maximum contaminant levels: arsenic in 5 samples, uranium in 4 samples, and fluoride in 1 sample. Two constituents were detected at concentrations above CDPH notification levels (boron in 9 samples and vanadium in 1 sample), and two were above USEPA lifetime health advisory levels (molybdenum in 3 samples and strontium in 1 sample). Most of the samples from OWENS wells had concentrations of major elements, TDS, and trace elements below the non-enforceable standards set for aesthetic concerns. Samples from nine grid wells had concentrations of manganese, iron, or TDS above the SMCL-CAs.

  13. Slowdowns in diversification rates from real phylogenies may not be real.

    PubMed

    Cusimano, Natalie; Renner, Susanne S

    2010-07-01

    Studies of diversification patterns often find a slowing in lineage accumulation toward the present. This seemingly pervasive pattern of rate downturns has been taken as evidence for adaptive radiations, density-dependent regulation, and metacommunity species interactions. The significance of rate downturns is evaluated with statistical tests (the gamma statistic and Monte Carlo constant rates (MCCR) test; birth-death likelihood models and Akaike Information Criterion [AIC] scores) that rely on null distributions, which assume that the included species are a random sample of the entire clade. Sampling in real phylogenies, however, often is nonrandom because systematists try to include early-diverging species or representatives of previous intrataxon classifications. We studied the effects of biased sampling, structured sampling, and random sampling by experimentally pruning simulated trees (60 and 150 species) as well as a completely sampled empirical tree (58 species) and then applying the gamma statistic/MCCR test and birth-death likelihood models/AIC scores to assess rate changes. For trees with random species sampling, the true model (i.e., the one fitting the complete phylogenies) could be inferred in most cases. Oversampling deep nodes, however, strongly biases inferences toward downturns, with simulations of structured and biased sampling suggesting that this occurs when sampling percentages drop below 80%. The magnitude of the effect and the sensitivity of diversification rate models is such that a useful rule of thumb may be not to infer rate downturns from real trees unless they have >80% species sampling.

  14. Adjusted Wald Confidence Interval for a Difference of Binomial Proportions Based on Paired Data

    ERIC Educational Resources Information Center

    Bonett, Douglas G.; Price, Robert M.

    2012-01-01

    Adjusted Wald intervals for binomial proportions in one-sample and two-sample designs have been shown to perform about as well as the best available methods. The adjusted Wald intervals are easy to compute and have been incorporated into introductory statistics courses. An adjusted Wald interval for paired binomial proportions is proposed here and…

  15. Groundwater-quality data in the Santa Barbara study unit, 2011: results from the California GAMA Program

    USGS Publications Warehouse

    Davis, Tracy A.; Kulongoski, Justin T.; Belitz, Kenneth

    2013-01-01

    Groundwater quality in the 48-square-mile Santa Barbara study unit was investigated by the U.S. Geological Survey (USGS) from January to February 2011, as part of the California State Water Resources Control Board (SWRCB) Groundwater Ambient Monitoring and Assessment (GAMA) Program’s Priority Basin Project (PBP). The GAMA-PBP was developed in response to the California Groundwater Quality Monitoring Act of 2001 and is being conducted in collaboration with the SWRCB and Lawrence Livermore National Laboratory (LLNL). The Santa Barbara study unit was the thirty-fourth study unit to be sampled as part of the GAMA-PBP. The GAMA Santa Barbara study was designed to provide a spatially unbiased assessment of untreated-groundwater quality in the primary aquifer system, and to facilitate statistically consistent comparisons of untreated-groundwater quality throughout California. The primary aquifer system is defined as those parts of the aquifers corresponding to the perforation intervals of wells listed in the California Department of Public Health (CDPH) database for the Santa Barbara study unit. Groundwater quality in the primary aquifer system may differ from the quality in the shallower or deeper water-bearing zones; shallow groundwater may be more vulnerable to surficial contamination. In the Santa Barbara study unit located in Santa Barbara and Ventura Counties, groundwater samples were collected from 24 wells. Eighteen of the wells were selected by using a spatially distributed, randomized grid-based method to provide statistical representation of the study unit (grid wells), and six wells were selected to aid in evaluation of water-quality issues (understanding wells). The groundwater samples were analyzed for organic constituents (volatile organic compounds [VOCs], pesticides and pesticide degradates, and pharmaceutical compounds); constituents of special interest (perchlorate and N-nitrosodimethylamine [NDMA]); naturally occurring inorganic constituents (trace elements, nutrients, major and minor ions, silica, total dissolved solids [TDS], alkalinity, and arsenic, chromium, and iron species); and radioactive constituents (radon-222 and gross alpha and gross beta radioactivity). Naturally occurring isotopes (stable isotopes of hydrogen and oxygen in water, stables isotopes of inorganic carbon and boron dissolved in water, isotope ratios of dissolved strontium, tritium activities, and carbon-14 abundances) and dissolved noble gases also were measured to help identify the sources and ages of the sampled groundwater. In total, 281 constituents and water-quality indicators were measured. Three types of quality-control samples (blanks, replicates, and matrix spikes) were collected at up to 12 percent of the wells in the Santa Barbara study unit, and the results for these samples were used to evaluate the quality of the data for the groundwater samples. Blanks rarely contained detectable concentrations of any constituent, suggesting that contamination from sample collection procedures was not a significant source of bias in the data for the groundwater samples. Replicate samples generally were within the limits of acceptable analytical reproducibility. Matrix-spike recoveries were within the acceptable range (70 to 130 percent) for approximately 82 percent of the compounds. This study did not attempt to evaluate the quality of water delivered to consumers; after withdrawal from the ground, untreated groundwater typically is treated, disinfected, and (or) blended with other waters to maintain water quality. Regulatory benchmarks apply to water that is served to the consumer, not to untreated groundwater. However, to provide some context for the results, concentrations of constituents measured in the untreated groundwater were compared with regulatory and non-regulatory health-based benchmarks established by the U.S. Environmental Protection Agency (USEPA) and CDPH and to non-regulatory benchmarks established for aesthetic concerns by CDPH. Comparisons between data collected for this study and benchmarks for drinking water are for illustrative purposes only and are not indicative of compliance or non-compliance with those benchmarks. All organic constituents and most inorganic constituents that were detected in groundwater samples from the 18 grid wells in the Santa Barbara study unit were detected at concentrations less than drinking-water benchmarks. Of the 220 organic and special-interest constituents sampled for at the 18 grid wells, 13 were detected in groundwater samples; concentrations of all detected constituents were less than regulatory and non-regulatory health-based benchmarks. In total, VOCs were detected in 61 percent of the 18 grid wells sampled, pesticides and pesticide degradates were detected in 11 percent, and perchlorate was detected in 67 percent. Polar pesticides and their degradates, pharmaceutical compounds, and NDMA were not detected in any of the grid wells sampled in the Santa Barbara study unit. Eighteen grid wells were sampled for trace elements, major and minor ions, nutrients, and radioactive constituents; most detected concentrations were less than health-based benchmarks. Exceptions are one detection of boron greater than the CDPH notification level (NL-CA) of 1,000 micrograms per liter (μg/L) and one detection of fluoride greater than the CDPH maximum contaminant level (MCL-CA) of 2 milligrams per liter (mg/L). Results for constituents with non-regulatory benchmarks set for aesthetic concerns from the grid wells showed that iron concentrations greater than the CDPH secondary maximum contaminant level (SMCL-CA) of 300 μg/L were detected in three grid wells. Manganese concentrations greater than the SMCL-CA of 50 μg/L were detected in seven grid wells. Chloride was detected at a concentration greater than the SMCL-CA recommended benchmark of 250 mg/L in four grid wells. Sulfate concentrations greater than the SMCL-CA recommended benchmark of 250 mg/L were measured in eight grid wells, and the concentration in one of these wells was also greater than the SMCL-CA upper benchmark of 500 mg/L. TDS concentrations greater than the SMCL-CA recommended benchmark of 500 mg/L were measured in 17 grid wells, and concentrations in six of these wells were also greater than the SMCL-CA upper benchmark of 1,000 mg/L.

  16. Improving qPCR telomere length assays: Controlling for well position effects increases statistical power.

    PubMed

    Eisenberg, Dan T A; Kuzawa, Christopher W; Hayes, M Geoffrey

    2015-01-01

    Telomere length (TL) is commonly measured using quantitative PCR (qPCR). Although, easier than the southern blot of terminal restriction fragments (TRF) TL measurement method, one drawback of qPCR is that it introduces greater measurement error and thus reduces the statistical power of analyses. To address a potential source of measurement error, we consider the effect of well position on qPCR TL measurements. qPCR TL data from 3,638 people run on a Bio-Rad iCycler iQ are reanalyzed here. To evaluate measurement validity, correspondence with TRF, age, and between mother and offspring are examined. First, we present evidence for systematic variation in qPCR TL measurements in relation to thermocycler well position. Controlling for these well-position effects consistently improves measurement validity and yields estimated improvements in statistical power equivalent to increasing sample sizes by 16%. We additionally evaluated the linearity of the relationships between telomere and single copy gene control amplicons and between qPCR and TRF measures. We find that, unlike some previous reports, our data exhibit linear relationships. We introduce the standard error in percent, a superior method for quantifying measurement error as compared to the commonly used coefficient of variation. Using this measure, we find that excluding samples with high measurement error does not improve measurement validity in our study. Future studies using block-based thermocyclers should consider well position effects. Since additional information can be gleaned from well position corrections, rerunning analyses of previous results with well position correction could serve as an independent test of the validity of these results. © 2015 Wiley Periodicals, Inc.

  17. Thematic accuracy of the 1992 National Land-Cover Data for the eastern United States: Statistical methodology and regional results

    USGS Publications Warehouse

    Stehman, S.V.; Wickham, J.D.; Smith, J.H.; Yang, L.

    2003-01-01

    The accuracy of the 1992 National Land-Cover Data (NLCD) map is assessed via a probability sampling design incorporating three levels of stratification and two stages of selection. Agreement between the map and reference land-cover labels is defined as a match between the primary or alternate reference label determined for a sample pixel and a mode class of the mapped 3×3 block of pixels centered on the sample pixel. Results are reported for each of the four regions comprising the eastern United States for both Anderson Level I and II classifications. Overall accuracies for Levels I and II are 80% and 46% for New England, 82% and 62% for New York/New Jersey (NY/NJ), 70% and 43% for the Mid-Atlantic, and 83% and 66% for the Southeast.

  18. Out-of-time-order fluctuation-dissipation theorem

    NASA Astrophysics Data System (ADS)

    Tsuji, Naoto; Shitara, Tomohiro; Ueda, Masahito

    2018-01-01

    We prove a generalized fluctuation-dissipation theorem for a certain class of out-of-time-ordered correlators (OTOCs) with a modified statistical average, which we call bipartite OTOCs, for general quantum systems in thermal equilibrium. The difference between the bipartite and physical OTOCs defined by the usual statistical average is quantified by a measure of quantum fluctuations known as the Wigner-Yanase skew information. Within this difference, the theorem describes a universal relation between chaotic behavior in quantum systems and a nonlinear-response function that involves a time-reversed process. We show that the theorem can be generalized to higher-order n -partite OTOCs as well as in the form of generalized covariance.

  19. Time-Dependent Selection of an Optimal Set of Sources to Define a Stable Celestial Reference Frame

    NASA Technical Reports Server (NTRS)

    Le Bail, Karine; Gordon, David

    2010-01-01

    Temporal statistical position stability is required for VLBI sources to define a stable Celestial Reference Frame (CRF) and has been studied in many recent papers. This study analyzes the sources from the latest realization of the International Celestial Reference Frame (ICRF2) with the Allan variance, in addition to taking into account the apparent linear motions of the sources. Focusing on the 295 defining sources shows how they are a good compromise of different criteria, such as statistical stability and sky distribution, as well as having a sufficient number of sources, despite the fact that the most stable sources of the entire ICRF2 are mostly in the Northern Hemisphere. Nevertheless, the selection of a stable set is not unique: studying different solutions (GSF005a and AUG24 from GSFC and OPA from the Paris Observatory) over different time periods (1989.5 to 2009.5 and 1999.5 to 2009.5) leads to selections that can differ in up to 20% of the sources. Observing, recording, and network improvement are some of the causes, showing better stability for the CRF over the last decade than the last twenty years. But this may also be explained by the assumption of stationarity that is not necessarily right for some sources.

  20. A proposed method to minimize waste from institutional radiation safety surveillance programs through the application of expected value statistics.

    PubMed

    Emery, R J

    1997-03-01

    Institutional radiation safety programs routinely use wipe test sampling and liquid scintillation counting analysis to indicate the presence of removable radioactive contamination. Significant volumes of liquid waste can be generated by such surveillance activities, and the subsequent disposal of these materials can sometimes be difficult and costly. In settings where large numbers of negative results are regularly obtained, the limited grouping of samples for analysis based on expected value statistical techniques is possible. To demonstrate the plausibility of the approach, single wipe samples exposed to varying amounts of contamination were analyzed concurrently with nine non-contaminated samples. Although the sample grouping inevitably leads to increased quenching with liquid scintillation counting systems, the effect did not impact the ability to detect removable contamination in amounts well below recommended action levels. Opportunities to further improve this cost effective semi-quantitative screening procedure are described, including improvements in sample collection procedures, enhancing sample-counting media contact through mixing and extending elution periods, increasing sample counting times, and adjusting institutional action levels.

  1. Advanced statistical methods for improved data analysis of NASA astrophysics missions

    NASA Technical Reports Server (NTRS)

    Feigelson, Eric D.

    1992-01-01

    The investigators under this grant studied ways to improve the statistical analysis of astronomical data. They looked at existing techniques, the development of new techniques, and the production and distribution of specialized software to the astronomical community. Abstracts of nine papers that were produced are included, as well as brief descriptions of four software packages. The articles that are abstracted discuss analytical and Monte Carlo comparisons of six different linear least squares fits, a (second) paper on linear regression in astronomy, two reviews of public domain software for the astronomer, subsample and half-sample methods for estimating sampling distributions, a nonparametric estimation of survival functions under dependent competing risks, censoring in astronomical data due to nondetections, an astronomy survival analysis computer package called ASURV, and improving the statistical methodology of astronomical data analysis.

  2. Association between obesity and depressive disorder in adolescents at high risk for depression.

    PubMed

    Hammerton, G; Thapar, A; Thapar, A K

    2014-04-01

    To examine the relationship between Body Mass Index (BMI) and depressive disorder in adolescents at high risk for depression. Prospective longitudinal 3-wave study of offspring of parents with recurrent depression. Replication in population-based cohort study. Three hundred and thirty-seven families where offspring were aged 9-17 years at baseline and 10-19 years at the final data point. Replication sample of adolescents from population-based cohort study aged 11-13 years at first assessment and 14-17 years at follow-up. High risk sample used BMI, skin-fold thickness, Diagnostic and Statistical Manual of Mental Disorders, fourth edition (DSM-IV)-defined major depressive disorder and depression symptoms using the Child and Adolescent Psychiatric Assessment (CAPA). Replication sample used BMI, DSM-IV depressive disorder and depression symptoms using the Development and Well-Being Assessment (DAWBA). Two hundred and eighty-nine adolescents were included in the primary analyses. The mean BMI for each age group in this sample were significantly higher than population norms. There was no significant longitudinal association between categories of weight (or BMI) and new onset depressive disorder or depression symptoms. Similar results were found for skin-fold thickness. The association was also tested in a replication population-based sample and found to be non-significant in the subsample of offspring with mothers who had experienced recurrent depression in the past. BMI at age 12 years was, however, a significant predictor of depression symptoms but not of depressive disorder at age 15 years for the total unselected population. BMI does not significantly predict the development of depression in the offspring of parents with recurrent depression.

  3. Assessing signal-to-noise in quantitative proteomics: multivariate statistical analysis in DIGE experiments.

    PubMed

    Friedman, David B

    2012-01-01

    All quantitative proteomics experiments measure variation between samples. When performing large-scale experiments that involve multiple conditions or treatments, the experimental design should include the appropriate number of individual biological replicates from each condition to enable the distinction between a relevant biological signal from technical noise. Multivariate statistical analyses, such as principal component analysis (PCA), provide a global perspective on experimental variation, thereby enabling the assessment of whether the variation describes the expected biological signal or the unanticipated technical/biological noise inherent in the system. Examples will be shown from high-resolution multivariable DIGE experiments where PCA was instrumental in demonstrating biologically significant variation as well as sample outliers, fouled samples, and overriding technical variation that would not be readily observed using standard univariate tests.

  4. Can natural variability trigger effects on fish and fish habitat as defined in environment Canada's metal mining environmental effects monitoring program?

    PubMed

    Mackey, Robin; Rees, Cassandra; Wells, Kelly; Pham, Samantha; England, Kent

    2013-01-01

    The Metal Mining Effluent Regulations (MMER) took effect in 2002 and require most metal mining operations in Canada to complete environmental effects monitoring (EEM) programs. An "effect" under the MMER EEM program is considered any positive or negative statistically significant difference in fish population, fish usability, or benthic invertebrate community EEM-defined endpoints. Two consecutive studies with the same statistically significant differences trigger more intensive monitoring, including the characterization of extent and magnitude and investigation of cause. Standard EEM study designs do not require multiple reference areas or preexposure sampling, thus results and conclusions about mine effects are highly contingent on the selection of a near perfect reference area and are at risk of falsely labeling natural variation as mine related "effects." A case study was completed to characterize the natural variability in EEM-defined endpoints during preexposure or baseline conditions. This involved completing a typical EEM study in future reference and exposure lakes surrounding a proposed uranium (U) mine in northern Saskatchewan, Canada. Moon Lake was sampled as the future exposure area as it is currently proposed to receive effluent from the U mine. Two reference areas were used: Slush Lake for both the fish population and benthic invertebrate community surveys and Lake C as a second reference area for the benthic invertebrate community survey. Moon Lake, Slush Lake, and Lake C are located in the same drainage basin in close proximity to one another. All 3 lakes contained similar water quality, fish communities, aquatic habitat, and a sediment composition largely comprised of fine-textured particles. The fish population survey consisted of a nonlethal northern pike (Esox lucius) and a lethal yellow perch (Perca flavescens) survey. A comparison of the 5 benthic invertebrate community effect endpoints, 4 nonlethal northern pike population effect endpoints, and 10 lethal yellow perch effect endpoints resulted in the observation of several statistically significant differences at the future exposure area relative to the reference area and/or areas. When the data from 2 reference areas assessed for the benthic invertebrate community survey were pooled, no significant differences in effect endpoints were observed. These results demonstrate weaknesses in the definition of an "effect" used by the MMER EEM program and in the use of a single reference area. Determination of the ecological significance of statistical differences identified as part of EEM programs conducted during the operational period should consider preexisting (background) natural variability between reference and exposure areas. Copyright © 2012 SETAC.

  5. Soft x-ray speckle from rough surfaces

    NASA Astrophysics Data System (ADS)

    Porter, Matthew Stanton

    Dynamic light scattering has been of great use in determining diffusion times for polymer solutions. At the same time, polymer thin films are becoming of increasing importance, especially in the semiconductor industry where they are used as photoresists and interlevel dielectrics. As the dimensions of these devices decrease we will reach a point where lasers will no longer be able to probe the length scales of interest. Current laser wavelengths limit the size of observable diffusion lengths to 180-700 nm. This dissertation will discuss attempts at pushing dynamic fight scattering experiments into the soft x-ray region so that we can examine fluctuations in polymer thin films on the molecular length scale. The dissertation explores the possibility of carrying out a dynamic light scattering experiment in the soft x-ray regime. A detailed account of how to meet the basic requirements for a coherent scattering experiment in the soft x-ray regime win be given. In addition, a complete description of the chamber design will be discussed. We used our custom designed scattering chamber to collect reproducible coherent soft x-ray scattering data from etched silicon wafers and from polystyrene coated silicon wafers. The data from the silicon wafers followed the statistics for a well-developed speckle pattern while the data from the polystyrene films exhibited Poisson statistics. We used the data from both the etched wafers and the polystyrene coated wafers to place a lower limit of ~20 Å on the RMS surface roughness of samples which will produce well defined speckle patterns for the current detector setup. Future experiments which use the criteria set forth in this dissertation have the opportunity to be even more successful than this dissertation project.

  6. A survey of the properties of early-type galaxies

    NASA Technical Reports Server (NTRS)

    Bregman, Joel N.; Roberts, M. S.; Hogg, D. E.

    1990-01-01

    A compilation of the properties of elliptical and early disk galaxies was completed. In addition to material from the literature, such as Infrared Astronomy Satellite (IRAS) fluxes, the compilation includes recent measurements of HI and CO, as well as a review of the x ray properties by Forman and Jones. The data are used to evaluate the gas content of early systems and to search for correlations with x ray emission. The interstellar medium in early-type galaxies is generally dominated by hot interstellar gas (T approx. 10 to the 7th power K; c.f. the review by Fabbiano 1989 and references therein). In addition, a significant fraction of these galaxies show infrared emission (Knapp, et al., 1989), optical emission lines, and visible dust. Sensitive studies in HI and CO of a number of these galaxies have been completed recently, resulting in several detections, particularly of the later types. Researchers wish to understand the connection among these different forms of the interstellar medium, and to examine the theoretical picture of the fate of the hot gas. To do so, they compiled observations of several forms of interstellar matter for a well-defined sample of early-type galaxies. Here they present a statistical analysis of this data base and discuss the implications of the results.

  7. Cephalopod dynamic camouflage: bridging the continuum between background matching and disruptive coloration

    PubMed Central

    Hanlon, R.T.; Chiao, C.-C.; Mäthger, L.M.; Barbosa, A.; Buresch, K.C.; Chubb, C.

    2008-01-01

    Individual cuttlefish, octopus and squid have the versatile capability to use body patterns for background matching and disruptive coloration. We define—qualitatively and quantitatively—the chief characteristics of the three major body pattern types used for camouflage by cephalopods: uniform and mottle patterns for background matching, and disruptive patterns that primarily enhance disruptiveness but aid background matching as well. There is great variation within each of the three body pattern types, but by defining their chief characteristics we lay the groundwork to test camouflage concepts by correlating background statistics with those of the body pattern. We describe at least three ways in which background matching can be achieved in cephalopods. Disruptive patterns in cuttlefish possess all four of the basic components of ‘disruptiveness’, supporting Cott's hypotheses, and we provide field examples of disruptive coloration in which the body pattern contrast exceeds that of the immediate surrounds. Based upon laboratory testing as well as thousands of images of camouflaged cephalopods in the field (a sample is provided on a web archive), we note that size, contrast and edges of background objects are key visual cues that guide cephalopod camouflage patterning. Mottle and disruptive patterns are frequently mixed, suggesting that background matching and disruptive mechanisms are often used in the same pattern. PMID:19008200

  8. Adaptive graph-based multiple testing procedures

    PubMed Central

    Klinglmueller, Florian; Posch, Martin; Koenig, Franz

    2016-01-01

    Multiple testing procedures defined by directed, weighted graphs have recently been proposed as an intuitive visual tool for constructing multiple testing strategies that reflect the often complex contextual relations between hypotheses in clinical trials. Many well-known sequentially rejective tests, such as (parallel) gatekeeping tests or hierarchical testing procedures are special cases of the graph based tests. We generalize these graph-based multiple testing procedures to adaptive trial designs with an interim analysis. These designs permit mid-trial design modifications based on unblinded interim data as well as external information, while providing strong family wise error rate control. To maintain the familywise error rate, it is not required to prespecify the adaption rule in detail. Because the adaptive test does not require knowledge of the multivariate distribution of test statistics, it is applicable in a wide range of scenarios including trials with multiple treatment comparisons, endpoints or subgroups, or combinations thereof. Examples of adaptations are dropping of treatment arms, selection of subpopulations, and sample size reassessment. If, in the interim analysis, it is decided to continue the trial as planned, the adaptive test reduces to the originally planned multiple testing procedure. Only if adaptations are actually implemented, an adjusted test needs to be applied. The procedure is illustrated with a case study and its operating characteristics are investigated by simulations. PMID:25319733

  9. Selected quality assurance data for water samples collected by the US Geological Survey, Idaho National Engineering Laboratory, Idaho, 1980 to 1988

    USGS Publications Warehouse

    Wegner, S.J.

    1989-01-01

    Multiple water samples from 115 wells and 3 surface water sites were collected between 1980 and 1988 for the ongoing quality assurance program at the Idaho National Engineering Laboratory. The reported results from the six laboratories involved were analyzed for agreement using descriptive statistics. The constituents and properties included: tritium, plutonium-238, plutonium-239, -240 (undivided), strontium-90, americium-241, cesium-137, total dissolved chromium, selected dissolved trace metals, sodium, chloride, nitrate, selected purgeable organic compounds, and specific conductance. Agreement could not be calculated for purgeable organic compounds, trace metals, some nitrates and blank sample analyses because analytical uncertainties were not consistently reported. However, differences between results for most of these data were calculated. The blank samples were not analyzed for differences. The laboratory results analyzed using descriptive statistics showed a median agreement between all useable data pairs of 95%. (USGS)

  10. Stable Estimation of a Covariance Matrix Guided by Nuclear Norm Penalties

    PubMed Central

    Chi, Eric C.; Lange, Kenneth

    2014-01-01

    Estimation of a covariance matrix or its inverse plays a central role in many statistical methods. For these methods to work reliably, estimated matrices must not only be invertible but also well-conditioned. The current paper introduces a novel prior to ensure a well-conditioned maximum a posteriori (MAP) covariance estimate. The prior shrinks the sample covariance estimator towards a stable target and leads to a MAP estimator that is consistent and asymptotically efficient. Thus, the MAP estimator gracefully transitions towards the sample covariance matrix as the number of samples grows relative to the number of covariates. The utility of the MAP estimator is demonstrated in two standard applications – discriminant analysis and EM clustering – in this sampling regime. PMID:25143662

  11. QADATA user's manual; an interactive computer program for the retrieval and analysis of the results from the external blind sample quality- assurance project of the U.S. Geological Survey

    USGS Publications Warehouse

    Lucey, K.J.

    1990-01-01

    The U.S. Geological Survey conducts an external blind sample quality assurance project for its National Water Quality Laboratory in Denver, Colorado, based on the analysis of reference water samples. Reference samples containing selected inorganic and nutrient constituents are disguised as environmental samples at the Survey 's office in Ocala, Florida, and are sent periodically through other Survey offices to the laboratory. The results of this blind sample project indicate the quality of analytical data produced by the laboratory. This report provides instructions on the use of QADATA, an interactive, menu-driven program that allows users to retrieve the results of the blind sample quality- assurance project. The QADATA program, which is available on the U.S. Geological Survey 's national computer network, accesses a blind sample data base that contains more than 50,000 determinations from the last five water years for approximately 40 constituents at various concentrations. The data can be retrieved from the database for any user- defined time period and for any or all available constituents. After the user defines the retrieval, the program prepares statistical tables, control charts, and precision plots and generates a report which can be transferred to the user 's office through the computer network. A discussion of the interpretation of the program output is also included. This quality assurance information will permit users to document the quality of the analytical results received from the laboratory. The blind sample data is entered into the database within weeks after being produced by the laboratory and can be retrieved to meet the needs of specific projects or programs. (USGS)

  12. Survival of Acinetobacter baumannii on dry surfaces.

    PubMed Central

    Wendt, C; Dietze, B; Dietz, E; Rüden, H

    1997-01-01

    Acinetobacter spp. have frequently been reported to be the causative agents of hospital outbreaks. The circumstances of some outbreaks demonstrated the long survival of Acinetobacter in a dry, inanimate environment. In laboratory experiments, we compared the abilities of five Acinetobacter baumannii strains, three Acinetobacter sp. strains from the American Type Culture Collection (ATCC), one Escherichia coli ATCC strain, and one Enterococcus faecium ATCC strain to survive under dry conditions. Bacterial solutions of the 10 strains were inoculated onto four different material samples (ceramic, polyvinyl chloride, rubber, and stainless steel) and stored under defined conditions. We investigated the bacterial counts of the material samples immediately after inoculation, after drying, and after 4 h, 1 day, and 1, 2, 4, 8, and 16 weeks of storage. A statistical model was used to distribute the 40 resulting curves among four types of survival curves. The type of survival curve was significantly associated with the bacterial strain but not with the material. The ability of the A. baumannii strains to survive under dry conditions varied greatly and correlated well with the source of the strain. Strains isolated from dry sources survived better than those isolated from wet sources. An outbreak strain that had caused hospital-acquired respiratory tract infections survived better than the strains from wet sources, but not as well as strains from dry sources. Resistance to dry conditions may promote the transmissibility of a strain, but it is not sufficient to make a strain an epidemic one. However, in the case of an outbreak, sources of Acinetobacter must be expected in the dry environment. PMID:9163451

  13. Statistical generation of training sets for measuring NO3(-), NH4(+) and major ions in natural waters using an ion selective electrode array.

    PubMed

    Mueller, Amy V; Hemond, Harold F

    2016-05-18

    Knowledge of ionic concentrations in natural waters is essential to understand watershed processes. Inorganic nitrogen, in the form of nitrate and ammonium ions, is a key nutrient as well as a participant in redox, acid-base, and photochemical processes of natural waters, leading to spatiotemporal patterns of ion concentrations at scales as small as meters or hours. Current options for measurement in situ are costly, relying primarily on instruments adapted from laboratory methods (e.g., colorimetric, UV absorption); free-standing and inexpensive ISE sensors for NO3(-) and NH4(+) could be attractive alternatives if interferences from other constituents were overcome. Multi-sensor arrays, coupled with appropriate non-linear signal processing, offer promise in this capacity but have not yet successfully achieved signal separation for NO3(-) and NH4(+)in situ at naturally occurring levels in unprocessed water samples. A novel signal processor, underpinned by an appropriate sensor array, is proposed that overcomes previous limitations by explicitly integrating basic chemical constraints (e.g., charge balance). This work further presents a rationalized process for the development of such in situ instrumentation for NO3(-) and NH4(+), including a statistical-modeling strategy for instrument design, training/calibration, and validation. Statistical analysis reveals that historical concentrations of major ionic constituents in natural waters across New England strongly covary and are multi-modal. This informs the design of a statistically appropriate training set, suggesting that the strong covariance of constituents across environmental samples can be exploited through appropriate signal processing mechanisms to further improve estimates of minor constituents. Two artificial neural network architectures, one expanded to incorporate knowledge of basic chemical constraints, were tested to process outputs of a multi-sensor array, trained using datasets of varying degrees of statistical representativeness to natural water samples. The accuracy of ANN results improves monotonically with the statistical representativeness of the training set (error decreases by ∼5×), while the expanded neural network architecture contributes a further factor of 2-3.5 decrease in error when trained with the most representative sample set. Results using the most statistically accurate set of training samples (which retain environmentally relevant ion concentrations but avoid the potential interference of humic acids) demonstrated accurate, unbiased quantification of nitrate and ammonium at natural environmental levels (±20% down to <10 μM), as well as the major ions Na(+), K(+), Ca(2+), Mg(2+), Cl(-), and SO4(2-), in unprocessed samples. These results show promise for the development of new in situ instrumentation for the support of scientific field work.

  14. Malaria Risk Assessment for the Republic of Korea Based on Models of Mosquito Distribution

    DTIC Science & Technology

    2008-06-01

    Yam;lda All. kleilli Rueda All. belellme Rueda VPH 0.8 • 0.6• ~ ~ 0.’ 0.2 0 H P V VPH Figure I, Illustration of the concept of the mal-area as it...the percentage of the sampled area that these parameters cover. The value for VPH could be used as a simplified index of malaria risk to compare...combinations of the VPH variables. These statistics will consist of the percentage of cells that contain a certain value for the user defined area

  15. Mathematics in modern immunology

    DOE PAGES

    Castro, Mario; Lythe, Grant; Molina-París, Carmen; ...

    2016-02-19

    Mathematical and statistical methods enable multidisciplinary approaches that catalyse discovery. Together with experimental methods, they identify key hypotheses, define measurable observables and reconcile disparate results. Here, we collect a representative sample of studies in T-cell biology that illustrate the benefits of modelling–experimental collaborations and that have proven valuable or even groundbreaking. Furthermore, we conclude that it is possible to find excellent examples of synergy between mathematical modelling and experiment in immunology, which have brought significant insight that would not be available without these collaborations, but that much remains to be discovered.

  16. Mathematics in modern immunology

    PubMed Central

    Castro, Mario; Lythe, Grant; Molina-París, Carmen; Ribeiro, Ruy M.

    2016-01-01

    Mathematical and statistical methods enable multidisciplinary approaches that catalyse discovery. Together with experimental methods, they identify key hypotheses, define measurable observables and reconcile disparate results. We collect a representative sample of studies in T-cell biology that illustrate the benefits of modelling–experimental collaborations and that have proven valuable or even groundbreaking. We conclude that it is possible to find excellent examples of synergy between mathematical modelling and experiment in immunology, which have brought significant insight that would not be available without these collaborations, but that much remains to be discovered. PMID:27051512

  17. Multiclass Bayes error estimation by a feature space sampling technique

    NASA Technical Reports Server (NTRS)

    Mobasseri, B. G.; Mcgillem, C. D.

    1979-01-01

    A general Gaussian M-class N-feature classification problem is defined. An algorithm is developed that requires the class statistics as its only input and computes the minimum probability of error through use of a combined analytical and numerical integration over a sequence simplifying transformations of the feature space. The results are compared with those obtained by conventional techniques applied to a 2-class 4-feature discrimination problem with results previously reported and 4-class 4-feature multispectral scanner Landsat data classified by training and testing of the available data.

  18. Issues in the Classification of Disease Instances with Ontologies

    PubMed Central

    Burgun, Anita; Bodenreider, Olivier; Jacquelinet, Christian

    2006-01-01

    Ontologies define classes of entities and their interrelations. They are used to organize data according to a theory of the domain. Towards that end, ontologies provide class definitions (i.e., the necessary and sufficient conditions for defining class membership). In medical ontologies, it is often difficult to establish such definitions for diseases. We use three examples (anemia, leukemia and schizophrenia) to illustrate the limitations of ontologies as classification resources. We show that eligibility criteria are often more useful than the Aristotelian definitions traditionally used in ontologies. Examples of eligibility criteria for diseases include complex predicates such as ‘ x is an instance of the class C when at least n criteria among m are verified’ and ‘symptoms must last at least one month if not treated, but less than one month, if effectively treated’. References to normality and abnormality are often found in disease definitions, but the operational definition of these references (i.e., the statistical and contextual information necessary to define them) is rarely provided. We conclude that knowledge bases that include probabilistic and statistical knowledge as well as rule-based criteria are more useful than Aristotelian definitions for representing the predicates defined by necessary and sufficient conditions. Rich knowledge bases are needed to clarify the relations between individuals and classes in various studies and applications. However, as ontologies represent relations among classes, they can play a supporting role in disease classification services built primarily on knowledge bases. PMID:16160339

  19. [Flavouring estimation of quality of grape wines with use of methods of mathematical statistics].

    PubMed

    Yakuba, Yu F; Khalaphyan, A A; Temerdashev, Z A; Bessonov, V V; Malinkin, A D

    2016-01-01

    The questions of forming of wine's flavour integral estimation during the tasting are discussed, the advantages and disadvantages of the procedures are declared. As investigating materials we used the natural white and red wines of Russian manufactures, which were made with the traditional technologies from Vitis Vinifera, straight hybrids, blending and experimental wines (more than 300 different samples). The aim of the research was to set the correlation between the content of wine's nonvolatile matter and wine's tasting quality rating by mathematical statistics methods. The content of organic acids, amino acids and cations in wines were considered as the main factors influencing on the flavor. Basically, they define the beverage's quality. The determination of those components in wine's samples was done by the electrophoretic method «CAPEL». Together with the analytical checking of wine's samples quality the representative group of specialists simultaneously carried out wine's tasting estimation using 100 scores system. The possibility of statistical modelling of correlation of wine's tasting estimation based on analytical data of amino acids and cations determination reasonably describing the wine's flavour was examined. The statistical modelling of correlation between the wine's tasting estimation and the content of major cations (ammonium, potassium, sodium, magnesium, calcium), free amino acids (proline, threonine, arginine) and the taking into account the level of influence on flavour and analytical valuation within fixed limits of quality accordance were done with Statistica. Adequate statistical models which are able to predict tasting estimation that is to determine the wine's quality using the content of components forming the flavour properties have been constructed. It is emphasized that along with aromatic (volatile) substances the nonvolatile matter - mineral substances and organic substances - amino acids such as proline, threonine, arginine influence on wine's flavour properties. It has been shown the nonvolatile components contribute in organoleptic and flavour quality estimation of wines as aromatic volatile substances but they take part in forming the expert's evaluation.

  20. Antiphase domains and reverse thermoremanent magnetism in ilmenite-hematite minerals

    USGS Publications Warehouse

    Lawson, C.A.; Nord, G.L.; Dowty, Eric; Hargraves, R.B.

    1981-01-01

    Examination of synthetic ilmenite-hematite samples by transmission electron microscopy has for the first time revealed the presence of well-defined antiphase domains and antiphase domain boundaries in this mineral system. Samples quenched from 1300??C have a high density of domain boundaries, whereas samples quenched from 900??C have a much lower density. Only the high-temperature samples acquire reverse thermoremanent magnetism when cooled in an applied magnetic field. The presence of a high density of domain boundaries seems to be a necessary condition for the acquisition of reverse thermoremanent magnetism.

  1. Determination of secondary electron emission characteristics of lunar soil samples

    NASA Technical Reports Server (NTRS)

    Gold, T.; Baron, R. L.; Bilson, E.

    1979-01-01

    A procedure is described for the determination of the 'apparent crossover voltage', i.e. the value of the primary (bombarding) electron energy at which an insulating sample surface changes the average sign of its charge. This apparent crossover point is characteristic of the secondary emission properties of insulating powders such as the lunar soil samples. Lunar core samples from well-defined, distinct soil layers are found to differ significantly in their secondary emission properties. This observation supports the suggestion that soil layers were deposited by an electrostatic transport process.

  2. A Primer on Receiver Operating Characteristic Analysis and Diagnostic Efficiency Statistics for Pediatric Psychology: We Are Ready to ROC

    PubMed Central

    2014-01-01

    Objective To offer a practical demonstration of receiver operating characteristic (ROC) analyses, diagnostic efficiency statistics, and their application to clinical decision making using a popular parent checklist to assess for potential mood disorder. Method Secondary analyses of data from 589 families seeking outpatient mental health services, completing the Child Behavior Checklist and semi-structured diagnostic interviews. Results Internalizing Problems raw scores discriminated mood disorders significantly better than did age- and gender-normed T scores, or an Affective Problems score. Internalizing scores <8 had a diagnostic likelihood ratio <0.3, and scores >30 had a diagnostic likelihood ratio of 7.4. Conclusions This study illustrates a series of steps in defining a clinical problem, operationalizing it, selecting a valid study design, and using ROC analyses to generate statistics that support clinical decisions. The ROC framework offers important advantages for clinical interpretation. Appendices include sample scripts using SPSS and R to check assumptions and conduct ROC analyses. PMID:23965298

  3. A primer on receiver operating characteristic analysis and diagnostic efficiency statistics for pediatric psychology: we are ready to ROC.

    PubMed

    Youngstrom, Eric A

    2014-03-01

    To offer a practical demonstration of receiver operating characteristic (ROC) analyses, diagnostic efficiency statistics, and their application to clinical decision making using a popular parent checklist to assess for potential mood disorder. Secondary analyses of data from 589 families seeking outpatient mental health services, completing the Child Behavior Checklist and semi-structured diagnostic interviews. Internalizing Problems raw scores discriminated mood disorders significantly better than did age- and gender-normed T scores, or an Affective Problems score. Internalizing scores <8 had a diagnostic likelihood ratio <0.3, and scores >30 had a diagnostic likelihood ratio of 7.4. This study illustrates a series of steps in defining a clinical problem, operationalizing it, selecting a valid study design, and using ROC analyses to generate statistics that support clinical decisions. The ROC framework offers important advantages for clinical interpretation. Appendices include sample scripts using SPSS and R to check assumptions and conduct ROC analyses.

  4. Lévy meets poisson: a statistical artifact may lead to erroneous recategorization of Lévy walk as Brownian motion.

    PubMed

    Gautestad, Arild O

    2013-03-01

    The flow of GPS data on animal space is challenging old paradigms, such as the issue of the scale-free Lévy walk versus scale-specific Brownian motion. Since these movement classes often require different protocols with respect to ecological analyses, further theoretical development in this field is important. I describe central concepts such as scale-specific versus scale-free movement and the difference between mechanistic and statistical-mechanical levels of analysis. Next, I report how a specific sampling scheme may have produced much confusion: a Lévy walk may be wrongly categorized as Brownian motion if the duration of a move, or bout, is used as a proxy for step length and a move is subjectively defined. Hence, the categorization and recategorization of movement class compliance surrounding the Lévy walk controversy may have been based on a statistical artifact. This issue may be avoided by collecting relocations at a fixed rate at a temporal scale that minimizes over- and undersampling.

  5. [The principal components analysis--method to classify the statistical variables with applications in medicine].

    PubMed

    Dascălu, Cristina Gena; Antohe, Magda Ecaterina

    2009-01-01

    Based on the eigenvalues and the eigenvectors analysis, the principal component analysis has the purpose to identify the subspace of the main components from a set of parameters, which are enough to characterize the whole set of parameters. Interpreting the data for analysis as a cloud of points, we find through geometrical transformations the directions where the cloud's dispersion is maximal--the lines that pass through the cloud's center of weight and have a maximal density of points around them (by defining an appropriate criteria function and its minimization. This method can be successfully used in order to simplify the statistical analysis on questionnaires--because it helps us to select from a set of items only the most relevant ones, which cover the variations of the whole set of data. For instance, in the presented sample we started from a questionnaire with 28 items and, applying the principal component analysis we identified 7 principal components--or main items--fact that simplifies significantly the further data statistical analysis.

  6. Statistical summaries of fatigue data for design purposes

    NASA Technical Reports Server (NTRS)

    Wirsching, P. H.

    1983-01-01

    Two methods are discussed for constructing a design curve on the safe side of fatigue data. Both the tolerance interval and equivalent prediction interval (EPI) concepts provide such a curve while accounting for both the distribution of the estimators in small samples and the data scatter. The EPI is also useful as a mechanism for providing necessary statistics on S-N data for a full reliability analysis which includes uncertainty in all fatigue design factors. Examples of statistical analyses of the general strain life relationship are presented. The tolerance limit and EPI techniques for defining a design curve are demonstrated. Examples usng WASPALOY B and RQC-100 data demonstrate that a reliability model could be constructed by considering the fatigue strength and fatigue ductility coefficients as two independent random variables. A technique given for establishing the fatigue strength for high cycle lives relies on an extrapolation technique and also accounts for "runners." A reliability model or design value can be specified.

  7. Psychological and behavioral differences between low back pain populations: a comparative analysis of chiropractic, primary and secondary care patients.

    PubMed

    Eklund, Andreas; Bergström, Gunnar; Bodin, Lennart; Axén, Iben

    2015-10-19

    Psychological, behavioral and social factors have long been considered important in the development of persistent pain. Little is known about how chiropractic low back pain (LBP) patients compare to other LBP patients in terms of psychological/behavioral characteristics. In this cross-sectional study, the aim was to investigate patients with LBP as regards to psychosocial/behavioral characteristics by describing a chiropractic primary care population and comparing this sample to three other populations using the MPI-S instrument. Thus, four different samples were compared. A: Four hundred eighty subjects from chiropractic primary care clinics. B: One hundred twenty-eight subjects from a gainfully employed population (sick listed with high risk of developing chronicity). C: Two hundred seventy-three subjects from a secondary care rehabilitation clinic. D: Two hundred thirty-five subjects from secondary care clinics. The Swedish version of the Multidimensional Pain Inventory (MPI-S) was used to collect data. Subjects were classified using a cluster analytic strategy into three pre-defined subgroups (named adaptive copers, dysfunctional and interpersonally distressed). The data show statistically significant overall differences across samples for the subgroups based on psychological and behavioral characteristics. The cluster classifications placed (in terms of the proportions of the adaptive copers and dysfunctional subgroups) sample A between B and the two secondary care samples C and D. The chiropractic primary care sample was more affected by pain and worse off with regards to psychological and behavioral characteristics compared to the other primary care sample. Based on our findings from the MPI-S instrument the 4 samples may be considered statistically and clinically different. Sample A comes from an ongoing trial registered at clinical trials.gov; NCT01539863 , February 22, 2012.

  8. Aqueous solubility calculation for petroleum mixtures in soil using comprehensive two-dimensional gas chromatography analysis data.

    PubMed

    Mao, Debin; Lookman, Richard; Van De Weghe, Hendrik; Vanermen, Guido; De Brucker, Nicole; Diels, Ludo

    2009-04-03

    An assessment of aqueous solubility (leaching potential) of soil contaminations with petroleum hydrocarbons (TPH) is important in the context of the evaluation of (migration) risks and soil/groundwater remediation. Field measurements using monitoring wells often overestimate real TPH concentrations in case of presence of pure oil in the screened interval of the well. This paper presents a method to calculate TPH equilibrium concentrations in groundwater using soil analysis by high-performance liquid chromatography followed by comprehensive two-dimensional gas chromatography (HPLC-GCXGC). The oil in the soil sample is divided into 79 defined hydrocarbon fractions on two GCXGC color plots. To each of these fractions a representative water solubility is assigned. Overall equilibrium water solubility of the non-aqueous phase liquid (NAPL) present in the sample and the water phase's chemical composition (in terms of the 79 fractions defined) are then calculated using Raoult's law. The calculation method was validated using soil spiked with 13 different TPH mixtures and 1 field-contaminated soil. Measured water solubilities using a column recirculation equilibration experiment agreed well to calculated equilibrium concentrations and water phase TPH composition.

  9. Fighting obesity campaign in Turkey: evaluation of media campaign efficacy.

    PubMed

    Arikan, Inci; Karakaya, Kağan; Erata, Mustafa; Tüzün, Hakan; Baran, Emine; Levent, Göçmen; Yeşil, Harika Kökalan

    2014-09-01

    This study aims to determine the frequency of behaviour change and related factors generated in the population through the "Fighting Obesity Campaign" of the Turkish Ministry of Health. Twelve statistical regions from NUTS-1 and 18 provinces were selected for the study sample. At least one province from each region was randomly selected, and stratawere defined as urban or rural. Of the sample selected, 2,038 respondents completed a face-to-face survey. Logistic regression analysis was used to analyse the data. Changing behaviour as result of the campaign was defined as the dependent variable. Behaviour change was defined as an individual taking at least one action to increase physical activity, calculate her/his Body Mass Index (BMI) or minimise meal portions. Of the sample selected, 84% of participants lived in urban areas. Of total sample selected, 49.8% were men and 50.2% were women. According to BMI categorisation, 41.4% of participants were underweight or normal weight, 34.3% were overweight and 24.3% were obese. Of the total participants, 85.2% learned about the "Fighting-Obesity Campaign" through television, 28.1% through radio, 11.0% from newspapers, 6.0% from billboards, and 19.2% from other sources. This study revealed that 28.5% of the participants adopted desired behavioural changes after exposure to the campaign. Logistic regression results demonstrated that behaviour change is greater among women, individuals living in urban settings, group of persons approving public spots, obese individuals, and among the 20-39 age group. Media campaigns may cause behavioural changes by increasing motivation to prevent obesity within the target population. Con- tinuing these campaigns can lead to success at the national level.

  10. Evaluating regional trends in ground-water nitrate concentrations of the Columbia Basin ground water management area, Washington

    USGS Publications Warehouse

    Frans, Lonna M.; Helsel, Dennis R.

    2005-01-01

    Trends in nitrate concentrations in water from 474 wells in 17 subregions in the Columbia Basin Ground Water Management Area (GWMA) in three counties in eastern Washington were evaluated using a variety of statistical techniques, including the Friedman test and the Kendall test. The Kendall test was modified from its typical 'seasonal' version into a 'regional' version by using well locations in place of seasons. No statistically significant trends in nitrate concentrations were identified in samples from wells in the GWMA, the three counties, or the 17 subregions from 1998 to 2002 when all data were included in the analysis. For wells in which nitrate concentrations were greater than 10 milligrams per liter (mg/L), however, a significant downward trend of -0.4 mg/L per year was observed between 1998 and 2002 for the GWMA as a whole, as well as for Adams County (-0.35 mg/L per year) and for Franklin County (-0.46 mg/L per year). Trend analysis for a smaller but longer-term 51-well dataset in Franklin County found a statistically significant upward trend in nitrate concentrations of 0.1 mg/L per year between 1986 and 2003. The largest increase of nitrate concentrations occurred between 1986 and 1991. No statistically significant differences were observed in this dataset between 1998 and 2003 indicating that the increase in nitrate concentrations has leveled off.

  11. Effect of the use of combination uridine triphosphate, cytidine monophosphate, and hydroxycobalamin on the recovery of neurosensory disturbance after bilateral sagittal split osteotomy: a randomized, double-blind trial.

    PubMed

    Vieira, C L; Vasconcelos, B C do E; Leão, J C; Laureano Filho, J R

    2016-02-01

    The change in neurosensory lesions that develop after bilateral sagittal split osteotomy (BSSO) was explored, and the influence of the application of combination uridine triphosphate (UTP), cytidine monophosphate (CMP), and hydroxycobalamin (vitamin B12) on patient outcomes was assessed. This was a randomized, controlled, double-blind trial. The study sample comprised 12 patients, each evaluated on both sides (thus 24 sides). All patients fulfilled defined selection criteria. Changes in the lesions were measured both subjectively and objectively. The sample was divided into two patient groups: an experimental group receiving medication and a control group receiving placebo. The statistical analysis was performed using SPSS software. Lesions in both groups improved and no statistically significant difference between the groups was observed at any time. 'Severe' injuries in the experimental group were more likely to exhibit a significant improvement after 6 months. Based on the results of the present study, it is concluded that the combination UTP, CMP, and hydroxycobalamin did not influence recovery from neurosensory disorders. Copyright © 2015. Published by Elsevier Ltd.

  12. Validation of a multilevel sampling device to determine the vertical variability of chlorinated solvent in a contaminated aquifer.

    PubMed

    Barnier, C; Palmier, C; Atteia, O

    2013-01-01

    The vertical heterogeneity of contaminant concentrations in aquifers is well known, but obtaining representative samples is still a subject of debate. In this paper, the question arises from sites where numerous fully screened wells exist and there is a need to define the vertical distribution of contaminants. For this purpose, several wells were investigated with different techniques on a site contaminated with chlorinated solvents. A core-bored well shows that a tetrachloroethene (PCE) phase is sitting on and infiltrating a less permeable layer. Downstream of the cored well, the following sampling techniques were compared on fully screened wells: low flow pumping at several depths, pumping between packers and a new multilevel sampler for fully screened wells. Concerning low flow rate pumping, very low gradients were found, which may be due to the existence of vertical flow inside the well or in the gravel pack. Sampling between packers gave results comparable with the cores, separating a layer with PCE and trichloroethene from another one with cis 1,2-dichloroethene and vinyl chloride as major compounds. Detailed sampling according to pumped volume shows that even between packers, cleaning of the inter-packer volume is necessary before each sampling. Lastly, the proposed new multilevel sampler gives results similar to the packers but has the advantages of much faster sampling and a constant vertical positioning, which is fairly important for long-term monitoring in highly stratified aquifers.

  13. Towards interoperable and reproducible QSAR analyses: Exchange of datasets.

    PubMed

    Spjuth, Ola; Willighagen, Egon L; Guha, Rajarshi; Eklund, Martin; Wikberg, Jarl Es

    2010-06-30

    QSAR is a widely used method to relate chemical structures to responses or properties based on experimental observations. Much effort has been made to evaluate and validate the statistical modeling in QSAR, but these analyses treat the dataset as fixed. An overlooked but highly important issue is the validation of the setup of the dataset, which comprises addition of chemical structures as well as selection of descriptors and software implementations prior to calculations. This process is hampered by the lack of standards and exchange formats in the field, making it virtually impossible to reproduce and validate analyses and drastically constrain collaborations and re-use of data. We present a step towards standardizing QSAR analyses by defining interoperable and reproducible QSAR datasets, consisting of an open XML format (QSAR-ML) which builds on an open and extensible descriptor ontology. The ontology provides an extensible way of uniquely defining descriptors for use in QSAR experiments, and the exchange format supports multiple versioned implementations of these descriptors. Hence, a dataset described by QSAR-ML makes its setup completely reproducible. We also provide a reference implementation as a set of plugins for Bioclipse which simplifies setup of QSAR datasets, and allows for exporting in QSAR-ML as well as old-fashioned CSV formats. The implementation facilitates addition of new descriptor implementations from locally installed software and remote Web services; the latter is demonstrated with REST and XMPP Web services. Standardized QSAR datasets open up new ways to store, query, and exchange data for subsequent analyses. QSAR-ML supports completely reproducible creation of datasets, solving the problems of defining which software components were used and their versions, and the descriptor ontology eliminates confusions regarding descriptors by defining them crisply. This makes is easy to join, extend, combine datasets and hence work collectively, but also allows for analyzing the effect descriptors have on the statistical model's performance. The presented Bioclipse plugins equip scientists with graphical tools that make QSAR-ML easily accessible for the community.

  14. Towards interoperable and reproducible QSAR analyses: Exchange of datasets

    PubMed Central

    2010-01-01

    Background QSAR is a widely used method to relate chemical structures to responses or properties based on experimental observations. Much effort has been made to evaluate and validate the statistical modeling in QSAR, but these analyses treat the dataset as fixed. An overlooked but highly important issue is the validation of the setup of the dataset, which comprises addition of chemical structures as well as selection of descriptors and software implementations prior to calculations. This process is hampered by the lack of standards and exchange formats in the field, making it virtually impossible to reproduce and validate analyses and drastically constrain collaborations and re-use of data. Results We present a step towards standardizing QSAR analyses by defining interoperable and reproducible QSAR datasets, consisting of an open XML format (QSAR-ML) which builds on an open and extensible descriptor ontology. The ontology provides an extensible way of uniquely defining descriptors for use in QSAR experiments, and the exchange format supports multiple versioned implementations of these descriptors. Hence, a dataset described by QSAR-ML makes its setup completely reproducible. We also provide a reference implementation as a set of plugins for Bioclipse which simplifies setup of QSAR datasets, and allows for exporting in QSAR-ML as well as old-fashioned CSV formats. The implementation facilitates addition of new descriptor implementations from locally installed software and remote Web services; the latter is demonstrated with REST and XMPP Web services. Conclusions Standardized QSAR datasets open up new ways to store, query, and exchange data for subsequent analyses. QSAR-ML supports completely reproducible creation of datasets, solving the problems of defining which software components were used and their versions, and the descriptor ontology eliminates confusions regarding descriptors by defining them crisply. This makes is easy to join, extend, combine datasets and hence work collectively, but also allows for analyzing the effect descriptors have on the statistical model's performance. The presented Bioclipse plugins equip scientists with graphical tools that make QSAR-ML easily accessible for the community. PMID:20591161

  15. 42 CFR 402.109 - Statistical sampling.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 42 Public Health 2 2011-10-01 2011-10-01 false Statistical sampling. 402.109 Section 402.109... Statistical sampling. (a) Purpose. CMS or OIG may introduce the results of a statistical sampling study to... or caused to be presented. (b) Prima facie evidence. The results of the statistical sampling study...

  16. 42 CFR 402.109 - Statistical sampling.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 42 Public Health 2 2010-10-01 2010-10-01 false Statistical sampling. 402.109 Section 402.109... Statistical sampling. (a) Purpose. CMS or OIG may introduce the results of a statistical sampling study to... or caused to be presented. (b) Prima facie evidence. The results of the statistical sampling study...

  17. The woman's birth experience---the effect of interpersonal relationships and continuity of care.

    PubMed

    Dahlberg, Unn; Aune, Ingvild

    2013-04-01

    the aim of the present study was to gain a deeper understanding of how relational continuity in the childbearing process may influence the woman's birth experience. RESEARCH DESIGN/SETTING: a Q-methodological approach was chosen, as it allows the researcher to systematically assess subjectivity. 23 women were invited to sort a sample of 48 statements regarding their subjective view of birth experience after having participated in a pilot project in Norway, where six midwifery students provided continuity of care to 58 women throughout the childbearing process. The sorting patterns were subsequently factor-analysed, using the statistical software 'PQ' which reveals one strong and one weaker factor. The consensus statements and the defining statements for the two factors were later interpreted. both factors seemed to represent experiences of psychological trust and a feeling of team work along with the midwifery student. Both factors indicated the importance of quality in the relation. Factor one represented experiences of presence and emotional support in the relationship. It also represented a feeling of personal growth for the women. Factor two was defined by experiences of predictability in the relation and process, as well as the feeling of interdependency in the relation. According to quality in the relation, women defining factor two experienced that the content, not only the continuity in the relation, was important for the birth experience. relational continuity is a key concept in the context of a positive birth experience. Quality in the relation gives the woman a possibility to experience positivity during the childbearing process. Continuity in care and personal growth related to birth promote empowerment for both the woman and her partner. Relational continuity gives an opportunity for midwives to provide care in a more holistic manner. Copyright © 2012 Elsevier Ltd. All rights reserved.

  18. Statistical photocalibration of photodetectors for radiometry without calibrated light sources

    NASA Astrophysics Data System (ADS)

    Yielding, Nicholas J.; Cain, Stephen C.; Seal, Michael D.

    2018-01-01

    Calibration of CCD arrays for identifying bad pixels and achieving nonuniformity correction is commonly accomplished using dark frames. This kind of calibration technique does not achieve radiometric calibration of the array since only the relative response of the detectors is computed. For this, a second calibration is sometimes utilized by looking at sources with known radiances. This process can be used to calibrate photodetectors as long as a calibration source is available and is well-characterized. A previous attempt at creating a procedure for calibrating a photodetector using the underlying Poisson nature of the photodetection required calculations of the skewness of the photodetector measurements. Reliance on the third moment of measurement meant that thousands of samples would be required in some cases to compute that moment. A photocalibration procedure is defined that requires only first and second moments of the measurements. The technique is applied to image data containing a known light source so that the accuracy of the technique can be surmised. It is shown that the algorithm can achieve accuracy of nearly 2.7% of the predicted number of photons using only 100 frames of image data.

  19. Descriptive statistics: the specification of statistical measures and their presentation in tables and graphs. Part 7 of a series on evaluation of scientific publications.

    PubMed

    Spriestersbach, Albert; Röhrig, Bernd; du Prel, Jean-Baptist; Gerhold-Ay, Aslihan; Blettner, Maria

    2009-09-01

    Descriptive statistics are an essential part of biometric analysis and a prerequisite for the understanding of further statistical evaluations, including the drawing of inferences. When data are well presented, it is usually obvious whether the author has collected and evaluated them correctly and in keeping with accepted practice in the field. Statistical variables in medicine may be of either the metric (continuous, quantitative) or categorical (nominal, ordinal) type. Easily understandable examples are given. Basic techniques for the statistical description of collected data are presented and illustrated with examples. The goal of a scientific study must always be clearly defined. The definition of the target value or clinical endpoint determines the level of measurement of the variables in question. Nearly all variables, whatever their level of measurement, can be usefully presented graphically and numerically. The level of measurement determines what types of diagrams and statistical values are appropriate. There are also different ways of presenting combinations of two independent variables graphically and numerically. The description of collected data is indispensable. If the data are of good quality, valid and important conclusions can already be drawn when they are properly described. Furthermore, data description provides a basis for inferential statistics.

  20. Novel biomarker identification using metabolomic profiling to differentiate radiation necrosis and recurrent tumor following Gamma Knife radiosurgery.

    PubMed

    Lu, Alex Y; Turban, Jack L; Damisah, Eyiyemisi C; Li, Jie; Alomari, Ahmed K; Eid, Tore; Vortmeyer, Alexander O; Chiang, Veronica L

    2017-08-01

    OBJECTIVE Following an initial response of brain metastases to Gamma Knife radiosurgery, regrowth of the enhancing lesion as detected on MRI may represent either radiation necrosis (a treatment-related inflammatory change) or recurrent tumor. Differentiation of radiation necrosis from tumor is vital for management decision making but remains difficult by imaging alone. In this study, gas chromatography with time-of-flight mass spectrometry (GC-TOF) was used to identify differential metabolite profiles of the 2 tissue types obtained by surgical biopsy to find potential targets for noninvasive imaging. METHODS Specimens of pure radiation necrosis and pure tumor obtained from patient brain biopsies were flash-frozen and validated histologically. These formalin-free tissue samples were then analyzed using GC-TOF. The metabolite profiles of radiation necrosis and tumor samples were compared using multivariate and univariate statistical analysis. Statistical significance was defined as p ≤ 0.05. RESULTS For the metabolic profiling, GC-TOF was performed on 7 samples of radiation necrosis and 7 samples of tumor. Of the 141 metabolites identified, 17 (12.1%) were found to be statistically significantly different between comparison groups. Of these metabolites, 6 were increased in tumor, and 11 were increased in radiation necrosis. An unsupervised hierarchical clustering analysis found that tumor had elevated levels of metabolites associated with energy metabolism, whereas radiation necrosis had elevated levels of metabolites that were fatty acids and antioxidants/cofactors. CONCLUSIONS To the authors' knowledge, this is the first tissue-based metabolomics study of radiation necrosis and tumor. Radiation necrosis and recurrent tumor following Gamma Knife radiosurgery for brain metastases have unique metabolite profiles that may be targeted in the future to develop noninvasive metabolic imaging techniques.

Top