Recommended protocols for sampling macrofungi
Gregory M. Mueller; John Paul Schmit; Sabine M. Hubndorf Leif Ryvarden; Thomas E. O' Dell; D. Jean Lodge; Patrick R. Leacock; Milagro Mata; Loengrin Umania; Qiuxin (Florence) Wu; Daniel L. Czederpiltz
2004-01-01
This chapter discusses several issues regarding reommended protocols for sampling macrofungi: Opportunistic sampling of macrofungi, sampling conspicuous macrofungi using fixed-size, sampling small Ascomycetes using microplots, and sampling a fixed number of downed logs.
Shen, You-xin; Liu, Wei-li; Li, Yu-hui; Guan, Hui-lin
2014-01-01
A large number of small-sized samples invariably shows that woody species are absent from forest soil seed banks, leading to a large discrepancy with the seedling bank on the forest floor. We ask: 1) Does this conventional sampling strategy limit the detection of seeds of woody species? 2) Are large sample areas and sample sizes needed for higher recovery of seeds of woody species? We collected 100 samples that were 10 cm (length) × 10 cm (width) × 10 cm (depth), referred to as larger number of small-sized samples (LNSS) in a 1 ha forest plot, and placed them to germinate in a greenhouse, and collected 30 samples that were 1 m × 1 m × 10 cm, referred to as small number of large-sized samples (SNLS) and placed them (10 each) in a nearby secondary forest, shrub land and grass land. Only 15.7% of woody plant species of the forest stand were detected by the 100 LNSS, contrasting with 22.9%, 37.3% and 20.5% woody plant species being detected by SNLS in the secondary forest, shrub land and grassland, respectively. The increased number of species vs. sampled areas confirmed power-law relationships for forest stand, the LNSS and SNLS at all three recipient sites. Our results, although based on one forest, indicate that conventional LNSS did not yield a high percentage of detection for woody species, but SNLS strategy yielded a higher percentage of detection for woody species in the seed bank if samples were exposed to a better field germination environment. A 4 m2 minimum sample area derived from power equations is larger than the sampled area in most studies in the literature. Increased sample size also is needed to obtain an increased sample area if the number of samples is to remain relatively low.
Okubo, Torahiko; Osaki, Takako; Nozaki, Eriko; Uemura, Akira; Sakai, Kouhei; Matushita, Mizue; Matsuo, Junji; Nakamura, Shinji; Kamiya, Shigeru; Yamaguchi, Hiroyuki
2017-01-01
Although human occupancy is a source of airborne bacteria, the role of walkers on bacterial communities in built environments is poorly understood. Therefore, we visualized the impact of walker occupancy combined with other factors (temperature, humidity, atmospheric pressure, dust particles) on airborne bacterial features in the Sapporo underground pedestrian space in Sapporo, Japan. Air samples (n = 18; 4,800L/each sample) were collected at 8:00 h to 20:00 h on 3 days (regular sampling) and at early morning / late night (5:50 h to 7:50 h / 22:15 h to 24:45 h) on a day (baseline sampling), and the number of CFUs (colony forming units) OTUs (operational taxonomic units) and other factors were determined. The results revealed that temperature, humidity, and atmospheric pressure changed with weather. The number of walkers increased greatly in the morning and evening on each regular sampling day, although total walker numbers did not differ significantly among regular sampling days. A slight increase in small dust particles (0.3-0.5μm) was observed on the days with higher temperature regardless of regular or baseline sampling. At the period on regular sampling, CFU levels varied irregularly among days, and the OTUs of 22-phylum types were observed, with the majority being from Firmicutes or Proteobacteria (γ-), including Staphylococcus sp. derived from human individuals. The data obtained from regular samplings reveled that although no direct interaction of walker occupancy and airborne CFU and OTU features was observed upon Pearson's correlation analysis, cluster analysis indicated an obvious lineage consisting of walker occupancy, CFU numbers, OTU types, small dust particles, and seasonal factors (including temperature and humidity). Meanwhile, at the period on baseline sampling both walker and CFU numbers were similarly minimal. Taken together, the results revealed a positive correlation of walker occupancy with airborne bacteria that increased with increases in temperature and humidity in the presence of airborne small particles. Moreover, the results indicated that small dust particles at high temperature and humidity may be a crucial factor responsible for stabilizing the bacteria released from walkers in built environments. The findings presented herein advance our knowledge and understanding of the relationship between humans and bacterial communities in built environments, and will help improve public health in urban communities.
Okubo, Torahiko; Osaki, Takako; Nozaki, Eriko; Uemura, Akira; Sakai, Kouhei; Matushita, Mizue; Matsuo, Junji; Nakamura, Shinji; Kamiya, Shigeru
2017-01-01
Although human occupancy is a source of airborne bacteria, the role of walkers on bacterial communities in built environments is poorly understood. Therefore, we visualized the impact of walker occupancy combined with other factors (temperature, humidity, atmospheric pressure, dust particles) on airborne bacterial features in the Sapporo underground pedestrian space in Sapporo, Japan. Air samples (n = 18; 4,800L/each sample) were collected at 8:00 h to 20:00 h on 3 days (regular sampling) and at early morning / late night (5:50 h to 7:50 h / 22:15 h to 24:45 h) on a day (baseline sampling), and the number of CFUs (colony forming units) OTUs (operational taxonomic units) and other factors were determined. The results revealed that temperature, humidity, and atmospheric pressure changed with weather. The number of walkers increased greatly in the morning and evening on each regular sampling day, although total walker numbers did not differ significantly among regular sampling days. A slight increase in small dust particles (0.3–0.5μm) was observed on the days with higher temperature regardless of regular or baseline sampling. At the period on regular sampling, CFU levels varied irregularly among days, and the OTUs of 22-phylum types were observed, with the majority being from Firmicutes or Proteobacteria (γ-), including Staphylococcus sp. derived from human individuals. The data obtained from regular samplings reveled that although no direct interaction of walker occupancy and airborne CFU and OTU features was observed upon Pearson's correlation analysis, cluster analysis indicated an obvious lineage consisting of walker occupancy, CFU numbers, OTU types, small dust particles, and seasonal factors (including temperature and humidity). Meanwhile, at the period on baseline sampling both walker and CFU numbers were similarly minimal. Taken together, the results revealed a positive correlation of walker occupancy with airborne bacteria that increased with increases in temperature and humidity in the presence of airborne small particles. Moreover, the results indicated that small dust particles at high temperature and humidity may be a crucial factor responsible for stabilizing the bacteria released from walkers in built environments. The findings presented herein advance our knowledge and understanding of the relationship between humans and bacterial communities in built environments, and will help improve public health in urban communities. PMID:28922412
Accuracy or precision: Implications of sample design and methodology on abundance estimation
Kowalewski, Lucas K.; Chizinski, Christopher J.; Powell, Larkin A.; Pope, Kevin L.; Pegg, Mark A.
2015-01-01
Sampling by spatially replicated counts (point-count) is an increasingly popular method of estimating population size of organisms. Challenges exist when sampling by point-count method, and it is often impractical to sample entire area of interest and impossible to detect every individual present. Ecologists encounter logistical limitations that force them to sample either few large-sample units or many small sample-units, introducing biases to sample counts. We generated a computer environment and simulated sampling scenarios to test the role of number of samples, sample unit area, number of organisms, and distribution of organisms in the estimation of population sizes using N-mixture models. Many sample units of small area provided estimates that were consistently closer to true abundance than sample scenarios with few sample units of large area. However, sample scenarios with few sample units of large area provided more precise abundance estimates than abundance estimates derived from sample scenarios with many sample units of small area. It is important to consider accuracy and precision of abundance estimates during the sample design process with study goals and objectives fully recognized, although and with consequence, consideration of accuracy and precision of abundance estimates is often an afterthought that occurs during the data analysis process.
Eblen, Denise R; Barlow, Kristina E; Naugle, Alecia Larew
2006-11-01
The U.S. Food Safety and Inspection Service (FSIS) pathogen reduction-hazard analysis critical control point systems final rule, published in 1996, established Salmonella performance standards for broiler chicken, cow and bull, market hog, and steer and heifer carcasses and for ground beef, chicken, and turkey meat. In 1998, the FSIS began testing to verify that establishments are meeting performance standards. Samples are collected in sets in which the number of samples is defined but varies according to product class. A sample set fails when the number of positive Salmonella samples exceeds the maximum number of positive samples allowed under the performance standard. Salmonella sample sets collected at 1,584 establishments from 1998 through 2003 were examined to identify factors associated with failure of one or more sets. Overall, 1,282 (80.9%) of establishments never had failed sets. In establishments that did experience set failure(s), generally the failed sets were collected early in the establishment testing history, with the exception of broiler establishments where failure(s) occurred both early and late in the course of testing. Small establishments were more likely to have experienced a set failure than were large or very small establishments, and broiler establishments were more likely to have failed than were ground beef, market hog, or steer-heifer establishments. Agency response to failed Salmonella sample sets in the form of in-depth verification reviews and related establishment-initiated corrective actions have likely contributed to declines in the number of establishments that failed sets. A focus on food safety measures in small establishments and broiler processing establishments should further reduce the number of sample sets that fail to meet the Salmonella performance standard.
Anomaly detection in reconstructed quantum states using a machine-learning technique
NASA Astrophysics Data System (ADS)
Hara, Satoshi; Ono, Takafumi; Okamoto, Ryo; Washio, Takashi; Takeuchi, Shigeki
2014-02-01
The accurate detection of small deviations in given density matrices is important for quantum information processing. Here we propose a method based on the concept of data mining. We demonstrate that the proposed method can more accurately detect small erroneous deviations in reconstructed density matrices, which contain intrinsic fluctuations due to the limited number of samples, than a naive method of checking the trace distance from the average of the given density matrices. This method has the potential to be a key tool in broad areas of physics where the detection of small deviations of quantum states reconstructed using a limited number of samples is essential.
A high-throughput microRNA expression profiling system.
Guo, Yanwen; Mastriano, Stephen; Lu, Jun
2014-01-01
As small noncoding RNAs, microRNAs (miRNAs) regulate diverse biological functions, including physiological and pathological processes. The expression and deregulation of miRNA levels contain rich information with diagnostic and prognostic relevance and can reflect pharmacological responses. The increasing interest in miRNA-related research demands global miRNA expression profiling on large numbers of samples. We describe here a robust protocol that supports high-throughput sample labeling and detection on hundreds of samples simultaneously. This method employs 96-well-based miRNA capturing from total RNA samples and on-site biochemical reactions, coupled with bead-based detection in 96-well format for hundreds of miRNAs per sample. With low-cost, high-throughput, high detection specificity, and flexibility to profile both small and large numbers of samples, this protocol can be adapted in a wide range of laboratory settings.
Tung, Yi-Chung; Torisawa, Yu-suke; Futai, Nobuyuki; Takayama, Shuichi
2007-11-01
This paper describes a micro flow cytometer system designed for efficient and non-damaging analysis of samples with small numbers of precious cells. The system utilizes actuation of Braille-display pins for micro-scale fluid manipulation and a fluorescence microscope with a CCD camera for optical detection. The microfluidic chip is fully disposable and is composed of a polydimethylsiloxane (PDMS) slab with microchannel features sealed against a thin deformable PDMS membrane. The channels are designed with diffusers to alleviate pulsatile flow behaviors inherent in pin actuator-based peristaltic pumping schemes to maximize hydrodynamic focusing of samples with minimal disturbances in the laminar streams within the channel. A funnel connected to the microfluidic channel is designed for efficient loading of samples with small number of cells and is also positioned on the chip to prevent physical damages of the samples by the squeezing actions of Braille pins during actuation. The sample loading scheme was characterized by both computational fluidic dynamics (CFD) simulation and experimental observation. A fluorescein solution was first used for flow field investigation, followed by use of fluorescence beads with known relative intensities for optical detection performance calibration. Murine myoblast cells (C2C12) were exploited to investigate cell viability for the sample loading scheme of the device. Furthermore, human promyelocytic leukemia (HL60) cells stained by hypotonic DNA staining buffer were also tested in the system for cell cycle analysis. The ability to efficiently analyze cellular samples where the number of cells is small was demonstrated by analyzing cells from a single embryoid body derived from mouse embryonic stem cells. Consequently, the designed microfluidic device reported in this paper is promising for easy-to-use, small sample size flow cytometric analysis, and has potential to be further integrated with other Braille display-based microfluidic devices to facilitate a multi-functional lab-on-a-chip for mammalian cell manipulations.
Correlated Observations, the Law of Small Numbers and Bank Runs
2016-01-01
Empirical descriptions and studies suggest that generally depositors observe a sample of previous decisions before deciding if to keep their funds deposited or to withdraw them. These observed decisions may exhibit different degrees of correlation across depositors. In our model depositors decide sequentially and are assumed to follow the law of small numbers in the sense that they believe that a bank run is underway if the number of observed withdrawals in their sample is large. Theoretically, with highly correlated samples and infinite depositors runs occur with certainty, while with random samples it needs not be the case, as for many parameter settings the likelihood of bank runs is zero. We investigate the intermediate cases and find that i) decreasing the correlation and ii) increasing the sample size reduces the likelihood of bank runs, ceteris paribus. Interestingly, the multiplicity of equilibria, a feature of the canonical Diamond-Dybvig model that we use also, disappears almost completely in our setup. Our results have relevant policy implications. PMID:27035435
Correlated Observations, the Law of Small Numbers and Bank Runs.
Horváth, Gergely; Kiss, Hubert János
2016-01-01
Empirical descriptions and studies suggest that generally depositors observe a sample of previous decisions before deciding if to keep their funds deposited or to withdraw them. These observed decisions may exhibit different degrees of correlation across depositors. In our model depositors decide sequentially and are assumed to follow the law of small numbers in the sense that they believe that a bank run is underway if the number of observed withdrawals in their sample is large. Theoretically, with highly correlated samples and infinite depositors runs occur with certainty, while with random samples it needs not be the case, as for many parameter settings the likelihood of bank runs is zero. We investigate the intermediate cases and find that i) decreasing the correlation and ii) increasing the sample size reduces the likelihood of bank runs, ceteris paribus. Interestingly, the multiplicity of equilibria, a feature of the canonical Diamond-Dybvig model that we use also, disappears almost completely in our setup. Our results have relevant policy implications.
Conversion of Small Algal Oil Sample to JP-8
2012-01-01
cracking of Algal Oil to SPK Hydroprocessing Lab Plant uop Nitrogen Hydrogen Product ., __ Small Scale Lab Hydprocessing plant - Down flow trickle ... bed configuration - Capable of retaining 25 cc of catalyst bed Meter UOP ·CONFIDENTIAL File Number The catalytic deoxygenation stage of the...content which combined with the samples acidity, is a challenge to reactor metallurgy. None the less, an attempt was made to convert this sample to
The Metals in Replicate Samples data set contains the analytical results of measurements of up to 2 metals in 172 replicate (duplicate) samples from 86 households. Measurements were made in samples of blood. Duplicate samples for a small percentage of the total number of sample...
Short-term memory for responses: the "choose-small" effect.
Fetterman, J G; MacEwen, D
1989-01-01
Pigeons' short-term memory for fixed-ratio requirements was assessed using a delayed symbolic matching-to-sample procedure. Different choices were reinforced after fixed-ratio 10 and fixed-ratio 40 requirements, and delays of 0, 5, or 20 s were sometimes placed between sample ratios and choice. All birds made disproportionate numbers of responses to the small-ratio choice alternative when delays were interposed between ratios and choice, and this bias increased as a function of delay. Preference for the small fixed-ratio alternative was also observed on "no-sample" trials, during which the choice alternatives were presented without a prior sample ratio. This "choose-small" bias is analogous to results obtained by Spetch and Wilkie (1983) with event duration as the discriminative stimulus. The choose-small bias was attenuated when the houselight was turned on during delays, but overall accuracy was not influenced systematically by the houselight manipulation. PMID:2584917
NASA Astrophysics Data System (ADS)
Herd, C. D. K.; Tornabene, L. L.; Bowling, T. J.; Walton, E. L.; Sharp, T. G.; Melosh, H. J.; Hamilton, J. S.; Viviano, C. E.; Ehlmann, B. L.
2018-04-01
We have made advances in constraining the potential source craters of the martian meteorites to a relatively small number. Our results have implications for Mars chronology and the prioritization of samples for Mars Sample Return.
Tyree, M T; Dixon, M A; Tyree, E L; Johnson, R
1984-08-01
Measurements are reported of ultrasonic acoustic emissions (AEs) measured from sapwood samples of Thuja occidentalis L. and Tsuga canadensis (L.) Carr. during air dehydration. The measurements were undertaken to test the following three hypotheses: (a) Each cavitation event produces one ultrasonic AE. (b) Large tracheids are more likely to cavitate than small tracheids. (c) When stem water potentials are >-0.4 MPa, a significant fraction of the water content of sapwood is held by ;capillary forces.' The last two hypotheses were recently discussed at length by M. H. Zimmermann. Experimental evidence consistent with all three hypotheses was obtained. The evidence for each hypothesis respectively is: (a) the cumulative number of AEs nearly equals the number of tracheids in small samples; (b) more water is lost per AE event at the beginning of the dehydration process than at the end, and (c) sapwood samples dehydrated from an initial water potential of 0 MPa lost significantly more water before AEs started than lost by samples dehydrated from an initial water potential of about -0.4 MPa. The extra water held by fully hydrated sapwood samples may have been capillary water as defined by Zimmerman.We also report an improved method for the measurement of the ;intensity' of ultrasonic AEs. Intensity is defined here as the area under the positive spikes of the AE signal (plotted as voltage versus time). This method was applied to produce a frequency histogram of the number of AEs versus intensity. A large fraction of the total number of AEs were of low intensity even in small samples (4 mm diameter by 10 mm length). This suggests that the effective ;listening distance' for most AEs was less than 5 to 10 mm.
Ultrasonic Acoustic Emissions from the Sapwood of Cedar and Hemlock 1
Tyree, Melvin T.; Dixon, Michael A.; Tyree, E. Loeta; Johnson, Robert
1984-01-01
Measurements are reported of ultrasonic acoustic emissions (AEs) measured from sapwood samples of Thuja occidentalis L. and Tsuga canadensis (L.) Carr. during air dehydration. The measurements were undertaken to test the following three hypotheses: (a) Each cavitation event produces one ultrasonic AE. (b) Large tracheids are more likely to cavitate than small tracheids. (c) When stem water potentials are >−0.4 MPa, a significant fraction of the water content of sapwood is held by `capillary forces.' The last two hypotheses were recently discussed at length by M. H. Zimmermann. Experimental evidence consistent with all three hypotheses was obtained. The evidence for each hypothesis respectively is: (a) the cumulative number of AEs nearly equals the number of tracheids in small samples; (b) more water is lost per AE event at the beginning of the dehydration process than at the end, and (c) sapwood samples dehydrated from an initial water potential of 0 MPa lost significantly more water before AEs started than lost by samples dehydrated from an initial water potential of about −0.4 MPa. The extra water held by fully hydrated sapwood samples may have been capillary water as defined by Zimmerman. We also report an improved method for the measurement of the `intensity' of ultrasonic AEs. Intensity is defined here as the area under the positive spikes of the AE signal (plotted as voltage versus time). This method was applied to produce a frequency histogram of the number of AEs versus intensity. A large fraction of the total number of AEs were of low intensity even in small samples (4 mm diameter by 10 mm length). This suggests that the effective `listening distance' for most AEs was less than 5 to 10 mm. PMID:16663774
Total Extracellular Small RNA Profiles from Plasma, Saliva, and Urine of Healthy Subjects
Yeri, Ashish; Courtright, Amanda; Reiman, Rebecca; Carlson, Elizabeth; Beecroft, Taylor; Janss, Alex; Siniard, Ashley; Richholt, Ryan; Balak, Chris; Rozowsky, Joel; Kitchen, Robert; Hutchins, Elizabeth; Winarta, Joseph; McCoy, Roger; Anastasi, Matthew; Kim, Seungchan; Huentelman, Matthew; Van Keuren-Jensen, Kendall
2017-01-01
Interest in circulating RNAs for monitoring and diagnosing human health has grown significantly. There are few datasets describing baseline expression levels for total cell-free circulating RNA from healthy control subjects. In this study, total extracellular RNA (exRNA) was isolated and sequenced from 183 plasma samples, 204 urine samples and 46 saliva samples from 55 male college athletes ages 18–25 years. Many participants provided more than one sample, allowing us to investigate variability in an individual’s exRNA expression levels over time. Here we provide a systematic analysis of small exRNAs present in each biofluid, as well as an analysis of exogenous RNAs. The small RNA profile of each biofluid is distinct. We find that a large number of RNA fragments in plasma (63%) and urine (54%) have sequences that are assigned to YRNA and tRNA fragments respectively. Surprisingly, while many miRNAs can be detected, there are few miRNAs that are consistently detected in all samples from a single biofluid, and profiles of miRNA are different for each biofluid. Not unexpectedly, saliva samples have high levels of exogenous sequence that can be traced to bacteria. These data significantly contribute to the current number of sequenced exRNA samples from normal healthy individuals. PMID:28303895
Zhang, Cuicui; Liang, Xuefeng; Matsuyama, Takashi
2014-12-08
Multi-camera networks have gained great interest in video-based surveillance systems for security monitoring, access control, etc. Person re-identification is an essential and challenging task in multi-camera networks, which aims to determine if a given individual has already appeared over the camera network. Individual recognition often uses faces as a trial and requires a large number of samples during the training phrase. This is difficult to fulfill due to the limitation of the camera hardware system and the unconstrained image capturing conditions. Conventional face recognition algorithms often encounter the "small sample size" (SSS) problem arising from the small number of training samples compared to the high dimensionality of the sample space. To overcome this problem, interest in the combination of multiple base classifiers has sparked research efforts in ensemble methods. However, existing ensemble methods still open two questions: (1) how to define diverse base classifiers from the small data; (2) how to avoid the diversity/accuracy dilemma occurring during ensemble. To address these problems, this paper proposes a novel generic learning-based ensemble framework, which augments the small data by generating new samples based on a generic distribution and introduces a tailored 0-1 knapsack algorithm to alleviate the diversity/accuracy dilemma. More diverse base classifiers can be generated from the expanded face space, and more appropriate base classifiers are selected for ensemble. Extensive experimental results on four benchmarks demonstrate the higher ability of our system to cope with the SSS problem compared to the state-of-the-art system.
Zhang, Cuicui; Liang, Xuefeng; Matsuyama, Takashi
2014-01-01
Multi-camera networks have gained great interest in video-based surveillance systems for security monitoring, access control, etc. Person re-identification is an essential and challenging task in multi-camera networks, which aims to determine if a given individual has already appeared over the camera network. Individual recognition often uses faces as a trial and requires a large number of samples during the training phrase. This is difficult to fulfill due to the limitation of the camera hardware system and the unconstrained image capturing conditions. Conventional face recognition algorithms often encounter the “small sample size” (SSS) problem arising from the small number of training samples compared to the high dimensionality of the sample space. To overcome this problem, interest in the combination of multiple base classifiers has sparked research efforts in ensemble methods. However, existing ensemble methods still open two questions: (1) how to define diverse base classifiers from the small data; (2) how to avoid the diversity/accuracy dilemma occurring during ensemble. To address these problems, this paper proposes a novel generic learning-based ensemble framework, which augments the small data by generating new samples based on a generic distribution and introduces a tailored 0–1 knapsack algorithm to alleviate the diversity/accuracy dilemma. More diverse base classifiers can be generated from the expanded face space, and more appropriate base classifiers are selected for ensemble. Extensive experimental results on four benchmarks demonstrate the higher ability of our system to cope with the SSS problem compared to the state-of-the-art system. PMID:25494350
Optimal number of features as a function of sample size for various classification rules.
Hua, Jianping; Xiong, Zixiang; Lowey, James; Suh, Edward; Dougherty, Edward R
2005-04-15
Given the joint feature-label distribution, increasing the number of features always results in decreased classification error; however, this is not the case when a classifier is designed via a classification rule from sample data. Typically (but not always), for fixed sample size, the error of a designed classifier decreases and then increases as the number of features grows. The potential downside of using too many features is most critical for small samples, which are commonplace for gene-expression-based classifiers for phenotype discrimination. For fixed sample size and feature-label distribution, the issue is to find an optimal number of features. Since only in rare cases is there a known distribution of the error as a function of the number of features and sample size, this study employs simulation for various feature-label distributions and classification rules, and across a wide range of sample and feature-set sizes. To achieve the desired end, finding the optimal number of features as a function of sample size, it employs massively parallel computation. Seven classifiers are treated: 3-nearest-neighbor, Gaussian kernel, linear support vector machine, polynomial support vector machine, perceptron, regular histogram and linear discriminant analysis. Three Gaussian-based models are considered: linear, nonlinear and bimodal. In addition, real patient data from a large breast-cancer study is considered. To mitigate the combinatorial search for finding optimal feature sets, and to model the situation in which subsets of genes are co-regulated and correlation is internal to these subsets, we assume that the covariance matrix of the features is blocked, with each block corresponding to a group of correlated features. Altogether there are a large number of error surfaces for the many cases. These are provided in full on a companion website, which is meant to serve as resource for those working with small-sample classification. For the companion website, please visit http://public.tgen.org/tamu/ofs/ e-dougherty@ee.tamu.edu.
Questioning the utility of pooling samples in microarray experiments with cell lines.
Lusa, L; Cappelletti, V; Gariboldi, M; Ferrario, C; De Cecco, L; Reid, J F; Toffanin, S; Gallus, G; McShane, L M; Daidone, M G; Pierotti, M A
2006-01-01
We describe a microarray experiment using the MCF-7 breast cancer cell line in two different experimental conditions for which the same number of independent pools as the number of individual samples was hybridized on Affymetrix GeneChips. Unexpectedly, when using individual samples, the number of probe sets found to be differentially expressed between treated and untreated cells was about three times greater than that found using pools. These findings indicate that pooling samples in microarray experiments where the biological variability is expected to be small might not be helpful and could even decrease one's ability to identify differentially expressed genes.
NASA Astrophysics Data System (ADS)
Ulyanov, Sergey; Ulianova, Onega; Filonova, Nadezhda; Moiseeva, Yulia; Zaitsev, Sergey; Saltykov, Yury; Polyanina, Tatiana; Lyapina, Anna; Kalduzova, Irina; Larionova, Olga; Utz, Sergey; Feodorova, Valentina
2018-04-01
Theory of diffusing wave spectroscopy has been firstly adapted to the problem of rapid detection of Chlamydia trachomatis bacteria in blood samples of Chlamydia patients. Formula for correlation function of temporal fluctuations of speckle intensity is derived for the case of small number of scattering events. Dependence of bandwidth of spectrum on average number of scatterers is analyzed. Set-up for detection of the presence of C. trachomatis cells in aqueous suspension is designed. Good agreement between theoretical results and experimental data is shown. Possibility of detection of the presence of C. trachomatis cells in probing volume using diffusing wave spectroscopy with a small number of scatterers is successfully demonstrated for the first time.
Simulation analyses of space use: Home range estimates, variability, and sample size
Bekoff, Marc; Mech, L. David
1984-01-01
Simulations of space use by animals were run to determine the relationship among home range area estimates, variability, and sample size (number of locations). As sample size increased, home range size increased asymptotically, whereas variability decreased among mean home range area estimates generated by multiple simulations for the same sample size. Our results suggest that field workers should ascertain between 100 and 200 locations in order to estimate reliably home range area. In some cases, this suggested guideline is higher than values found in the few published studies in which the relationship between home range area and number of locations is addressed. Sampling differences for small species occupying relatively small home ranges indicate that fewer locations may be sufficient to allow for a reliable estimate of home range. Intraspecific variability in social status (group member, loner, resident, transient), age, sex, reproductive condition, and food resources also have to be considered, as do season, habitat, and differences in sampling and analytical methods. Comparative data still are needed.
Counting glomeruli and podocytes: rationale and methodologies
Puelles, Victor G.; Bertram, John F.
2015-01-01
Purpose of review There is currently much interest in the numbers of both glomeruli and podocytes. This interest stems from greater understanding of the effects of suboptimal fetal events on nephron endowment, the associations between low nephron number and chronic cardiovascular and kidney disease in adults, and the emergence of the podocyte depletion hypothesis. Recent findings Obtaining accurate and precise estimates of glomerular and podocyte number has proven surprisingly difficult. When whole kidneys or large tissue samples are available, design-based stereological methods are considered gold-standard because they are based on principles that negate systematic bias. However, these methods are often tedious and time-consuming, and oftentimes inapplicable when dealing with small samples such as biopsies. Therefore, novel methods suitable for small tissue samples, and innovative approaches to facilitate high through put measurements, such as magnetic resonance imaging (MRI) to estimate glomerular number and flow cytometry to estimate podocyte number, have recently been described. Summary This review describes current gold-standard methods for estimating glomerular and podocyte number, as well as methods developed in the past 3 years. We are now better placed than ever before to accurately and precisely estimate glomerular and podocyte number, and to examine relationships between these measurements and kidney health and disease. PMID:25887899
Automated storm water sampling on small watersheds
Harmel, R.D.; King, K.W.; Slade, R.M.
2003-01-01
Few guidelines are currently available to assist in designing appropriate automated storm water sampling strategies for small watersheds. Therefore, guidance is needed to develop strategies that achieve an appropriate balance between accurate characterization of storm water quality and loads and limitations of budget, equipment, and personnel. In this article, we explore the important sampling strategy components (minimum flow threshold, sampling interval, and discrete versus composite sampling) and project-specific considerations (sampling goal, sampling and analysis resources, and watershed characteristics) based on personal experiences and pertinent field and analytical studies. These components and considerations are important in achieving the balance between sampling goals and limitations because they determine how and when samples are taken and the potential sampling error. Several general recommendations are made, including: setting low minimum flow thresholds, using flow-interval or variable time-interval sampling, and using composite sampling to limit the number of samples collected. Guidelines are presented to aid in selection of an appropriate sampling strategy based on user's project-specific considerations. Our experiences suggest these recommendations should allow implementation of a successful sampling strategy for most small watershed sampling projects with common sampling goals.
Carpenter, Danielle; Walker, Susan; Prescott, Natalie; Schalkwijk, Joost; Armour, John Al
2011-08-18
Copy number variation (CNV) contributes to the variation observed between individuals and can influence human disease progression, but the accurate measurement of individual copy numbers is technically challenging. In the work presented here we describe a modification to a previously described paralogue ratio test (PRT) method for genotyping the CCL3L1/CCL4L1 copy variable region, which we use to ascertain CCL3L1/CCL4L1 copy number in 1581 European samples. As the products of CCL3L1 and CCL4L1 potentially play a role in autoimmunity we performed case control association studies with Crohn's disease, rheumatoid arthritis and psoriasis clinical cohorts. We evaluate the PRT methodology used, paying particular attention to accuracy and precision, and highlight the problems of differential bias in copy number measurements. Our PRT methods for measuring copy number were of sufficient precision to detect very slight but systematic differential bias between results from case and control DNA samples in one study. We find no evidence for an association between CCL3L1 copy number and Crohn's disease, rheumatoid arthritis or psoriasis. Differential bias of this small magnitude, but applied systematically across large numbers of samples, would create a serious risk of false positive associations in copy number, if measured using methods of lower precision, or methods relying on single uncorroborated measurements. In this study the small differential bias detected by PRT in one sample set was resolved by a simple pre-treatment by restriction enzyme digestion.
2011-01-01
Background Copy number variation (CNV) contributes to the variation observed between individuals and can influence human disease progression, but the accurate measurement of individual copy numbers is technically challenging. In the work presented here we describe a modification to a previously described paralogue ratio test (PRT) method for genotyping the CCL3L1/CCL4L1 copy variable region, which we use to ascertain CCL3L1/CCL4L1 copy number in 1581 European samples. As the products of CCL3L1 and CCL4L1 potentially play a role in autoimmunity we performed case control association studies with Crohn's disease, rheumatoid arthritis and psoriasis clinical cohorts. Results We evaluate the PRT methodology used, paying particular attention to accuracy and precision, and highlight the problems of differential bias in copy number measurements. Our PRT methods for measuring copy number were of sufficient precision to detect very slight but systematic differential bias between results from case and control DNA samples in one study. We find no evidence for an association between CCL3L1 copy number and Crohn's disease, rheumatoid arthritis or psoriasis. Conclusions Differential bias of this small magnitude, but applied systematically across large numbers of samples, would create a serious risk of false positive associations in copy number, if measured using methods of lower precision, or methods relying on single uncorroborated measurements. In this study the small differential bias detected by PRT in one sample set was resolved by a simple pre-treatment by restriction enzyme digestion. PMID:21851606
Nieva, Jorge; Wendel, Marco; Luttgen, Madelyn S; Marrinucci, Dena; Bazhenova, Lyudmila; Kolatkar, Anand; Santala, Roger; Whittenberger, Brock; Burke, James; Torrey, Melissa; Bethel, Kelly; Kuhn, Peter
2012-02-01
Sampling circulating tumor cells (CTCs) from peripheral blood is ideally accomplished using assays that detect high numbers of cells and preserve them for downstream characterization. We sought to evaluate a method using enrichment free fluorescent labeling of CTCs followed by automated digital microscopy in patients with non-small cell lung cancer. Twenty-eight patients with non-small cell lung cancer and hematogenously seeded metastasis were analyzed with multiple blood draws. We detected CTCs in 68% of analyzed samples and found a propensity for increased CTC detection as the disease progressed in individual patients. CTCs were present at a median concentration of 1.6 CTCs ml⁻¹ of analyzed blood in the patient population. Higher numbers of detected CTCs were associated with an unfavorable prognosis.
A Maximum Entropy Test for Evaluating Higher-Order Correlations in Spike Counts
Onken, Arno; Dragoi, Valentin; Obermayer, Klaus
2012-01-01
Evaluating the importance of higher-order correlations of neural spike counts has been notoriously hard. A large number of samples are typically required in order to estimate higher-order correlations and resulting information theoretic quantities. In typical electrophysiology data sets with many experimental conditions, however, the number of samples in each condition is rather small. Here we describe a method that allows to quantify evidence for higher-order correlations in exactly these cases. We construct a family of reference distributions: maximum entropy distributions, which are constrained only by marginals and by linear correlations as quantified by the Pearson correlation coefficient. We devise a Monte Carlo goodness-of-fit test, which tests - for a given divergence measure of interest - whether the experimental data lead to the rejection of the null hypothesis that it was generated by one of the reference distributions. Applying our test to artificial data shows that the effects of higher-order correlations on these divergence measures can be detected even when the number of samples is small. Subsequently, we apply our method to spike count data which were recorded with multielectrode arrays from the primary visual cortex of anesthetized cat during an adaptation experiment. Using mutual information as a divergence measure we find that there are spike count bin sizes at which the maximum entropy hypothesis can be rejected for a substantial number of neuronal pairs. These results demonstrate that higher-order correlations can matter when estimating information theoretic quantities in V1. They also show that our test is able to detect their presence in typical in-vivo data sets, where the number of samples is too small to estimate higher-order correlations directly. PMID:22685392
Sampled-data chain-observer design for a class of delayed nonlinear systems
NASA Astrophysics Data System (ADS)
Kahelras, M.; Ahmed-Ali, T.; Giri, F.; Lamnabhi-Lagarrigue, F.
2018-05-01
The problem of observer design is addressed for a class of triangular nonlinear systems with not-necessarily small delay and sampled output measurements. One more difficulty is that the system state matrix is dependent on the un-delayed output signal which is not accessible to measurement, making existing observers inapplicable. A new chain observer, composed of m elementary observers in series, is designed to compensate for output sampling and arbitrary large delays. The larger the time-delay the larger the number m. Each elementary observer includes an output predictor that is conceived to compensate for the effects of output sampling and a fractional delay. The predictors are defined by first-order ordinary differential equations (ODEs) much simpler than those of existing predictors which involve both output and state predictors. Using a small gain type analysis, sufficient conditions for the observer to be exponentially convergent are established in terms of the minimal number m of elementary observers and the maximum sampling interval.
Support vector regression to predict porosity and permeability: Effect of sample size
NASA Astrophysics Data System (ADS)
Al-Anazi, A. F.; Gates, I. D.
2012-02-01
Porosity and permeability are key petrophysical parameters obtained from laboratory core analysis. Cores, obtained from drilled wells, are often few in number for most oil and gas fields. Porosity and permeability correlations based on conventional techniques such as linear regression or neural networks trained with core and geophysical logs suffer poor generalization to wells with only geophysical logs. The generalization problem of correlation models often becomes pronounced when the training sample size is small. This is attributed to the underlying assumption that conventional techniques employing the empirical risk minimization (ERM) inductive principle converge asymptotically to the true risk values as the number of samples increases. In small sample size estimation problems, the available training samples must span the complexity of the parameter space so that the model is able both to match the available training samples reasonably well and to generalize to new data. This is achieved using the structural risk minimization (SRM) inductive principle by matching the capability of the model to the available training data. One method that uses SRM is support vector regression (SVR) network. In this research, the capability of SVR to predict porosity and permeability in a heterogeneous sandstone reservoir under the effect of small sample size is evaluated. Particularly, the impact of Vapnik's ɛ-insensitivity loss function and least-modulus loss function on generalization performance was empirically investigated. The results are compared to the multilayer perception (MLP) neural network, a widely used regression method, which operates under the ERM principle. The mean square error and correlation coefficients were used to measure the quality of predictions. The results demonstrate that SVR yields consistently better predictions of the porosity and permeability with small sample size than the MLP method. Also, the performance of SVR depends on both kernel function type and loss functions used.
Liu, Fang
2016-01-01
In both clinical development and post-marketing of a new therapy or a new treatment, incidence of an adverse event (AE) is always a concern. When sample sizes are small, large sample-based inferential approaches on an AE incidence proportion in a certain time period no longer apply. In this brief discussion, we introduce a simple Bayesian framework to quantify, in small sample studies and the rare AE case, (1) the confidence level that the incidence proportion of a particular AE p is over or below a threshold, (2) the lower or upper bounds on p with a certain level of confidence, and (3) the minimum required number of patients with an AE before we can be certain that p surpasses a specific threshold, or the maximum allowable number of patients with an AE after which we can no longer be certain that p is below a certain threshold, given a certain confidence level. The method is easy to understand and implement; the interpretation of the results is intuitive. This article also demonstrates the usefulness of simple Bayesian concepts when it comes to answering practical questions.
Sampling for Global Epidemic Models and the Topology of an International Airport Network
Bobashev, Georgiy; Morris, Robert J.; Goedecke, D. Michael
2008-01-01
Mathematical models that describe the global spread of infectious diseases such as influenza, severe acute respiratory syndrome (SARS), and tuberculosis (TB) often consider a sample of international airports as a network supporting disease spread. However, there is no consensus on how many cities should be selected or on how to select those cities. Using airport flight data that commercial airlines reported to the Official Airline Guide (OAG) in 2000, we have examined the network characteristics of network samples obtained under different selection rules. In addition, we have examined different size samples based on largest flight volume and largest metropolitan populations. We have shown that although the bias in network characteristics increases with the reduction of the sample size, a relatively small number of areas that includes the largest airports, the largest cities, the most-connected cities, and the most central cities is enough to describe the dynamics of the global spread of influenza. The analysis suggests that a relatively small number of cities (around 200 or 300 out of almost 3000) can capture enough network information to adequately describe the global spread of a disease such as influenza. Weak traffic flows between small airports can contribute to noise and mask other means of spread such as the ground transportation. PMID:18776932
Monitoring Species of Concern Using Noninvasive Genetic Sampling and Capture-Recapture Methods
2016-11-01
ABBREVIATIONS AICc Akaike’s Information Criterion with small sample size correction AZGFD Arizona Game and Fish Department BMGR Barry M. Goldwater...MNKA Minimum Number Known Alive N Abundance Ne Effective Population Size NGS Noninvasive Genetic Sampling NGS-CR Noninvasive Genetic...parameter estimates from capture-recapture models require sufficient sample sizes , capture probabilities and low capture biases. For NGS-CR, sample
NASA Astrophysics Data System (ADS)
She, Qi; Chen, Guanrong; Chan, Rosa H. M.
2016-02-01
The amount of publicly accessible experimental data has gradually increased in recent years, which makes it possible to reconsider many longstanding questions in neuroscience. In this paper, an efficient framework is presented for reconstructing functional connectivity using experimental spike-train data. A modified generalized linear model (GLM) with L1-norm penalty was used to investigate 10 datasets. These datasets contain spike-train data collected from the entorhinal-hippocampal region in the brains of rats performing different tasks. The analysis shows that entorhinal-hippocampal network of well-trained rats demonstrated significant small-world features. It is found that the connectivity structure generated by distance-dependent models is responsible for the observed small-world features of the reconstructed networks. The models are utilized to simulate a subset of units recorded from a large biological neural network using multiple electrodes. Two metrics for quantifying the small-world-ness both suggest that the reconstructed network from the sampled nodes estimates a more prominent small-world-ness feature than that of the original unknown network when the number of recorded neurons is small. Finally, this study shows that it is feasible to adjust the estimated small-world-ness results based on the number of neurons recorded to provide a more accurate reference of the network property.
Evaluation of the Biological Sampling Kit (BiSKit) for Large-Area Surface Sampling
Buttner, Mark P.; Cruz, Patricia; Stetzenbach, Linda D.; Klima-Comba, Amy K.; Stevens, Vanessa L.; Emanuel, Peter A.
2004-01-01
Current surface sampling methods for microbial contaminants are designed to sample small areas and utilize culture analysis. The total number of microbes recovered is low because a small area is sampled, making detection of a potential pathogen more difficult. Furthermore, sampling of small areas requires a greater number of samples to be collected, which delays the reporting of results, taxes laboratory resources and staffing, and increases analysis costs. A new biological surface sampling method, the Biological Sampling Kit (BiSKit), designed to sample large areas and to be compatible with testing with a variety of technologies, including PCR and immunoassay, was evaluated and compared to other surface sampling strategies. In experimental room trials, wood laminate and metal surfaces were contaminated by aerosolization of Bacillus atrophaeus spores, a simulant for Bacillus anthracis, into the room, followed by settling of the spores onto the test surfaces. The surfaces were sampled with the BiSKit, a cotton-based swab, and a foam-based swab. Samples were analyzed by culturing, quantitative PCR, and immunological assays. The results showed that the large surface area (1 m2) sampled with the BiSKit resulted in concentrations of B. atrophaeus in samples that were up to 10-fold higher than the concentrations obtained with the other methods tested. A comparison of wet and dry sampling with the BiSKit indicated that dry sampling was more efficient (efficiency, 18.4%) than wet sampling (efficiency, 11.3%). The sensitivities of detection of B. atrophaeus on metal surfaces were 42 ± 5.8 CFU/m2 for wet sampling and 100.5 ± 10.2 CFU/m2 for dry sampling. These results demonstrate that the use of a sampling device capable of sampling larger areas results in higher sensitivity than that obtained with currently available methods and has the advantage of sampling larger areas, thus requiring collection of fewer samples per site. PMID:15574898
Costa, Sofia R; Kerry, Brian R; Bardgett, Richard D; Davies, Keith G
2006-12-01
The Pasteuria group of endospore-forming bacteria has been studied as a biocontrol agent of plant-parasitic nematodes. Techniques have been developed for its detection and quantification in soil samples, and these mainly focus on observations of endospore attachment to nematodes. Characterization of Pasteuria populations has recently been performed with DNA-based techniques, which usually require the extraction of large numbers of spores. We describe a simple immunological method for the quantification and characterization of Pasteuria populations. Bayesian statistics were used to determine an extraction efficiency of 43% and a threshold of detection of 210 endospores g(-1) sand. This provided a robust means of estimating numbers of endospores in small-volume samples from a natural system. Based on visual assessment of endospore fluorescence, a quantitative method was developed to characterize endospore populations, which were shown to vary according to their host.
Use of Unlabeled Samples for Mitigating the Hughes Phenomenon
NASA Technical Reports Server (NTRS)
Landgrebe, David A.; Shahshahani, Behzad M.
1993-01-01
The use of unlabeled samples in improving the performance of classifiers is studied. When the number of training samples is fixed and small, additional feature measurements may reduce the performance of a statistical classifier. It is shown that by using unlabeled samples, estimates of the parameters can be improved and therefore this phenomenon may be mitigated. Various methods for using unlabeled samples are reviewed and experimental results are provided.
Detecting Small Amounts of Gene Flow from Phylogenies of Alleles
Slatkin, M.
1989-01-01
The method of coalescents is used to find the probability that none of the ancestors of alleles sampled from a population are immigrants. If that is the case for samples from two or more populations, then there would be concordance between the phylogenies of those alleles and the geographic locations from which they are drawn. This type of concordance has been found in several studies of mitochondrial DNA from natural populations. It is shown that if the number of sequences sampled from each population is reasonably large (10 or more), then this type of concordance suggests that the average number of individuals migrating between populations is likely to be relatively small (Nm < 1) but the possibility of occasional migrants cannot be excluded. The method is applied to the data of E. Bermingham and J. C. Avise on mtDNA from the bowfin, Amia calva. PMID:2714639
Differential sampling for fast frequency acquisition via adaptive extended least squares algorithm
NASA Technical Reports Server (NTRS)
Kumar, Rajendra
1987-01-01
This paper presents a differential signal model along with appropriate sampling techinques for least squares estimation of the frequency and frequency derivatives and possibly the phase and amplitude of a sinusoid received in the presence of noise. The proposed algorithm is recursive in mesurements and thus the computational requirement increases only linearly with the number of measurements. The dimension of the state vector in the proposed algorithm does not depend upon the number of measurements and is quite small, typically around four. This is an advantage when compared to previous algorithms wherein the dimension of the state vector increases monotonically with the product of the frequency uncertainty and the observation period. Such a computational simplification may possibly result in some loss of optimality. However, by applying the sampling techniques of the paper such a possible loss in optimality can made small.
USDA-ARS?s Scientific Manuscript database
Small, coded, pill-sized tracers embedded in grain are proposed as a method for grain traceability. A sampling process for a grain traceability system was designed and investigated by applying probability statistics using a science-based sampling approach to collect an adequate number of tracers fo...
Incorporating Biological Knowledge into Evaluation of Casual Regulatory Hypothesis
NASA Technical Reports Server (NTRS)
Chrisman, Lonnie; Langley, Pat; Bay, Stephen; Pohorille, Andrew; DeVincenzi, D. (Technical Monitor)
2002-01-01
Biological data can be scarce and costly to obtain. The small number of samples available typically limits statistical power and makes reliable inference of causal relations extremely difficult. However, we argue that statistical power can be increased substantially by incorporating prior knowledge and data from diverse sources. We present a Bayesian framework that combines information from different sources and we show empirically that this lets one make correct causal inferences with small sample sizes that otherwise would be impossible.
Small-airway obstruction and emphysema in chronic obstructive pulmonary disease.
McDonough, John E; Yuan, Ren; Suzuki, Masaru; Seyednejad, Nazgol; Elliott, W Mark; Sanchez, Pablo G; Wright, Alexander C; Gefter, Warren B; Litzky, Leslie; Coxson, Harvey O; Paré, Peter D; Sin, Don D; Pierce, Richard A; Woods, Jason C; McWilliams, Annette M; Mayo, John R; Lam, Stephen C; Cooper, Joel D; Hogg, James C
2011-10-27
The major sites of obstruction in chronic obstructive pulmonary disease (COPD) are small airways (<2 mm in diameter). We wanted to determine whether there was a relationship between small-airway obstruction and emphysematous destruction in COPD. We used multidetector computed tomography (CT) to compare the number of airways measuring 2.0 to 2.5 mm in 78 patients who had various stages of COPD, as judged by scoring on the Global Initiative for Chronic Obstructive Lung Disease (GOLD) scale, in isolated lungs removed from patients with COPD who underwent lung transplantation, and in donor (control) lungs. MicroCT was used to measure the extent of emphysema (mean linear intercept), the number of terminal bronchioles per milliliter of lung volume, and the minimum diameters and cross-sectional areas of terminal bronchioles. On multidetector CT, in samples from patients with COPD, as compared with control samples, the number of airways measuring 2.0 to 2.5 mm in diameter was reduced in patients with GOLD stage 1 disease (P=0.001), GOLD stage 2 disease (P=0.02), and GOLD stage 3 or 4 disease (P<0.001). MicroCT of isolated samples of lungs removed from patients with GOLD stage 4 disease showed a reduction of 81 to 99.7% in the total cross-sectional area of terminal bronchioles and a reduction of 72 to 89% in the number of terminal bronchioles (P<0.001). A comparison of the number of terminal bronchioles and dimensions at different levels of emphysematous destruction (i.e., an increasing value for the mean linear intercept) showed that the narrowing and loss of terminal bronchioles preceded emphysematous destruction in COPD (P<0.001). These results show that narrowing and disappearance of small conducting airways before the onset of emphysematous destruction can explain the increased peripheral airway resistance reported in COPD. (Funded by the National Heart, Lung, and Blood Institute and others.).
An empirical study using permutation-based resampling in meta-regression
2012-01-01
Background In meta-regression, as the number of trials in the analyses decreases, the risk of false positives or false negatives increases. This is partly due to the assumption of normality that may not hold in small samples. Creation of a distribution from the observed trials using permutation methods to calculate P values may allow for less spurious findings. Permutation has not been empirically tested in meta-regression. The objective of this study was to perform an empirical investigation to explore the differences in results for meta-analyses on a small number of trials using standard large sample approaches verses permutation-based methods for meta-regression. Methods We isolated a sample of randomized controlled clinical trials (RCTs) for interventions that have a small number of trials (herbal medicine trials). Trials were then grouped by herbal species and condition and assessed for methodological quality using the Jadad scale, and data were extracted for each outcome. Finally, we performed meta-analyses on the primary outcome of each group of trials and meta-regression for methodological quality subgroups within each meta-analysis. We used large sample methods and permutation methods in our meta-regression modeling. We then compared final models and final P values between methods. Results We collected 110 trials across 5 intervention/outcome pairings and 5 to 10 trials per covariate. When applying large sample methods and permutation-based methods in our backwards stepwise regression the covariates in the final models were identical in all cases. The P values for the covariates in the final model were larger in 78% (7/9) of the cases for permutation and identical for 22% (2/9) of the cases. Conclusions We present empirical evidence that permutation-based resampling may not change final models when using backwards stepwise regression, but may increase P values in meta-regression of multiple covariates for relatively small amount of trials. PMID:22587815
Carolyn Hull Sieg
1988-01-01
Small mammals and vegetation were sampled over two years in Rocky Mountain juniper woodlands and adjacent grasslands in South Dakota. Juniper woodlands provided specialized habitat for two woodland species, white-footed mice and bushy-tailed woodrats, and attracted a number of species generally associated with grasslands.
What is the extent of prokaryotic diversity?
Curtis, Thomas P; Head, Ian M; Lunn, Mary; Woodcock, Stephen; Schloss, Patrick D; Sloan, William T
2006-01-01
The extent of microbial diversity is an intrinsically fascinating subject of profound practical importance. The term ‘diversity’ may allude to the number of taxa or species richness as well as their relative abundance. There is uncertainty about both, primarily because sample sizes are too small. Non-parametric diversity estimators make gross underestimates if used with small sample sizes on unevenly distributed communities. One can make richness estimates over many scales using small samples by assuming a species/taxa-abundance distribution. However, no one knows what the underlying taxa-abundance distributions are for bacterial communities. Latterly, diversity has been estimated by fitting data from gene clone libraries and extrapolating from this to taxa-abundance curves to estimate richness. However, since sample sizes are small, we cannot be sure that such samples are representative of the community from which they were drawn. It is however possible to formulate, and calibrate, models that predict the diversity of local communities and of samples drawn from that local community. The calibration of such models suggests that migration rates are small and decrease as the community gets larger. The preliminary predictions of the model are qualitatively consistent with the patterns seen in clone libraries in ‘real life’. The validation of this model is also confounded by small sample sizes. However, if such models were properly validated, they could form invaluable tools for the prediction of microbial diversity and a basis for the systematic exploration of microbial diversity on the planet. PMID:17028084
2010-06-01
Sampling (MIS)? • Technique of combining many increments of soil from a number of points within exposure area • Developed by Enviro Stat (Trademarked...Demonstrating a reliable soil sampling strategy to accurately characterize contaminant concentrations in spatially extreme and heterogeneous...into a set of decision (exposure) units • One or several discrete or small- scale composite soil samples collected to represent each decision unit
ERIC Educational Resources Information Center
Gersten, Russell; Rolfhus, Eric; Clarke, Ben; Decker, Lauren E.; Wilkins, Chuck; Dimino, Joseph
2015-01-01
Replication studies are extremely rare in education. This randomized controlled trial (RCT) is a scale-up replication of Fuchs et al., which in a sample of 139 found a statistically significant positive impact for Number Rockets, a small-group intervention for at-risk first graders that focused on building understanding of number operations. The…
NASA Astrophysics Data System (ADS)
Bozorgzadeh, Nezam; Yanagimura, Yoko; Harrison, John P.
2017-12-01
The Hoek-Brown empirical strength criterion for intact rock is widely used as the basis for estimating the strength of rock masses. Estimations of the intact rock H-B parameters, namely the empirical constant m and the uniaxial compressive strength σc, are commonly obtained by fitting the criterion to triaxial strength data sets of small sample size. This paper investigates how such small sample sizes affect the uncertainty associated with the H-B parameter estimations. We use Monte Carlo (MC) simulation to generate data sets of different sizes and different combinations of H-B parameters, and then investigate the uncertainty in H-B parameters estimated from these limited data sets. We show that the uncertainties depend not only on the level of variability but also on the particular combination of parameters being investigated. As particular combinations of H-B parameters can informally be considered to represent specific rock types, we discuss that as the minimum number of required samples depends on rock type it should correspond to some acceptable level of uncertainty in the estimations. Also, a comparison of the results from our analysis with actual rock strength data shows that the probability of obtaining reliable strength parameter estimations using small samples may be very low. We further discuss the impact of this on ongoing implementation of reliability-based design protocols and conclude with suggestions for improvements in this respect.
GUM Analysis for TIMS Isotopic Ratios in BEP0 Graphite Qualification Samples, Round 2
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gerlach, David C.; Heasler, Patrick G.; Reid, Bruce D.
In May 2007, one set of three samples from NBL were addressed to Steve Petersen for TIMS analysis, and included BEP0 samples numbered 27008, 30986, and 50846. All cores were trimmed by tooling, and lightly cleaned by CO2 pellet blasting. Small discs were cut from the second set of samples for SIMS analysis, with the remainder of each used for TIMS preparation.
Forcino, Frank L; Leighton, Lindsey R; Twerdy, Pamela; Cahill, James F
2015-01-01
Community ecologists commonly perform multivariate techniques (e.g., ordination, cluster analysis) to assess patterns and gradients of taxonomic variation. A critical requirement for a meaningful statistical analysis is accurate information on the taxa found within an ecological sample. However, oversampling (too many individuals counted per sample) also comes at a cost, particularly for ecological systems in which identification and quantification is substantially more resource consuming than the field expedition itself. In such systems, an increasingly larger sample size will eventually result in diminishing returns in improving any pattern or gradient revealed by the data, but will also lead to continually increasing costs. Here, we examine 396 datasets: 44 previously published and 352 created datasets. Using meta-analytic and simulation-based approaches, the research within the present paper seeks (1) to determine minimal sample sizes required to produce robust multivariate statistical results when conducting abundance-based, community ecology research. Furthermore, we seek (2) to determine the dataset parameters (i.e., evenness, number of taxa, number of samples) that require larger sample sizes, regardless of resource availability. We found that in the 44 previously published and the 220 created datasets with randomly chosen abundances, a conservative estimate of a sample size of 58 produced the same multivariate results as all larger sample sizes. However, this minimal number varies as a function of evenness, where increased evenness resulted in increased minimal sample sizes. Sample sizes as small as 58 individuals are sufficient for a broad range of multivariate abundance-based research. In cases when resource availability is the limiting factor for conducting a project (e.g., small university, time to conduct the research project), statistically viable results can still be obtained with less of an investment.
Characterization Methods for Small Estuarine Systems in the Mid-Atlantic Region of the United States
Various statistical methods were applied to spatially discrete data from 14 intensively sampled small estuarine systems in the mid-Atlantic U.S. The number of sites per system ranged from 6 to 37. The surface area of the systems ranged from 1.9 to 193.4 km2. Parameters examined ...
A reliable method of analyzing dietaries of mycophagous small mammals
W. Colgan; A.B. Carey; James M. Trappe
1997-01-01
Two methods of analyzing the dietaries of mycophagous small mammals were compared. Fecal pellets were collected from 11 northern flying squirrels and 12 Townsend's chipmunks, all caught live. In 1 method, pellets from each individual were examined microscopically; in the other, samples from 3 or 4 individuals from each species were pooled and the number of slides...
Small-Sample Equating with Prior Information. Research Report. ETS RR-09-25
ERIC Educational Resources Information Center
Livingston, Samuel A.; Lewis, Charles
2009-01-01
This report proposes an empirical Bayes approach to the problem of equating scores on test forms taken by very small numbers of test takers. The equated score is estimated separately at each score point, making it unnecessary to model either the score distribution or the equating transformation. Prior information comes from equatings of other…
Collateral Information for Equating in Small Samples: A Preliminary Investigation
ERIC Educational Resources Information Center
Kim, Sooyeon; Livingston, Samuel A.; Lewis, Charles
2011-01-01
This article describes a preliminary investigation of an empirical Bayes (EB) procedure for using collateral information to improve equating of scores on test forms taken by small numbers of examinees. Resampling studies were done on two different forms of the same test. In each study, EB and non-EB versions of two equating methods--chained linear…
Modified Technique For Chemisorption Measurements
NASA Technical Reports Server (NTRS)
Schryer, David R.; Brown, Kenneth G.; Schryer, Jacqueline
1989-01-01
In measurements of chemisorption of CO on Pt/SnO2 catalyst observed that if small numbers of relatively large volumes of adsorbate gas are passed through sample, very little removal of CO detected. In these cases little or no CO has been chemisorbed on Pt/SnO2. Technique of using large number of small volumes of adsorbate gas to measure chemisorption applicable to many gas/material combinations other than CO on Pt/SnO2. Volume used chosen so that at least 10 percent of adsorbate gas removed during each exposure.
Estimation of reference intervals from small samples: an example using canine plasma creatinine.
Geffré, A; Braun, J P; Trumel, C; Concordet, D
2009-12-01
According to international recommendations, reference intervals should be determined from at least 120 reference individuals, which often are impossible to achieve in veterinary clinical pathology, especially for wild animals. When only a small number of reference subjects is available, the possible bias cannot be known and the normality of the distribution cannot be evaluated. A comparison of reference intervals estimated by different methods could be helpful. The purpose of this study was to compare reference limits determined from a large set of canine plasma creatinine reference values, and large subsets of this data, with estimates obtained from small samples selected randomly. Twenty sets each of 120 and 27 samples were randomly selected from a set of 1439 plasma creatinine results obtained from healthy dogs in another study. Reference intervals for the whole sample and for the large samples were determined by a nonparametric method. The estimated reference limits for the small samples were minimum and maximum, mean +/- 2 SD of native and Box-Cox-transformed values, 2.5th and 97.5th percentiles by a robust method on native and Box-Cox-transformed values, and estimates from diagrams of cumulative distribution functions. The whole sample had a heavily skewed distribution, which approached Gaussian after Box-Cox transformation. The reference limits estimated from small samples were highly variable. The closest estimates to the 1439-result reference interval for 27-result subsamples were obtained by both parametric and robust methods after Box-Cox transformation but were grossly erroneous in some cases. For small samples, it is recommended that all values be reported graphically in a dot plot or histogram and that estimates of the reference limits be compared using different methods.
Papini, Paolo; Faustini, Annunziata; Manganello, Rosa; Borzacchi, Giancarlo; Spera, Domenico; Perucci, Carlo A
2005-01-01
To determine the frequency of sampling in small water distribution systems (<5,000 inhabitants) and compare the results according to different hypotheses in bacteria distribution. We carried out two sampling programs to monitor the water distribution system in a town in Central Italy between July and September 1992; the Poisson distribution assumption implied 4 water samples, the assumption of negative binomial distribution implied 21 samples. Coliform organisms were used as indicators of water safety. The network consisted of two pipe rings and two wells fed by the same water source. The number of summer customers varied considerably from 3,000 to 20,000. The mean density was 2.33 coliforms/100 ml (sd= 5.29) for 21 samples and 3 coliforms/100 ml (sd= 6) for four samples. However the hypothesis of homogeneity was rejected (p-value <0.001) and the probability of II type error with the assumption of heterogeneity was higher with 4 samples (beta= 0.24) than with 21 (beta= 0.05). For this small network, determining the samples' size according to heterogeneity hypothesis strengthens the statement that water is drinkable compared with homogeneity assumption.
Improved argument-FFT frequency offset estimation for QPSK coherent optical Systems
NASA Astrophysics Data System (ADS)
Han, Jilong; Li, Wei; Yuan, Zhilin; Li, Haitao; Huang, Liyan; Hu, Qianggao
2016-02-01
A frequency offset estimation (FOE) algorithm based on fast Fourier transform (FFT) of the signal's argument is investigated, which does not require removing the modulated data phase. In this paper, we analyze the flaw of the argument-FFT algorithm and propose a combined FOE algorithm, in which the absolute of frequency offset (FO) is accurately calculated by argument-FFT algorithm with a relatively large number of samples and the sign of FO is determined by FFT-based interpolation discrete Fourier transformation (DFT) algorithm with a relatively small number of samples. Compared with the previous algorithms based on argument-FFT, the proposed one has low complexity and can still effectively work with a relatively less number of samples.
NASA Astrophysics Data System (ADS)
Lee, H.; Sheen, D.; Kim, S.
2013-12-01
The b-value in Gutenberg-Richter relation is an important parameter widely used not only in the interpretation of regional tectonic structure but in the seismic hazard analysis. In this study, we tested four methods for estimating the stable b-value in a small number of events using Monte-Carlo method. One is the Least-Squares method (LSM) which minimizes the observation error. Others are based on the Maximum Likelihood method (MLM) which maximizes the likelihood function: Utsu's (1965) method for continuous magnitudes and an infinite maximum magnitude, Page's (1968) for continuous magnitudes and a finite maximum magnitude, and Weichert's (1980) for interval magnitude and a finite maximum magnitude. A synthetic parent population of the earthquake catalog of million events from magnitude 2.0 to 7.0 with interval of 0.1 was generated for the Monte-Carlo simulation. The sample, the number of which was increased from 25 to 1000, was extracted from the parent population randomly. The resampling procedure was applied 1000 times with different random seed numbers. The mean and the standard deviation of the b-value were estimated for each sample group that has the same number of samples. As expected, the more samples were used, the more stable b-value was obtained. However, in a small number of events, the LSM gave generally low b-value with a large standard deviation while other MLMs gave more accurate and stable values. It was found that Utsu (1965) gives the most accurate and stable b-value even in a small number of events. It was also found that the selection of the minimum magnitude could be critical for estimating the correct b-value for Utsu's (1965) method and Page's (1968) if magnitudes were binned into an interval. Therefore, we applied Utsu (1965) to estimate the b-value using two instrumental earthquake catalogs, which have events occurred around the southern part of the Korean Peninsula from 1978 to 2011. By a careful choice of the minimum magnitude, the b-values of the earthquake catalogs of the Korea Meteorological Administration and Kim (2012) are estimated to be 0.72 and 0.74, respectively.
Directions for new developments on statistical design and analysis of small population group trials.
Hilgers, Ralf-Dieter; Roes, Kit; Stallard, Nigel
2016-06-14
Most statistical design and analysis methods for clinical trials have been developed and evaluated where at least several hundreds of patients could be recruited. These methods may not be suitable to evaluate therapies if the sample size is unavoidably small, which is usually termed by small populations. The specific sample size cut off, where the standard methods fail, needs to be investigated. In this paper, the authors present their view on new developments for design and analysis of clinical trials in small population groups, where conventional statistical methods may be inappropriate, e.g., because of lack of power or poor adherence to asymptotic approximations due to sample size restrictions. Following the EMA/CHMP guideline on clinical trials in small populations, we consider directions for new developments in the area of statistical methodology for design and analysis of small population clinical trials. We relate the findings to the research activities of three projects, Asterix, IDeAl, and InSPiRe, which have received funding since 2013 within the FP7-HEALTH-2013-INNOVATION-1 framework of the EU. As not all aspects of the wide research area of small population clinical trials can be addressed, we focus on areas where we feel advances are needed and feasible. The general framework of the EMA/CHMP guideline on small population clinical trials stimulates a number of research areas. These serve as the basis for the three projects, Asterix, IDeAl, and InSPiRe, which use various approaches to develop new statistical methodology for design and analysis of small population clinical trials. Small population clinical trials refer to trials with a limited number of patients. Small populations may result form rare diseases or specific subtypes of more common diseases. New statistical methodology needs to be tailored to these specific situations. The main results from the three projects will constitute a useful toolbox for improved design and analysis of small population clinical trials. They address various challenges presented by the EMA/CHMP guideline as well as recent discussions about extrapolation. There is a need for involvement of the patients' perspective in the planning and conduct of small population clinical trials for a successful therapy evaluation.
On the analysis of very small samples of Gaussian repeated measurements: an alternative approach.
Westgate, Philip M; Burchett, Woodrow W
2017-03-15
The analysis of very small samples of Gaussian repeated measurements can be challenging. First, due to a very small number of independent subjects contributing outcomes over time, statistical power can be quite small. Second, nuisance covariance parameters must be appropriately accounted for in the analysis in order to maintain the nominal test size. However, available statistical strategies that ensure valid statistical inference may lack power, whereas more powerful methods may have the potential for inflated test sizes. Therefore, we explore an alternative approach to the analysis of very small samples of Gaussian repeated measurements, with the goal of maintaining valid inference while also improving statistical power relative to other valid methods. This approach uses generalized estimating equations with a bias-corrected empirical covariance matrix that accounts for all small-sample aspects of nuisance correlation parameter estimation in order to maintain valid inference. Furthermore, the approach utilizes correlation selection strategies with the goal of choosing the working structure that will result in the greatest power. In our study, we show that when accurate modeling of the nuisance correlation structure impacts the efficiency of regression parameter estimation, this method can improve power relative to existing methods that yield valid inference. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.
DOE-2 sample run book: Version 2.1E
DOE Office of Scientific and Technical Information (OSTI.GOV)
Winkelmann, F.C.; Birdsall, B.E.; Buhl, W.F.
1993-11-01
The DOE-2 Sample Run Book shows inputs and outputs for a variety of building and system types. The samples start with a simple structure and continue to a high-rise office building, a medical building, three small office buildings, a bar/lounge, a single-family residence, a small office building with daylighting, a single family residence with an attached sunspace, a ``parameterized`` building using input macros, and a metric input/output example. All of the samples use Chicago TRY weather. The main purpose of the Sample Run Book is instructional. It shows the relationship of LOADS-SYSTEMS-PLANT-ECONOMICS inputs, displays various input styles, and illustrates manymore » of the basic and advanced features of the program. Many of the sample runs are preceded by a sketch of the building showing its general appearance and the zoning used in the input. In some cases we also show a 3-D rendering of the building as produced by the program DrawBDL. Descriptive material has been added as comments in the input itself. We find that a number of users have loaded these samples onto their editing systems and use them as ``templates`` for creating new inputs. Another way of using them would be to store various portions as files that can be read into the input using the {number_sign}{number_sign} include command, which is part of the Input Macro feature introduced in version DOE-2.lD. Note that the energy rate structures here are the same as in the DOE-2.lD samples, but have been rewritten using the new DOE-2.lE commands and keywords for ECONOMICS. The samples contained in this report are the same as those found on the DOE-2 release files. However, the output numbers that appear here may differ slightly from those obtained from the release files. The output on the release files can be used as a check set to compare results on your computer.« less
Haverkamp, Nicolas; Beauducel, André
2017-01-01
We investigated the effects of violations of the sphericity assumption on Type I error rates for different methodical approaches of repeated measures analysis using a simulation approach. In contrast to previous simulation studies on this topic, up to nine measurement occasions were considered. Effects of the level of inter-correlations between measurement occasions on Type I error rates were considered for the first time. Two populations with non-violation of the sphericity assumption, one with uncorrelated measurement occasions and one with moderately correlated measurement occasions, were generated. One population with violation of the sphericity assumption combines uncorrelated with highly correlated measurement occasions. A second population with violation of the sphericity assumption combines moderately correlated and highly correlated measurement occasions. From these four populations without any between-group effect or within-subject effect 5,000 random samples were drawn. Finally, the mean Type I error rates for Multilevel linear models (MLM) with an unstructured covariance matrix (MLM-UN), MLM with compound-symmetry (MLM-CS) and for repeated measures analysis of variance (rANOVA) models (without correction, with Greenhouse-Geisser-correction, and Huynh-Feldt-correction) were computed. To examine the effect of both the sample size and the number of measurement occasions, sample sizes of n = 20, 40, 60, 80, and 100 were considered as well as measurement occasions of m = 3, 6, and 9. With respect to rANOVA, the results plead for a use of rANOVA with Huynh-Feldt-correction, especially when the sphericity assumption is violated, the sample size is rather small and the number of measurement occasions is large. For MLM-UN, the results illustrate a massive progressive bias for small sample sizes ( n = 20) and m = 6 or more measurement occasions. This effect could not be found in previous simulation studies with a smaller number of measurement occasions. The proportionality of bias and number of measurement occasions should be considered when MLM-UN is used. The good news is that this proportionality can be compensated by means of large sample sizes. Accordingly, MLM-UN can be recommended even for small sample sizes for about three measurement occasions and for large sample sizes for about nine measurement occasions.
Leung, Ana Marie R.; Lu, Jinky Leilanie DP.
2016-01-01
OBJECTIVES This cross-sectional study aimed at the environmental health hazards at work and cyanide exposure of small-scale gold miners engaged in gold extraction from ores in a mining area in the Philippines. METHODS Methods consisted of structured questionnaire-guided interviews, work process observation tools, physical health assessment by medical doctors, and laboratory examination and blood cyanide determination in the blood samples of 34 indigenous small-scale gold miners from Benguet, Philippines. RESULTS The small-scale gold miners worked for a mean of 10.3 years, had a mean age of 36 years, with mean lifetime mining work hours of 18,564. All were involved in tunneling work (100%) while a considerable number were involved in mixing cyanide with the ore (44%). A considerable number were injured (35%) during the mining activity, and an alarming number (35%) had elevated blood cyanide level. The most prevalent hazard was exposure to chemicals, particularly to cyanide and nitric acid, which were usually handled with bare hands. CONCLUSION The small-scale gold miners were exposed to occupational and environmental hazards at work. PMID:27547035
Leung, Ana Marie R; Lu, Jinky Leilanie Dp
2016-01-01
This cross-sectional study aimed at the environmental health hazards at work and cyanide exposure of small-scale gold miners engaged in gold extraction from ores in a mining area in the Philippines. Methods consisted of structured questionnaire-guided interviews, work process observation tools, physical health assessment by medical doctors, and laboratory examination and blood cyanide determination in the blood samples of 34 indigenous small-scale gold miners from Benguet, Philippines. The small-scale gold miners worked for a mean of 10.3 years, had a mean age of 36 years, with mean lifetime mining work hours of 18,564. All were involved in tunneling work (100%) while a considerable number were involved in mixing cyanide with the ore (44%). A considerable number were injured (35%) during the mining activity, and an alarming number (35%) had elevated blood cyanide level. The most prevalent hazard was exposure to chemicals, particularly to cyanide and nitric acid, which were usually handled with bare hands. The small-scale gold miners were exposed to occupational and environmental hazards at work.
Loeb, Gregory M.
2018-01-01
Invasive, polyphagous crop pests subsist on a number of crop and non-crop resources. While knowing the full range of host species is important, a seasonal investigation into the use of non-crop plants adjacent to cropping systems provide key insights into some of the factors determining local population dynamics. This study investigated the infestation of non-crop plants by the invasive Drosophila suzukii (Matsumura), a pest of numerous economically important stone and small fruit crops, by sampling fruit-producing non-crop hosts adjacent to commercial plantings weekly from June through November in central New York over a two-year period. We found D. suzukii infestation rates (number of flies emerged/kg fruit) peaked mid-August through early September, with Rubus allegheniensis Porter and Lonicera morrowii Asa Gray showing the highest average infestation in both years. Interannual infestation patterns were similar despite a lower number of adults caught in monitoring traps the second year, suggesting D. suzukii host use may be density independent. PMID:29301358
On the theory and simulation of multiple Coulomb scattering of heavy-charged particles.
Striganov, S I
2005-01-01
The Moliere theory of multiple Coulomb scattering is modified to take into account the difference between processes of scattering off atomic nuclei and electrons. A simple analytical expression for angular distribution of charged particles passing through a thick absorber is found. It does not assume any special form for a differential scattering cross section and has a wider range of applicability than a gaussian approximation. A well-known method to simulate multiple Coulomb scatterings is based on treating 'soft' and 'hard' collisions differently. An angular deflection in a large number of 'soft' collisions is sampled using the proposed distribution function, a small number of 'hard' collision are simulated directly. A boundary between 'hard' and 'soft' collisions is defined, providing a precise sampling of a scattering angle (1% level) and a small number of 'hard' collisions. A corresponding simulating module takes into account projectile and nucleus charged distributions and exact kinematics of a projectile-electron interaction.
Dry particle generation with a 3-D printed fluidized bed generator
Roesch, Michael; Roesch, Carolin; Cziczo, Daniel J.
2017-06-02
We describe the design and testing of PRIZE (PRinted fluidIZed bed gEnerator), a compact fluidized bed aerosol generator manufactured using stereolithography (SLA) printing. Dispersing small quantities of powdered materials – due to either rarity or expense – is challenging due to a lack of small, low-cost dry aerosol generators. With this as motivation, we designed and built a generator that uses a mineral dust or other dry powder sample mixed with bronze beads that sit atop a porous screen. A particle-free airflow is introduced, dispersing the sample as airborne particles. The total particle number concentrations and size distributions were measured duringmore » different stages of the assembling process to show that the SLA 3-D printed generator did not generate particles until the mineral dust sample was introduced. Furthermore, time-series measurements with Arizona Test Dust (ATD) showed stable total particle number concentrations of 10–150 cm -3, depending on the sample mass, from the sub- to super-micrometer size range. Additional tests with collected soil dust samples are also presented. PRIZE is simple to assemble, easy to clean, inexpensive and deployable for laboratory and field studies that require dry particle generation.« less
Dry particle generation with a 3-D printed fluidized bed generator
DOE Office of Scientific and Technical Information (OSTI.GOV)
Roesch, Michael; Roesch, Carolin; Cziczo, Daniel J.
We describe the design and testing of PRIZE (PRinted fluidIZed bed gEnerator), a compact fluidized bed aerosol generator manufactured using stereolithography (SLA) printing. Dispersing small quantities of powdered materials – due to either rarity or expense – is challenging due to a lack of small, low-cost dry aerosol generators. With this as motivation, we designed and built a generator that uses a mineral dust or other dry powder sample mixed with bronze beads that sit atop a porous screen. A particle-free airflow is introduced, dispersing the sample as airborne particles. The total particle number concentrations and size distributions were measured duringmore » different stages of the assembling process to show that the SLA 3-D printed generator did not generate particles until the mineral dust sample was introduced. Furthermore, time-series measurements with Arizona Test Dust (ATD) showed stable total particle number concentrations of 10–150 cm -3, depending on the sample mass, from the sub- to super-micrometer size range. Additional tests with collected soil dust samples are also presented. PRIZE is simple to assemble, easy to clean, inexpensive and deployable for laboratory and field studies that require dry particle generation.« less
Monte Carlo simulation of induction time and metastable zone width; stochastic or deterministic?
NASA Astrophysics Data System (ADS)
Kubota, Noriaki
2018-03-01
The induction time and metastable zone width (MSZW) measured for small samples (say 1 mL or less) both scatter widely. Thus, these two are observed as stochastic quantities. Whereas, for large samples (say 1000 mL or more), the induction time and MSZW are observed as deterministic quantities. The reason for such experimental differences is investigated with Monte Carlo simulation. In the simulation, the time (under isothermal condition) and supercooling (under polythermal condition) at which a first single crystal is detected are defined as the induction time t and the MSZW ΔT for small samples, respectively. The number of crystals just at the moment of t and ΔT is unity. A first crystal emerges at random due to the intrinsic nature of nucleation, accordingly t and ΔT become stochastic. For large samples, the time and supercooling at which the number density of crystals N/V reaches a detector sensitivity (N/V)det are defined as t and ΔT for isothermal and polythermal conditions, respectively. The points of t and ΔT are those of which a large number of crystals have accumulated. Consequently, t and ΔT become deterministic according to the law of large numbers. Whether t and ΔT may stochastic or deterministic in actual experiments should not be attributed to change in nucleation mechanisms in molecular level. It could be just a problem caused by differences in the experimental definition of t and ΔT.
Li, Peng; Redden, David T.
2014-01-01
SUMMARY The sandwich estimator in generalized estimating equations (GEE) approach underestimates the true variance in small samples and consequently results in inflated type I error rates in hypothesis testing. This fact limits the application of the GEE in cluster-randomized trials (CRTs) with few clusters. Under various CRT scenarios with correlated binary outcomes, we evaluate the small sample properties of the GEE Wald tests using bias-corrected sandwich estimators. Our results suggest that the GEE Wald z test should be avoided in the analyses of CRTs with few clusters even when bias-corrected sandwich estimators are used. With t-distribution approximation, the Kauermann and Carroll (KC)-correction can keep the test size to nominal levels even when the number of clusters is as low as 10, and is robust to the moderate variation of the cluster sizes. However, in cases with large variations in cluster sizes, the Fay and Graubard (FG)-correction should be used instead. Furthermore, we derive a formula to calculate the power and minimum total number of clusters one needs using the t test and KC-correction for the CRTs with binary outcomes. The power levels as predicted by the proposed formula agree well with the empirical powers from the simulations. The proposed methods are illustrated using real CRT data. We conclude that with appropriate control of type I error rates under small sample sizes, we recommend the use of GEE approach in CRTs with binary outcomes due to fewer assumptions and robustness to the misspecification of the covariance structure. PMID:25345738
William L. Thompson
2003-01-01
Hankin and Reeves' (1988) approach to estimating fish abundance in small streams has been applied in stream fish studies across North America. However, their population estimator relies on two key assumptions: (1) removal estimates are equal to the true numbers of fish, and (2) removal estimates are highly correlated with snorkel counts within a subset of sampled...
Mars Immunoassay Life Detection Instrument (MILDI)
NASA Technical Reports Server (NTRS)
McKay, David; Steele, Andrew; Allen, Carlton; Thomas-Kepta, Kathie; Schweitzer, Mary; Priscu, John; Sears, Joe; Avci, Recep; Firman, Keith
2000-01-01
The direct detection of organic biomarkers for living or fossil microbes on Mars by an in situ instrument is a worthy goal for future lander missions. We have proposed an instrument based on immunological reactions to specific antibodies to cause activation of fluorescent stains. Antibodies are raised or acquired to a variety of general and specific substances that might be in Mars soil. These antibodies are then combined with various fluorescent stains and applied to small numbered spots on a small (two to three centimeters) test plate where they become firmly attached after drying. On Mars, a sample of soil from a trench or drill core is extracted with water and/or an organic solvent that is then applied to the test plate. Any substance, which has an antibody on the test plate, will react with its antibody and activate its fluorescent stain. A small ultraviolet light source will illuminate the test plate, which is observed with a small CCD camera. The numbered spots that fluoresce indicate the presence of the tested-for substance, and the intensity indicates relative amounts. The entire instrument can be quite small and light, on the order of ten cm in each dimension. A possible choice for light source may be small UV lasers at several wavelengths. Up to 1000 different sample spots can be placed on a plate 3 cm on a side, but a more practical number might be 100. Each antibody can have a redundant position for independent verification of reaction. Some of the wells or spots can contain simply standard fluorescent stains used to detect live cells, dead cells, DNA, etc. These the stains in these spots may be directly activated; no antibodies are necessary.
Determination of small quantities of fluoride in water: A modified zirconium-alizarin method
Lamar, W.L.; Seegmiller, C.G.
1941-01-01
The zirconium-alizarin method has been modified to facilitate the convenient and accurate determination of small amounts of fluoride in a large number of water samples. Sulfuric acid is used to acidify the samples to reduce the interference of sulfate. The pH is accurately controlled to give the most sensitive comparisons. Most natural waters can be analyzed by the modified procedure without resorting to correction curves. The fluoride content of waters containing less than 500 parts per million of sulfate, 500 parts per million of bicarbonate, and 1000 parts per million of chloride may be determined within a limit of about 0.1 part per million when a 100-ml. sample is used.
Evaluation of process errors in bed load sampling using a Dune Model
Gomez, Basil; Troutman, Brent M.
1997-01-01
Reliable estimates of the streamwide bed load discharge obtained using sampling devices are dependent upon good at-a-point knowledge across the full width of the channel. Using field data and information derived from a model that describes the geometric features of a dune train in terms of a spatial process observed at a fixed point in time, we show that sampling errors decrease as the number of samples collected increases, and the number of traverses of the channel over which the samples are collected increases. It also is preferable that bed load sampling be conducted at a pace which allows a number of bed forms to pass through the sampling cross section. The situations we analyze and simulate pertain to moderate transport conditions in small rivers. In such circumstances, bed load sampling schemes typically should involve four or five traverses of a river, and the collection of 20–40 samples at a rate of five or six samples per hour. By ensuring that spatial and temporal variability in the transport process is accounted for, such a sampling design reduces both random and systematic errors and hence minimizes the total error involved in the sampling process.
STUDY OF HOME DEMONSTRATION UNITS IN A SAMPLE OF 27 COUNTIES IN NEW YORK STATE, NUMBER 3.
ERIC Educational Resources Information Center
ALEXANDER, FRANK D.; HARSHAW, JEAN
AN EXPLORATORY STUDY EXAMINED CHARACTERISTICS OF 1,128 HOME DEMONSTRATION UNITS TO SUGGEST HYPOTHESES AND SCOPE FOR A MORE INTENSIVE STUDY OF A SMALL SAMPLE OF UNITS, AND TO PROVIDE GUIDANCE IN SAMPLING. DATA WERE OBTAINED FROM A SPECIALLY DESIGNED MEMBERSHIP CARD USED IN 1962. UNIT SIZE AVERAGED 23.6 MEMBERS BUT THE RANGE WAS FAIRLY GREAT. A NEED…
Data survey and sampling procedures to quantify recreation use of national forests in Alaska
Ginny Fay; Steve Colt; Eric M. White
2010-01-01
Estimating visitor numbers and collecting information on visitor attitudes in Alaska national forests is especially challenging because of the dispersed access to the forests by a relatively small number of visitors. The Tongass and Chugach National Forests are each millions of acres with miles of saltwater coastline and numerous lakes...
NASA Astrophysics Data System (ADS)
Kawahara, Toshio; Tabuchi, Norikazu; Arai, Takashi; Sato, Yoshikazu; Morimoto, Jun; Matsumura, Hideki
2005-02-01
We measured structure factors of hydrogenated amorphous Si by x-ray diffraction and analysed the obtained structures using a reverse Monte Carlo (RMC) technique. A small shoulder in the measured structure factor S(Q) was observed on the larger Q side of the first peak. The RMC results with an unconstrained model did not clearly show the small shoulder. Adding constraints for coordination numbers 2 and 3, the small shoulder was reproduced and the agreement with the experimental data became better. The ratio of the constrained coordination numbers was consistent with the ratio of Si-H and Si-H2 bonds which was estimated by the Fourier transformed infrared spectra of the same sample. This shoulder and the oscillation of the corresponding pair distribution function g(r) at large r seem to be related to the low randomness of cat-CVD deposited a-Si:H.
Liu, Lian; Zhang, Shao-Wu; Huang, Yufei; Meng, Jia
2017-08-31
As a newly emerged research area, RNA epigenetics has drawn increasing attention recently for the participation of RNA methylation and other modifications in a number of crucial biological processes. Thanks to high throughput sequencing techniques, such as, MeRIP-Seq, transcriptome-wide RNA methylation profile is now available in the form of count-based data, with which it is often of interests to study the dynamics at epitranscriptomic layer. However, the sample size of RNA methylation experiment is usually very small due to its costs; and additionally, there usually exist a large number of genes whose methylation level cannot be accurately estimated due to their low expression level, making differential RNA methylation analysis a difficult task. We present QNB, a statistical approach for differential RNA methylation analysis with count-based small-sample sequencing data. Compared with previous approaches such as DRME model based on a statistical test covering the IP samples only with 2 negative binomial distributions, QNB is based on 4 independent negative binomial distributions with their variances and means linked by local regressions, and in the way, the input control samples are also properly taken care of. In addition, different from DRME approach, which relies only the input control sample only for estimating the background, QNB uses a more robust estimator for gene expression by combining information from both input and IP samples, which could largely improve the testing performance for very lowly expressed genes. QNB showed improved performance on both simulated and real MeRIP-Seq datasets when compared with competing algorithms. And the QNB model is also applicable to other datasets related RNA modifications, including but not limited to RNA bisulfite sequencing, m 1 A-Seq, Par-CLIP, RIP-Seq, etc.
Rodic, Maja; Zhou, Xinlin; Tikhomirova, Tatiana; Wei, Wei; Malykh, Sergei; Ismatulina, Victoria; Sabirova, Elena; Davidova, Yulia; Tosto, Maria Grazia; Lemelin, Jean-Pascal; Kovas, Yulia
2015-01-01
The present study evaluated 626 5-7-year-old children in the UK, China, Russia, and Kyrgyzstan on a cognitive test battery measuring: (1) general skills; (2) non-symbolic number sense; (3) symbolic number understanding; (4) simple arithmetic - operating with numbers; and (5) familiarity with numbers. Although most inter-population differences were small, 13% of the variance in arithmetic skills could be explained by the sample, replicating the pattern, previously found with older children in PISA. Furthermore, the same cognitive skills were related to early arithmetic in these diverse populations. Only understanding of symbolic number explained variation in mathematical performance in all samples. We discuss the results in terms of potential influences of socio-demographic, linguistic and genetic factors on individual differences in mathematics. © 2014 John Wiley & Sons Ltd.
Radioactivity Registered With a Small Number of Events
NASA Astrophysics Data System (ADS)
Zlokazov, Victor; Utyonkov, Vladimir
2018-02-01
The synthesis of superheavy elements asks for the analysis of low statistics experimental data presumably obeying an unknown exponential distribution and to take the decision whether they originate from one source or have admixtures. Here we analyze predictions following from non-parametrical methods, employing only such fundamental sample properties as the sample mean, the median and the mode.
On Understanding the Notion of Limits and Infinitesimal Quantities
ERIC Educational Resources Information Center
Parameswaran, Revathy
2007-01-01
In this paper we explore the influence of students' personalized notion of "small" numbers based on real life experiences on their understanding of limits. Tests were conducted on two samples of students. The first sample, consisting of students in the XII grade, had been taught limits using an informal approach (i.e., without recourse to the…
Can Financial Need Analysis be Simplified?
ERIC Educational Resources Information Center
Orwig, M. D.; Jones, Paul K.
This paper examines the problem of collecting financial data on aid applicants. A 10% sample (12,383) of student records was taken from the 1968-69 alphabetic history file for the ACT Student Need Analysis Service. Random sub-samples were taken in certain phases of the study. A relatively small number of financial variables were found to predict…
Kim, Minkyoung; Choi, Seung-Hoon; Kim, Junhyoung; Choi, Kihang; Shin, Jae-Min; Kang, Sang-Kee; Choi, Yun-Jaie; Jung, Dong Hyun
2009-11-01
This study describes the application of a density-based algorithm to clustering small peptide conformations after a molecular dynamics simulation. We propose a clustering method for small peptide conformations that enables adjacent clusters to be separated more clearly on the basis of neighbor density. Neighbor density means the number of neighboring conformations, so if a conformation has too few neighboring conformations, then it is considered as noise or an outlier and is excluded from the list of cluster members. With this approach, we can easily identify clusters in which the members are densely crowded in the conformational space, and we can safely avoid misclustering individual clusters linked by noise or outliers. Consideration of neighbor density significantly improves the efficiency of clustering of small peptide conformations sampled from molecular dynamics simulations and can be used for predicting peptide structures.
Variational Approach to Enhanced Sampling and Free Energy Calculations
NASA Astrophysics Data System (ADS)
Valsson, Omar; Parrinello, Michele
2014-08-01
The ability of widely used sampling methods, such as molecular dynamics or Monte Carlo simulations, to explore complex free energy landscapes is severely hampered by the presence of kinetic bottlenecks. A large number of solutions have been proposed to alleviate this problem. Many are based on the introduction of a bias potential which is a function of a small number of collective variables. However constructing such a bias is not simple. Here we introduce a functional of the bias potential and an associated variational principle. The bias that minimizes the functional relates in a simple way to the free energy surface. This variational principle can be turned into a practical, efficient, and flexible sampling method. A number of numerical examples are presented which include the determination of a three-dimensional free energy surface. We argue that, beside being numerically advantageous, our variational approach provides a convenient and novel standpoint for looking at the sampling problem.
Zangar, Richard C.; Varnum, Susan M.; Covington, Chandice Y.; ...
2004-01-01
Identifying useful markers of cancer can be problematic due to limited amounts of sample. Some samples such as nipple aspirate fluid (NAF) or early-stage tumors are inherently small. Other samples such as serum are collected in larger volumes but archives of these samples are very valuable and only small amounts of each sample may be available for a single study. Also, given the diverse nature of cancer and the inherent variability in individual protein levels, it seems likely that the best approach to screen for cancer will be to determine the profile of a battery of proteins. As a result,more » a major challenge in identifying protein markers of disease is the ability to screen many proteins using very small amounts of sample. In this review, we outline some technological advances in proteomics that greatly advance this capability. Specifically, we propose a strategy for identifying markers of breast cancer in NAF that utilizes mass spectrometry (MS) to simultaneously screen hundreds or thousands of proteins in each sample. The best potential markers identified by the MS analysis can then be extensively characterized using an ELISA microarray assay. Because the microarray analysis is quantitative and large numbers of samples can be efficiently analyzed, this approach offers the ability to rapidly assess a battery of selected proteins in a manner that is directly relevant to traditional clinical assays.« less
Shaffer, Patrick; Valsson, Omar; Parrinello, Michele
2016-01-01
The capabilities of molecular simulations have been greatly extended by a number of widely used enhanced sampling methods that facilitate escaping from metastable states and crossing large barriers. Despite these developments there are still many problems which remain out of reach for these methods which has led to a vigorous effort in this area. One of the most important problems that remains unsolved is sampling high-dimensional free-energy landscapes and systems that are not easily described by a small number of collective variables. In this work we demonstrate a new way to compute free-energy landscapes of high dimensionality based on the previously introduced variationally enhanced sampling, and we apply it to the miniprotein chignolin. PMID:26787868
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gould, Andrew; Yee, Jennifer C., E-mail: gould@astronomy.ohio-state.edu, E-mail: jyee@astronomy.ohio-state.edu
While of order of a million asteroids have been discovered, the number in rigorously controlled samples that have precise orbits and rotation periods, as well as well-measured colors, is relatively small. In particular, less than a dozen main-belt asteroids with estimated diameters D < 3 km have excellent rotation periods. We show how existing and soon-to-be-acquired microlensing data can yield a large asteroid sample with precise orbits and rotation periods, which will include roughly 6% of all asteroids with maximum brightness I < 18.1 and lying within 10 Degree-Sign of the ecliptic. This sample will be dominated by small andmore » very small asteroids, down to D {approx} 1 km. We also show how asteroid astrometry could turn current narrow-angle OGLE proper motions of bulge stars into wide-angle proper motions. This would enable one to measure the proper-motion gradient across the Galactic bar.« less
Rosenblum, Michael A; Laan, Mark J van der
2009-01-07
The validity of standard confidence intervals constructed in survey sampling is based on the central limit theorem. For small sample sizes, the central limit theorem may give a poor approximation, resulting in confidence intervals that are misleading. We discuss this issue and propose methods for constructing confidence intervals for the population mean tailored to small sample sizes. We present a simple approach for constructing confidence intervals for the population mean based on tail bounds for the sample mean that are correct for all sample sizes. Bernstein's inequality provides one such tail bound. The resulting confidence intervals have guaranteed coverage probability under much weaker assumptions than are required for standard methods. A drawback of this approach, as we show, is that these confidence intervals are often quite wide. In response to this, we present a method for constructing much narrower confidence intervals, which are better suited for practical applications, and that are still more robust than confidence intervals based on standard methods, when dealing with small sample sizes. We show how to extend our approaches to much more general estimation problems than estimating the sample mean. We describe how these methods can be used to obtain more reliable confidence intervals in survey sampling. As a concrete example, we construct confidence intervals using our methods for the number of violent deaths between March 2003 and July 2006 in Iraq, based on data from the study "Mortality after the 2003 invasion of Iraq: A cross sectional cluster sample survey," by Burnham et al. (2006).
High-efficiency multiphoton boson sampling
NASA Astrophysics Data System (ADS)
Wang, Hui; He, Yu; Li, Yu-Huai; Su, Zu-En; Li, Bo; Huang, He-Liang; Ding, Xing; Chen, Ming-Cheng; Liu, Chang; Qin, Jian; Li, Jin-Peng; He, Yu-Ming; Schneider, Christian; Kamp, Martin; Peng, Cheng-Zhi; Höfling, Sven; Lu, Chao-Yang; Pan, Jian-Wei
2017-06-01
Boson sampling is considered as a strong candidate to demonstrate 'quantum computational supremacy' over classical computers. However, previous proof-of-principle experiments suffered from small photon number and low sampling rates owing to the inefficiencies of the single-photon sources and multiport optical interferometers. Here, we develop two central components for high-performance boson sampling: robust multiphoton interferometers with 99% transmission rate and actively demultiplexed single-photon sources based on a quantum dot-micropillar with simultaneously high efficiency, purity and indistinguishability. We implement and validate three-, four- and five-photon boson sampling, and achieve sampling rates of 4.96 kHz, 151 Hz and 4 Hz, respectively, which are over 24,000 times faster than previous experiments. Our architecture can be scaled up for a larger number of photons and with higher sampling rates to compete with classical computers, and might provide experimental evidence against the extended Church-Turing thesis.
Cheng, Ningtao; Wu, Leihong; Cheng, Yiyu
2013-01-01
The promise of microarray technology in providing prediction classifiers for cancer outcome estimation has been confirmed by a number of demonstrable successes. However, the reliability of prediction results relies heavily on the accuracy of statistical parameters involved in classifiers. It cannot be reliably estimated with only a small number of training samples. Therefore, it is of vital importance to determine the minimum number of training samples and to ensure the clinical value of microarrays in cancer outcome prediction. We evaluated the impact of training sample size on model performance extensively based on 3 large-scale cancer microarray datasets provided by the second phase of MicroArray Quality Control project (MAQC-II). An SSNR-based (scale of signal-to-noise ratio) protocol was proposed in this study for minimum training sample size determination. External validation results based on another 3 cancer datasets confirmed that the SSNR-based approach could not only determine the minimum number of training samples efficiently, but also provide a valuable strategy for estimating the underlying performance of classifiers in advance. Once translated into clinical routine applications, the SSNR-based protocol would provide great convenience in microarray-based cancer outcome prediction in improving classifier reliability. PMID:23861920
1980-01-01
In a multi-laboratory trial, minerals-modified glutamate medium (MMGM) was compared with lauryl tryptose lactose broth (LTLB) in the multiple tube method for the enumeration of coliform organisms, including Escherichia coli, in water. Samples of raw and chlorinated waters yielded a total of 2313 positive tube-reactions with MMGM and 2174 with LTLB. These were interpreted either as E. coli; other coliform organisms; or as false positive reactions. The results at first reading (18 or 24 h) and at 48 h have been analysed statistically in terms of (i) most probable numbers of coliform organisms; (ii) positive reactions and their interpretation; and (iii) whether or not the sample yielded any E. coli or other coliform organisms. All three analyses indicated the same trends. For the detection of E. coli in raw waters LTLB was better than MMGM at 18-24 h, but MMGM was better at 48 h with waters containing small numbers of coliform organisms; for raw waters with greater numbers of organisms, both media performed equally well. Analysis of a subset of samples read at both 18 and 24 h indicated that the superiority of LTLB over MMGM with raw waters disappeared by 24 h. For chlorinated waters, LTLB yielded more positive gas reactions at 18-24 h, but fewer of these were E. coli than with MMGM; at 48 h MMGM was clearly better than LTLB for total coliform organisms including E. coli--especially if the numbers were small. MMGM therefore remains the medium of choice for the detection of E. coli as an indicator of faecal contamination of chlorinated drinking water supplies. It is also better for the detection of small numbers of E. coli in other waters. PMID:6762395
Reducing the number of reconstructions needed for estimating channelized observer performance
NASA Astrophysics Data System (ADS)
Pineda, Angel R.; Miedema, Hope; Brenner, Melissa; Altaf, Sana
2018-03-01
A challenge for task-based optimization is the time required for each reconstructed image in applications where reconstructions are time consuming. Our goal is to reduce the number of reconstructions needed to estimate the area under the receiver operating characteristic curve (AUC) of the infinitely-trained optimal channelized linear observer. We explore the use of classifiers which either do not invert the channel covariance matrix or do feature selection. We also study the assumption that multiple low contrast signals in the same image of a non-linear reconstruction do not significantly change the estimate of the AUC. We compared the AUC of several classifiers (Hotelling, logistic regression, logistic regression using Firth bias reduction and the least absolute shrinkage and selection operator (LASSO)) with a small number of observations both for normal simulated data and images from a total variation reconstruction in magnetic resonance imaging (MRI). We used 10 Laguerre-Gauss channels and the Mann-Whitney estimator for AUC. For this data, our results show that at small sample sizes feature selection using the LASSO technique can decrease bias of the AUC estimation with increased variance and that for large sample sizes the difference between these classifiers is small. We also compared the use of multiple signals in a single reconstructed image to reduce the number of reconstructions in a total variation reconstruction for accelerated imaging in MRI. We found that AUC estimation using multiple low contrast signals in the same image resulted in similar AUC estimates as doing a single reconstruction per signal leading to a 13x reduction in the number of reconstructions needed.
Dal Grande, Eleonora; Chittleborough, Catherine R; Campostrini, Stefano; Taylor, Anne W
2016-04-18
Emerging communication technologies have had an impact on population-based telephone surveys worldwide. Our objective was to examine the potential biases of health estimates in South Australia, a state of Australia, obtained via current landline telephone survey methodologies and to report on the impact of mobile-only household on household surveys. Data from an annual multi-stage, systematic, clustered area, face-to-face population survey, Health Omnibus Survey (approximately 3000 interviews annually), included questions about telephone ownership to assess the population that were non-contactable by current telephone sampling methods (2006 to 2013). Univariable analyses (2010 to 2013) and trend analyses were conducted for sociodemographic and health indicator variables in relation to telephone status. Relative coverage biases (RCB) of two hypothetical telephone samples was undertaken by examining the prevalence estimates of health status and health risk behaviours (2010 to 2013): directory-listed numbers, consisting mainly of landline telephone numbers and a small proportion of mobile telephone numbers; and a random digit dialling (RDD) sample of landline telephone numbers which excludes mobile-only households. Telephone (landline and mobile) coverage in South Australia is very high (97%). Mobile telephone ownership increased slightly (7.4%), rising from 89.7% in 2006 to 96.3% in 2013; mobile-only households increased by 431% over the eight year period from 5.2% in 2006 to 27.6% in 2013. Only half of the households have either a mobile or landline number listed in the telephone directory. There were small differences in the prevalence estimates for current asthma, arthritis, diabetes and obesity between the hypothetical telephone samples and the overall sample. However, prevalence estimate for diabetes was slightly underestimated (RCB value of -0.077) in 2013. Mixed RCB results were found for having a mental health condition for both telephone samples. Current smoking prevalence was lower for both hypothetical telephone samples in absolute differences and RCB values: -0.136 to -0.191 for RDD landline samples and -0.129 to -0.313 for directory-listed samples. These findings suggest landline-based sampling frames used in Australia, when appropriately weighted, produce reliable representative estimates for some health indicators but not for all. Researchers need to be aware of their limitations and potential biased estimates.
An Analysis Of Coast Guard Enlisted Retention
1993-03-01
Instrument Identification Number(i f applicable , Address (cirv. state, and ZIP code) 10 Source of Funding Numbers Program Element No Project No ITask...46 E. SAMPLE RESTRICTIONS ........ .............. 48 F. DATA LIMITATIONS AND PROBLEMS ... ......... .. 52 1. PMIS Data Base...civilian employment suggest retention behavior may be similar. Also, the small personnel inventories of some of the rates would limit the model’s
ERIC Educational Resources Information Center
Hasselhorn, Marcus; Linke-Hasselhorn, Kathrin
2013-01-01
Eight six-year old German children with development disabilities regarding such number competencies as have been demonstrated to be among the most relevant precursor skills for the acquisition of elementary mathematics received intensive training with the program "Mengen, zählen, Zahlen" ["quantities, counting, numbers"] (MZZ,…
USDA-ARS?s Scientific Manuscript database
For any analytical system the population mean (mu) number of entities (e.g., cells or molecules) per tested volume, surface area, or mass also defines the population standard deviation (sigma = square root of mu ). For a preponderance of analytical methods, sigma is very small relative to mu due to...
Radiation sensitivity of foodborne pathogens in meat byproducts with different packaging
NASA Astrophysics Data System (ADS)
Yong, Hae In; Kim, Hyun-Joo; Nam, Ki Chang; Kwon, Joong Ho; Jo, Cheorun
2015-10-01
The aim of this study was to determine radiation sensitivity of Escherichia coli O157:H7 and Listeria monocytogenes in edible meat byproducts. Seven beef byproducts (heart, liver, lung, lumen, omasum, large intestine, and small intestine) and four pork byproducts (heart, large intestine, liver, and small intestine) were used. Electron beam irradiation significantly reduced the numbers of pathogenic microorganisms in meat byproducts and no viable cells were detected in both aerobically- and vacuum-packaged samples irradiated at 4 kGy. Meat byproducts packed under vacuum had higher D10 value than the ones packed aerobically. No significant difference was observed between the D10 values of E. coli O157:H7 and L. monocytogenes inoculated in either aerobically or vacuum packaged samples. These results suggest that low-dose electron beam irradiation can significantly decrease microbial numbers and reduce the risk of meat byproduct contamination by the foodborne pathogens.
Zhao, Qi; Liu, Yuanning; Zhang, Ning; Hu, Menghan; Zhang, Hao; Joshi, Trupti; Xu, Dong
2018-01-01
In recent years, an increasing number of studies have reported the presence of plant miRNAs in human samples, which resulted in a hypothesis asserting the existence of plant-derived exogenous microRNA (xenomiR). However, this hypothesis is not widely accepted in the scientific community due to possible sample contamination and the small sample size with lack of rigorous statistical analysis. This study provides a systematic statistical test that can validate (or invalidate) the plant-derived xenomiR hypothesis by analyzing 388 small RNA sequencing data from human samples in 11 types of body fluids/tissues. A total of 166 types of plant miRNAs were found in at least one human sample, of which 14 plant miRNAs represented more than 80% of the total plant miRNAs abundance in human samples. Plant miRNA profiles were characterized to be tissue-specific in different human samples. Meanwhile, the plant miRNAs identified from microbiome have an insignificant abundance compared to those from humans, while plant miRNA profiles in human samples were significantly different from those in plants, suggesting that sample contamination is an unlikely reason for all the plant miRNAs detected in human samples. This study also provides a set of testable synthetic miRNAs with isotopes that can be detected in situ after being fed to animals.
ERIC Educational Resources Information Center
Fiedler, Klaus; Kareev, Yaakov
2006-01-01
Adaptive decision making requires that contingencies between decision options and their relative assets be assessed accurately and quickly. The present research addresses the challenging notion that contingencies may be more visible from small than from large samples of observations. An algorithmic account for such a seemingly paradoxical effect…
ERIC Educational Resources Information Center
Roisman, Glenn I.; Fraley, R. Chris
2008-01-01
A number of relatively small-sample, genetically sensitive studies of infant attachment security have been published in the past several years that challenge the view that all psychological phenotypes are heritable and that environmental influences on child development--to the extent that they can be detected--serve to make siblings dissimilar.…
D. Lee Taylor; Michael G. Booth; Jack W. McFarland; Ian C. Herriott; Niall J. Lennon; Chad Nusbaum; Thomas G. Marr
2008-01-01
High throughput sequencing methods are widely used in analyses of microbial diversity but are generally applied to small numbers of samples, which precludes charaterization of patterns of microbial diversity across space and time. We have designed a primer-tagging approach that allows pooling and subsequent sorting of numerous samples, which is directed to...
TNO/Centaurs grouping tested with asteroid data sets
NASA Astrophysics Data System (ADS)
Fulchignoni, M.; Birlan, M.; Barucci, M. A.
2001-11-01
Recently, we have discussed the possible subdivision in few groups of a sample of 22 TNO and Centaurs for which the BVRIJ photometry were available (Barucci et al., 2001, A&A, 371,1150). We obtained this results using the multivariate statistics adopted to define the current asteroid taxonomy, namely the Principal Components Analysis and the G-mode method (Tholen & Barucci, 1989, in ASTEROIDS II). How these methods work with a very small statistical sample as the TNO/Centaurs one? Theoretically, the number of degrees of freedom of the sample is correct. In fact it is 88 in our case and have to be larger then 50 to cope with the requirements of the G-mode. Does the random sampling of the small number of members of a large population contain enough information to reveal some structure in the population? We extracted several samples of 22 asteroids out of a data-base of 86 objects of known taxonomic type for which BVRIJ photometry is available from ECAS (Zellner et al. 1985, ICARUS 61, 355), SMASS II (S.W. Bus, 1999, PhD Thesis, MIT), and the Bell et al. Atlas of the asteroid infrared spectra. The objects constituting the first sample were selected in order to give a good representation of the major asteroid taxonomic classes (at least three samples each class): C,S,D,A, and G. Both methods were able to distinguish all these groups confirming the validity of the adopted methods. The S class is hard to individuate as a consequence of the choice of I and J variables, which imply a lack of information on the absorption band at 1 micron. The other samples were obtained by random choice of the objects. Not all the major groups were well represented (less than three samples per groups), but the general trend of the asteroid taxonomy has been always obtained. We conclude that the quoted grouping of TNO/Centaurs is representative of some physico-chemical structure of the outer solar system small body population.
[Microbiological quality of the air in "small gastronomy point"].
Wójcik-Stopczyńska, Barbara
2006-01-01
The aim of this work was the estimation of microbial contamination of the air in "small gastronomy point". The study included three places, which have been separated on the ground of their function: 1. area of subsidiaries, 2. area of distribution (sale and serving meal), 3. area of consumption. The total numbers of aerobic mesophilic bacteria, yeasts and moulds were determined by sedimentation method. Taxonomy units of fungal aerosol were also estimated. The samples of air were collected in 16 investigation points in the morning (8-8.30) and in the afternoon (14-14.30). Four series of measurements were carried out and in general 128 of air samples were tested. The results showed that numbers of bacteria, yeasts and moulds were variable and received respectively 30-3397, 0-254 and 0-138 cfu x m(-3). Microbial contamination of air changed depending on area character (the highest average count of bacteria occurred in the air of consumption area and fungi in subsidiaries area), time of a day (contamination of the air increased in the afternoon) and determination date. Only in single samples the numbers of bacteria and fungi were higher than recommended level. Pigmentary bacteria had high participation in total count of bacteria and filamentous fungi were represented mostly by Penicillium sp. and Cladosporium sp.
Modeling the genealogy of a cultural trait.
Aguilar, Elliot; Ghirlanda, Stefano
2015-05-01
The mathematical study of genealogies has yielded important insights in population biology, such as the ability to estimate the time to the most recent common ancestor (MRCA) of a sample of genetic sequences or of a group of individuals. Here we introduce a model of cultural genealogies that is a step toward answering similar questions for cultural traits. In our model individuals can inherit from a variable, potentially large number of ancestors, rather than from a fixed, small number of ancestors (one or two) as is typical of genetic evolution. We first show that, given a sample of individuals, a cultural common ancestor does not necessarily exist. We then introduce a related concept: the most recent unique ancestor (MRUA), i.e., the most recent single individual who is the earliest cultural ancestor of the sample. We show that, under neutral evolution, the time to the MRUA can be staggeringly larger than the time to MRCA in a single ancestor model, except when the average number of learning opportunities per individuals is small. Our results point out that the properties of cultural genealogies may be very different from those of genetic genealogies, with potential implications for reconstructing the histories of cultural traits. Copyright © 2014 Elsevier Inc. All rights reserved.
Jamali, Jamshid; Ayatollahi, Seyyed Mohammad Taghi; Jafari, Peyman
2017-01-01
Evaluating measurement equivalence (also known as differential item functioning (DIF)) is an important part of the process of validating psychometric questionnaires. This study aimed at evaluating the multiple indicators multiple causes (MIMIC) model for DIF detection when latent construct distribution is nonnormal and the focal group sample size is small. In this simulation-based study, Type I error rates and power of MIMIC model for detecting uniform-DIF were investigated under different combinations of reference to focal group sample size ratio, magnitude of the uniform-DIF effect, scale length, the number of response categories, and latent trait distribution. Moderate and high skewness in the latent trait distribution led to a decrease of 0.33% and 0.47% power of MIMIC model for detecting uniform-DIF, respectively. The findings indicated that, by increasing the scale length, the number of response categories and magnitude DIF improved the power of MIMIC model, by 3.47%, 4.83%, and 20.35%, respectively; it also decreased Type I error of MIMIC approach by 2.81%, 5.66%, and 0.04%, respectively. This study revealed that power of MIMIC model was at an acceptable level when latent trait distributions were skewed. However, empirical Type I error rate was slightly greater than nominal significance level. Consequently, the MIMIC was recommended for detection of uniform-DIF when latent construct distribution is nonnormal and the focal group sample size is small.
Jafari, Peyman
2017-01-01
Evaluating measurement equivalence (also known as differential item functioning (DIF)) is an important part of the process of validating psychometric questionnaires. This study aimed at evaluating the multiple indicators multiple causes (MIMIC) model for DIF detection when latent construct distribution is nonnormal and the focal group sample size is small. In this simulation-based study, Type I error rates and power of MIMIC model for detecting uniform-DIF were investigated under different combinations of reference to focal group sample size ratio, magnitude of the uniform-DIF effect, scale length, the number of response categories, and latent trait distribution. Moderate and high skewness in the latent trait distribution led to a decrease of 0.33% and 0.47% power of MIMIC model for detecting uniform-DIF, respectively. The findings indicated that, by increasing the scale length, the number of response categories and magnitude DIF improved the power of MIMIC model, by 3.47%, 4.83%, and 20.35%, respectively; it also decreased Type I error of MIMIC approach by 2.81%, 5.66%, and 0.04%, respectively. This study revealed that power of MIMIC model was at an acceptable level when latent trait distributions were skewed. However, empirical Type I error rate was slightly greater than nominal significance level. Consequently, the MIMIC was recommended for detection of uniform-DIF when latent construct distribution is nonnormal and the focal group sample size is small. PMID:28713828
Effect of Thermodiffusion Nitriding on Cytocompatibility of Ti-6Al-4V Titanium Alloy
NASA Astrophysics Data System (ADS)
Pohrelyuk, I. M.; Tkachuk, O. V.; Proskurnyak, R. V.; Boiko, N. M.; Kluchivska, O. Yu.; Stoika, R. S.
2016-04-01
The nitrided layer was formed on the surface of Ti-6Al-4V titanium alloy by the thermodiffusion saturation in nitrogen at the atmospheric pressure. The study of the vitality of pseudonormal human embryo kidney cells of the HEK293T line showed that their cultivation in the presence of the untreated alloy sample is accompanied by a statistically significant reduction in the number of living cells compared with the control sample (untreated cells), whereas their cultivation in the presence of the nitrided alloy sample does not change the cell number considerably. In addition, it was shown that cell behavior in the presence of the nitrided sample differs only slightly from the control sample, whereas the growth of cells in the presence of the untreated alloy differed significantly from that in the control sample, demonstrating small groups of cells instead of their big clusters.
Feng, Dai; Cortese, Giuliana; Baumgartner, Richard
2017-12-01
The receiver operating characteristic (ROC) curve is frequently used as a measure of accuracy of continuous markers in diagnostic tests. The area under the ROC curve (AUC) is arguably the most widely used summary index for the ROC curve. Although the small sample size scenario is common in medical tests, a comprehensive study of small sample size properties of various methods for the construction of the confidence/credible interval (CI) for the AUC has been by and large missing in the literature. In this paper, we describe and compare 29 non-parametric and parametric methods for the construction of the CI for the AUC when the number of available observations is small. The methods considered include not only those that have been widely adopted, but also those that have been less frequently mentioned or, to our knowledge, never applied to the AUC context. To compare different methods, we carried out a simulation study with data generated from binormal models with equal and unequal variances and from exponential models with various parameters and with equal and unequal small sample sizes. We found that the larger the true AUC value and the smaller the sample size, the larger the discrepancy among the results of different approaches. When the model is correctly specified, the parametric approaches tend to outperform the non-parametric ones. Moreover, in the non-parametric domain, we found that a method based on the Mann-Whitney statistic is in general superior to the others. We further elucidate potential issues and provide possible solutions to along with general guidance on the CI construction for the AUC when the sample size is small. Finally, we illustrate the utility of different methods through real life examples.
Effect of Sampling Plans on the Risk of Escherichia coli O157 Illness.
Kiermeier, Andreas; Sumner, John; Jenson, Ian
2015-07-01
Australia exports about 150,000 to 200,000 tons of manufacturing beef to the United States annually. Each lot is tested for Escherichia coli O157 using the N-60 sampling protocol, where 60 small pieces of surface meat from each lot of production are tested. A risk assessment of E. coli O157 illness from the consumption of hamburgers made from Australian manufacturing meat formed the basis to evaluate the effect of sample size and amount on the number of illnesses predicted. The sampling plans evaluated included no sampling (resulting in an estimated 55.2 illnesses per annum), the current N-60 plan (50.2 illnesses), N-90 (49.6 illnesses), N-120 (48.4 illnesses), and a more stringent N-60 sampling plan taking five 25-g samples from each of 12 cartons (47.4 illnesses per annum). While sampling may detect some highly contaminated lots, it does not guarantee that all such lots are removed from commerce. It is concluded that increasing the sample size or sample amount from the current N-60 plan would have a very small public health effect.
Methods for estimating the amount of vernal pool habitat in the northeastern United States
Van Meter, R.; Bailey, L.L.; Grant, E.H.C.
2008-01-01
The loss of small, seasonal wetlands is a major concern for a variety of state, local, and federal organizations in the northeastern U.S. Identifying and estimating the number of vernal pools within a given region is critical to developing long-term conservation and management strategies for these unique habitats and their faunal communities. We use three probabilistic sampling methods (simple random sampling, adaptive cluster sampling, and the dual frame method) to estimate the number of vernal pools on protected, forested lands. Overall, these methods yielded similar values of vernal pool abundance for each study area, and suggest that photographic interpretation alone may grossly underestimate the number of vernal pools in forested habitats. We compare the relative efficiency of each method and discuss ways of improving precision. Acknowledging that the objectives of a study or monitoring program ultimately determine which sampling designs are most appropriate, we recommend that some type of probabilistic sampling method be applied. We view the dual-frame method as an especially useful way of combining incomplete remote sensing methods, such as aerial photograph interpretation, with a probabilistic sample of the entire area of interest to provide more robust estimates of the number of vernal pools and a more representative sample of existing vernal pool habitats.
Variability of intertidal foraminferal assemblages in a salt marsh, Oregon, USA
Milker, Yvonne; Horton, Benjamin P.; Nelson, Alan R.; Engelhart, Simon E.; Witter, Robert C.
2015-01-01
We studied 18 sampling stations along a transect to investigate the similarity between live (rose Bengal stained) foraminiferal populations and dead assemblages, their small-scale spatial variations and the distribution of infaunal foraminifera in a salt marsh (Toms Creek marsh) at the upper end of the South Slough arm of the Coos Bay estuary, Oregon, USA. We aimed to test to what extent taphonomic processes, small-scale variability and infaunal distribution influence the accuracy of sea-level reconstructions based on intertidal foraminifera. Cluster analyses have shown that dead assemblages occur in distinct zones with respect to elevation, a prerequisite for using foraminifera as sea-level indicators. Our nonparametric multivariate analysis of variance showed that small-scale spatial variability has only a small influence on live (rose Bengal stained) populations and dead assemblages. The dissimilarity was higher, however, between live (rose Bengal stained) populations in the middle marsh. We observed early diagenetic dissolution of calcareous tests in the dead assemblages. If comparable post-depositional processes and similar minor spatial variability also characterize fossil assemblages, then dead assemblage are the best modern analogs for paleoenvironmental reconstructions. The Toms Creek tidal flat and low marsh vascular plant zones are dominated by Miliammina fusca, the middle marsh is dominated by Balticammina pseudomacrescens and Trochammina inflata, and the high marsh and upland–marsh transition zone are dominated by Trochamminita irregularis. Analysis of infaunal foraminifera showed that most living specimens are found in the surface sediments and the majority of live (rose Bengal stained) infaunal specimens are restricted to the upper 10 cm, but living individuals are found to depths of 50 cm. The dominant infaunal specimens are similar to those in the corresponding surface samples and no species have been found living solely infaunally. The total numbers of infaunal foraminifera are small compared to the total numbers of dead specimens in the surface samples. This suggests that surface samples adequately represent the modern intertidal environment in Toms Creek.
Stability and bias of classification rates in biological applications of discriminant analysis
Williams, B.K.; Titus, K.; Hines, J.E.
1990-01-01
We assessed the sampling stability of classification rates in discriminant analysis by using a factorial design with factors for multivariate dimensionality, dispersion structure, configuration of group means, and sample size. A total of 32,400 discriminant analyses were conducted, based on data from simulated populations with appropriate underlying statistical distributions. Simulation results indicated strong bias in correct classification rates when group sample sizes were small and when overlap among groups was high. We also found that stability of the correct classification rates was influenced by these factors, indicating that the number of samples required for a given level of precision increases with the amount of overlap among groups. In a review of 60 published studies, we found that 57% of the articles presented results on classification rates, though few of them mentioned potential biases in their results. Wildlife researchers should choose the total number of samples per group to be at least 2 times the number of variables to be measured when overlap among groups is low. Substantially more samples are required as the overlap among groups increases
Jiang, Wenyu; Simon, Richard
2007-12-20
This paper first provides a critical review on some existing methods for estimating the prediction error in classifying microarray data where the number of genes greatly exceeds the number of specimens. Special attention is given to the bootstrap-related methods. When the sample size n is small, we find that all the reviewed methods suffer from either substantial bias or variability. We introduce a repeated leave-one-out bootstrap (RLOOB) method that predicts for each specimen in the sample using bootstrap learning sets of size ln. We then propose an adjusted bootstrap (ABS) method that fits a learning curve to the RLOOB estimates calculated with different bootstrap learning set sizes. The ABS method is robust across the situations we investigate and provides a slightly conservative estimate for the prediction error. Even with small samples, it does not suffer from large upward bias as the leave-one-out bootstrap and the 0.632+ bootstrap, and it does not suffer from large variability as the leave-one-out cross-validation in microarray applications. Copyright (c) 2007 John Wiley & Sons, Ltd.
A Sub-Sampling Approach for Data Acquisition in Gamma Ray Emission Tomography
NASA Astrophysics Data System (ADS)
Fysikopoulos, Eleftherios; Kopsinis, Yannis; Georgiou, Maria; Loudos, George
2016-06-01
State of the art data acquisition systems for small animal imaging gamma ray detectors often rely on free running Analog to Digital Converters (ADCs) and high density Field Programmable Gate Arrays (FPGA) devices for digital signal processing. In this work, a sub-sampling acquisition approach, which exploits a priori information regarding the shape of the obtained detector pulses is proposed. Output pulses shape depends on the response of the scintillation crystal, photodetector's properties and amplifier/shaper operation. Using these known characteristics of the detector pulses prior to digitization, one can model the voltage pulse derived from the shaper (a low-pass filter, last in the front-end electronics chain), in order to reduce the desirable sampling rate of ADCs. Fitting with a small number of measurements, pulse shape estimation is then feasible. In particular, the proposed sub-sampling acquisition approach relies on a bi-exponential modeling of the pulse shape. We show that the properties of the pulse that are relevant for Single Photon Emission Computed Tomography (SPECT) event detection (i.e., position and energy) can be calculated by collecting just a small fraction of the number of samples usually collected in data acquisition systems used so far. Compared to the standard digitization process, the proposed sub-sampling approach allows the use of free running ADCs with sampling rate reduced by a factor of 5. Two small detectors consisting of Cerium doped Gadolinium Aluminum Gallium Garnet (Gd3Al2Ga3O12 : Ce or GAGG:Ce) pixelated arrays (array elements: 2 × 2 × 5 mm3 and 1 × 1 × 10 mm3 respectively) coupled to a Position Sensitive Photomultiplier Tube (PSPMT) were used for experimental evaluation. The two detectors were used to obtain raw images and energy histograms under 140 keV and 661.7 keV irradiation respectively. The sub-sampling acquisition technique (10 MHz sampling rate) was compared with a standard acquisition method (52 MHz sampling rate), in terms of energy resolution and image signal to noise ratio for both gamma ray energies. The Levenberg-Marquardt (LM) non-linear least-squares algorithm was used, in post processing, in order to fit the acquired data with the proposed model. The results showed that analog pulses prior to digitization are being estimated with high accuracy after fitting with the bi-exponential model.
NASA Astrophysics Data System (ADS)
Deng, Chengbin; Wu, Changshan
2013-12-01
Urban impervious surface information is essential for urban and environmental applications at the regional/national scales. As a popular image processing technique, spectral mixture analysis (SMA) has rarely been applied to coarse-resolution imagery due to the difficulty of deriving endmember spectra using traditional endmember selection methods, particularly within heterogeneous urban environments. To address this problem, we derived endmember signatures through a least squares solution (LSS) technique with known abundances of sample pixels, and integrated these endmember signatures into SMA for mapping large-scale impervious surface fraction. In addition, with the same sample set, we carried out objective comparative analyses among SMA (i.e. fully constrained and unconstrained SMA) and machine learning (i.e. Cubist regression tree and Random Forests) techniques. Analysis of results suggests three major conclusions. First, with the extrapolated endmember spectra from stratified random training samples, the SMA approaches performed relatively well, as indicated by small MAE values. Second, Random Forests yields more reliable results than Cubist regression tree, and its accuracy is improved with increased sample sizes. Finally, comparative analyses suggest a tentative guide for selecting an optimal approach for large-scale fractional imperviousness estimation: unconstrained SMA might be a favorable option with a small number of samples, while Random Forests might be preferred if a large number of samples are available.
Bellier, Edwige; Grøtan, Vidar; Engen, Steinar; Schartau, Ann Kristin; Diserud, Ola H; Finstad, Anders G
2012-10-01
Obtaining accurate estimates of diversity indices is difficult because the number of species encountered in a sample increases with sampling intensity. We introduce a novel method that requires that the presence of species in a sample to be assessed while the counts of the number of individuals per species are only required for just a small part of the sample. To account for species included as incidence data in the species abundance distribution, we modify the likelihood function of the classical Poisson log-normal distribution. Using simulated community assemblages, we contrast diversity estimates based on a community sample, a subsample randomly extracted from the community sample, and a mixture sample where incidence data are added to a subsample. We show that the mixture sampling approach provides more accurate estimates than the subsample and at little extra cost. Diversity indices estimated from a freshwater zooplankton community sampled using the mixture approach show the same pattern of results as the simulation study. Our method efficiently increases the accuracy of diversity estimates and comprehension of the left tail of the species abundance distribution. We show how to choose the scale of sample size needed for a compromise between information gained, accuracy of the estimates and cost expended when assessing biological diversity. The sample size estimates are obtained from key community characteristics, such as the expected number of species in the community, the expected number of individuals in a sample and the evenness of the community.
Multiplex-Ready Technology for mid-throughput genotyping of molecular markers.
Bonneau, Julien; Hayden, Matthew
2014-01-01
Screening molecular markers across large populations in breeding programs is generally time consuming and expensive. The Multiplex-Ready Technology (MRT) (Hayden et al., BMC genomics 9:80, 2008) was created to optimize polymorphism screening and genotyping using standardized PCR reaction conditions. The flexibility of this method maximizes the number of markers (up to 24 markers SSR or SNP, ideally small PCR product <500 bp and highly polymorphic) by using fluorescent dye (VIC, FAM, NED, and PET) and a semiautomated DNA fragment analyzer (ABI3730) capillary electrophoresis for large numbers of DNA samples (96 or 384 samples).
Virdis, Salvatore; Scarano, Christian; Spanu, Vincenzo; Murittu, Gavino; Spanu, Carlo; Ibba, Ignazio; De Santis, Enrico Pietro Luigi
2014-12-09
In the present work the results of a survey conducted in Sardinia Region on Aflatoxin M 1 (AFM 1 ) contamination in milk of small ruminants from 2005 to 2013 are reported. A total of 517 sheep and 88 goat milk samples from bulk tank, tank trucks and silo tank milk were collected. Analyses were performed by the Regional Farmers Association laboratory using high-performance liquid chromatography following the ISO 14501:1998 standard. None of the sheep milk samples analysed during 2005-2012 showed AFM 1 contamination. In sheep milk samples collected in 2013, 8 out of 172 (4.6%) were contaminated by AFM 1 with a concentration (mean±SD) of 12.59±14.05 ng/L. In one bulk tank milk sample 58.82 ng/L AFM 1 was detected, exceeding the EU limit. In none of goat milk samples analysed from 2010 to 2012 AFM 1 was detected. In 2013, 9 out of 66 goat milk samples (13.6%) showed an AFM 1 concentration of 47.21±19.58 ng/L. Two of these samples exceeded the EU limit, with concentrations of 62.09 and 138.6 ng/L. Higher contamination frequency and concentration rates were detected in bulk tank milk samples collected at farm than in bulk milk truck or silo samples, showing a dilution effect on AFM 1 milk content along small ruminants supply chain. The rate and levels of AFM 1 contamination in sheep and goat milk samples were lower than other countries. However, the small number of milk samples analysed for AFM 1 in Sardinia Region in 2005-2013 give evidence that food business operators check programmes should be improved to ensure an adequate monitoring of AFM 1 contamination in small ruminant dairy chain.
Direct comparison of nanoindentation and tensile test results on reactor-irradiated materials
NASA Astrophysics Data System (ADS)
Krumwiede, D. L.; Yamamoto, T.; Saleh, T. A.; Maloy, S. A.; Odette, G. R.; Hosemann, P.
2018-06-01
Nanoindentation testing has been used for decades to assess materials on a local scale and to obtain fundamental mechanical property parameters. Nuclear materials research often faces the challenge of testing rather small samples due to the hazardous nature, limited space in reactors, and shallow ion-irradiated zones, fostering the need for small-scale mechanical testing (SSMT). As such, correlating the results from SSMT to bulk properties is particularly of interest. This study compares macroscopic tensile test data (yield and flow stresses) to nanoindentation data (hardness) obtained on a number of different neutron-irradiated materials in order to understand the scaling behavior on radiation-damaged samples.
ERIC Educational Resources Information Center
Papenberg, Martin; Musch, Jochen
2017-01-01
In multiple-choice tests, the quality of distractors may be more important than their number. We therefore examined the joint influence of distractor quality and quantity on test functioning by providing a sample of 5,793 participants with five parallel test sets consisting of items that differed in the number and quality of distractors.…
Detection of small number of Giardia in biological materials prepared from stray dogs.
Esmailikia, Leila; Ebrahimzade, Elahe; Shayan, Parviz; Amininia, Narges
2017-12-20
Giardia lamblia is an intestinal protozoa with intermittent and low shedding especially in dogs, and the detection of Giardia is accompanied with problems such as sampling and diagnostic method. The objective of this study was to detection of Giardia in biological materials with low number of parasite using parasitological and molecular methods, and also to determine whether the examined stray dogs harbor known zoonotic genotype of Giardia. For this aim 85 fecal and duodenal samples were studied from which 1 was positive by Trichrome staining of stool, 4 were positive by staining of duodenal samples. The nested PCR analysis with primers derived from 18 SrRNA showed that the specific PCR product could be amplified in 4 stool and 4 duodenal samples. All positive samples in staining analysis were also positive in nested PCR. No amplification could be observed by nested PCR with primers derived from β giardin gene due to the single copy of gene. Interestingly, the extracted DNA from old fixed stained Giardia positive smears could be also amplified with primers derived from 18SrRNA gene. The sequence analysis of nested PCR products showed that they belong to the genotype D. In conclusion, it is to denote that the Trichrome or Giemsa methods were not suitable for the detection of small number of this parasite in stool and the nested PCR with primers derived from 18S rRNA gene can replace the traditional methods successfully. For detection of Giardia in stool, primers derived from β giardin will not be recommended.
Set size and culture influence children's attention to number.
Cantrell, Lisa; Kuwabara, Megumi; Smith, Linda B
2015-03-01
Much research evidences a system in adults and young children for approximately representing quantity. Here we provide evidence that the bias to attend to discrete quantity versus other dimensions may be mediated by set size and culture. Preschool-age English-speaking children in the United States and Japanese-speaking children in Japan were tested in a match-to-sample task where number was pitted against cumulative surface area in both large and small numerical set comparisons. Results showed that children from both cultures were biased to attend to the number of items for small sets. Large set responses also showed a general attention to number when ratio difficulty was easy. However, relative to the responses for small sets, attention to number decreased for both groups; moreover, both U.S. and Japanese children showed a significant bias to attend to total amount for difficult numerical ratio distances, although Japanese children shifted attention to total area at relatively smaller set sizes than U.S. children. These results add to our growing understanding of how quantity is represented and how such representation is influenced by context--both cultural and perceptual. Copyright © 2014 Elsevier Inc. All rights reserved.
A. Broido; Hsiukang Yow
1977-01-01
Even before weight loss in the low-temperature pyrolysis of cellulose becomes significant, the average degree of polymerization of the partially pyrolyzed samples drops sharply. The gel permeation chromatograms of nitrated derivatives of the samples can be described in terms of a small number of mixed size populationsâeach component fitted within reasonable limits by a...
Areal Control Using Generalized Least Squares As An Alternative to Stratification
Raymond L. Czaplewski
2001-01-01
Stratification for both variance reduction and areal control proliferates the number of strata, which causes small sample sizes in many strata. This might compromise statistical efficiency. Generalized least squares can, in principle, replace stratification for areal control.
McTwo: a two-step feature selection algorithm based on maximal information coefficient.
Ge, Ruiquan; Zhou, Manli; Luo, Youxi; Meng, Qinghan; Mai, Guoqin; Ma, Dongli; Wang, Guoqing; Zhou, Fengfeng
2016-03-23
High-throughput bio-OMIC technologies are producing high-dimension data from bio-samples at an ever increasing rate, whereas the training sample number in a traditional experiment remains small due to various difficulties. This "large p, small n" paradigm in the area of biomedical "big data" may be at least partly solved by feature selection algorithms, which select only features significantly associated with phenotypes. Feature selection is an NP-hard problem. Due to the exponentially increased time requirement for finding the globally optimal solution, all the existing feature selection algorithms employ heuristic rules to find locally optimal solutions, and their solutions achieve different performances on different datasets. This work describes a feature selection algorithm based on a recently published correlation measurement, Maximal Information Coefficient (MIC). The proposed algorithm, McTwo, aims to select features associated with phenotypes, independently of each other, and achieving high classification performance of the nearest neighbor algorithm. Based on the comparative study of 17 datasets, McTwo performs about as well as or better than existing algorithms, with significantly reduced numbers of selected features. The features selected by McTwo also appear to have particular biomedical relevance to the phenotypes from the literature. McTwo selects a feature subset with very good classification performance, as well as a small feature number. So McTwo may represent a complementary feature selection algorithm for the high-dimensional biomedical datasets.
Khondoker, Mizanur; Dobson, Richard; Skirrow, Caroline; Simmons, Andrew; Stahl, Daniel
2016-10-01
Recent literature on the comparison of machine learning methods has raised questions about the neutrality, unbiasedness and utility of many comparative studies. Reporting of results on favourable datasets and sampling error in the estimated performance measures based on single samples are thought to be the major sources of bias in such comparisons. Better performance in one or a few instances does not necessarily imply so on an average or on a population level and simulation studies may be a better alternative for objectively comparing the performances of machine learning algorithms. We compare the classification performance of a number of important and widely used machine learning algorithms, namely the Random Forests (RF), Support Vector Machines (SVM), Linear Discriminant Analysis (LDA) and k-Nearest Neighbour (kNN). Using massively parallel processing on high-performance supercomputers, we compare the generalisation errors at various combinations of levels of several factors: number of features, training sample size, biological variation, experimental variation, effect size, replication and correlation between features. For smaller number of correlated features, number of features not exceeding approximately half the sample size, LDA was found to be the method of choice in terms of average generalisation errors as well as stability (precision) of error estimates. SVM (with RBF kernel) outperforms LDA as well as RF and kNN by a clear margin as the feature set gets larger provided the sample size is not too small (at least 20). The performance of kNN also improves as the number of features grows and outplays that of LDA and RF unless the data variability is too high and/or effect sizes are too small. RF was found to outperform only kNN in some instances where the data are more variable and have smaller effect sizes, in which cases it also provide more stable error estimates than kNN and LDA. Applications to a number of real datasets supported the findings from the simulation study. © The Author(s) 2013.
Preparation of highly multiplexed small RNA sequencing libraries.
Persson, Helena; Søkilde, Rolf; Pirona, Anna Chiara; Rovira, Carlos
2017-08-01
MicroRNAs (miRNAs) are ~22-nucleotide-long small non-coding RNAs that regulate the expression of protein-coding genes by base pairing to partially complementary target sites, preferentially located in the 3´ untranslated region (UTR) of target mRNAs. The expression and function of miRNAs have been extensively studied in human disease, as well as the possibility of using these molecules as biomarkers for prognostication and treatment guidance. To identify and validate miRNAs as biomarkers, their expression must be screened in large collections of patient samples. Here, we develop a scalable protocol for the rapid and economical preparation of a large number of small RNA sequencing libraries using dual indexing for multiplexing. Combined with the use of off-the-shelf reagents, more samples can be sequenced simultaneously on large-scale sequencing platforms at a considerably lower cost per sample. Sample preparation is simplified by pooling libraries prior to gel purification, which allows for the selection of a narrow size range while minimizing sample variation. A comparison with publicly available data from benchmarking of miRNA analysis platforms showed that this method captures absolute and differential expression as effectively as commercially available alternatives.
Internal pilots for a class of linear mixed models with Gaussian and compound symmetric data
Gurka, Matthew J.; Coffey, Christopher S.; Muller, Keith E.
2015-01-01
SUMMARY An internal pilot design uses interim sample size analysis, without interim data analysis, to adjust the final number of observations. The approach helps to choose a sample size sufficiently large (to achieve the statistical power desired), but not too large (which would waste money and time). We report on recent research in cerebral vascular tortuosity (curvature in three dimensions) which would benefit greatly from internal pilots due to uncertainty in the parameters of the covariance matrix used for study planning. Unfortunately, observations correlated across the four regions of the brain and small sample sizes preclude using existing methods. However, as in a wide range of medical imaging studies, tortuosity data have no missing or mistimed data, a factorial within-subject design, the same between-subject design for all responses, and a Gaussian distribution with compound symmetry. For such restricted models, we extend exact, small sample univariate methods for internal pilots to linear mixed models with any between-subject design (not just two groups). Planning a new tortuosity study illustrates how the new methods help to avoid sample sizes that are too small or too large while still controlling the type I error rate. PMID:17318914
DOE Office of Scientific and Technical Information (OSTI.GOV)
McNally, N.; Liu, Xiang Yang; Choudary, P.V.
1997-01-01
The authors describe a microplate-based high-throughput procedure for rapid assay of the enzyme activities of nitrate reductase and nitrite reductase, using extremely small volumes of reagents. The new procedure offers the advantages of rapidity, small sample size-nanoliter volumes, low cost, and a dramatic increase in the throughput sample number that can be analyzed simultaneously. Additional advantages can be accessed by using microplate reader application software packages that permit assigning a group type to the wells, recording of the data on exportable data files and exercising the option of using the kinetic or endpoint reading modes. The assay can also bemore » used independently for detecting nitrite residues/contamination in environmental/food samples. 10 refs., 2 figs.« less
Ferrero, Giulio; Cordero, Francesca; Tarallo, Sonia; Arigoni, Maddalena; Riccardo, Federica; Gallo, Gaetano; Ronco, Guglielmo; Allasia, Marco; Kulkarni, Neha; Matullo, Giuseppe; Vineis, Paolo; Calogero, Raffaele A; Pardini, Barbara; Naccarati, Alessio
2018-01-09
The role of non-coding RNAs in different biological processes and diseases is continuously expanding. Next-generation sequencing together with the parallel improvement of bioinformatics analyses allows the accurate detection and quantification of an increasing number of RNA species. With the aim of exploring new potential biomarkers for disease classification, a clear overview of the expression levels of common/unique small RNA species among different biospecimens is necessary. However, except for miRNAs in plasma, there are no substantial indications about the pattern of expression of various small RNAs in multiple specimens among healthy humans. By analysing small RNA-sequencing data from 243 samples, we have identified and compared the most abundantly and uniformly expressed miRNAs and non-miRNA species of comparable size with the library preparation in four different specimens (plasma exosomes, stool, urine, and cervical scrapes). Eleven miRNAs were commonly detected among all different specimens while 231 miRNAs were globally unique across them. Classification analysis using these miRNAs provided an accuracy of 99.6% to recognize the sample types. piRNAs and tRNAs were the most represented non-miRNA small RNAs detected in all specimen types that were analysed, particularly in urine samples. With the present data, the most uniformly expressed small RNAs in each sample type were also identified. A signature of small RNAs for each specimen could represent a reference gene set in validation studies by RT-qPCR. Overall, the data reported hereby provide an insight of the constitution of the human miRNome and of other small non-coding RNAs in various specimens of healthy individuals.
Sampling procedures for throughfall monitoring: A simulation study
NASA Astrophysics Data System (ADS)
Zimmermann, Beate; Zimmermann, Alexander; Lark, Richard Murray; Elsenbeer, Helmut
2010-01-01
What is the most appropriate sampling scheme to estimate event-based average throughfall? A satisfactory answer to this seemingly simple question has yet to be found, a failure which we attribute to previous efforts' dependence on empirical studies. Here we try to answer this question by simulating stochastic throughfall fields based on parameters for statistical models of large monitoring data sets. We subsequently sampled these fields with different sampling designs and variable sample supports. We evaluated the performance of a particular sampling scheme with respect to the uncertainty of possible estimated means of throughfall volumes. Even for a relative error limit of 20%, an impractically large number of small, funnel-type collectors would be required to estimate mean throughfall, particularly for small events. While stratification of the target area is not superior to simple random sampling, cluster random sampling involves the risk of being less efficient. A larger sample support, e.g., the use of trough-type collectors, considerably reduces the necessary sample sizes and eliminates the sensitivity of the mean to outliers. Since the gain in time associated with the manual handling of troughs versus funnels depends on the local precipitation regime, the employment of automatically recording clusters of long troughs emerges as the most promising sampling scheme. Even so, a relative error of less than 5% appears out of reach for throughfall under heterogeneous canopies. We therefore suspect a considerable uncertainty of input parameters for interception models derived from measured throughfall, in particular, for those requiring data of small throughfall events.
Sampling and counting genome rearrangement scenarios
2015-01-01
Background Even for moderate size inputs, there are a tremendous number of optimal rearrangement scenarios, regardless what the model is and which specific question is to be answered. Therefore giving one optimal solution might be misleading and cannot be used for statistical inferring. Statistically well funded methods are necessary to sample uniformly from the solution space and then a small number of samples are sufficient for statistical inferring. Contribution In this paper, we give a mini-review about the state-of-the-art of sampling and counting rearrangement scenarios, focusing on the reversal, DCJ and SCJ models. Above that, we also give a Gibbs sampler for sampling most parsimonious labeling of evolutionary trees under the SCJ model. The method has been implemented and tested on real life data. The software package together with example data can be downloaded from http://www.renyi.hu/~miklosi/SCJ-Gibbs/ PMID:26452124
Vitamin D receptor gene and osteoporosis - author`s response
DOE Office of Scientific and Technical Information (OSTI.GOV)
Looney, J.E.; Yoon, Hyun Koo; Fischer, M.
1996-04-01
We appreciate the comments of Dr. Nguyen et al. about our recent study, but we disagree with their suggestion that the lack of an association between low bone density and the BB VDR genotype, which we reported, is an artifact generated by the small sample size. Furthermore, our results are consistent with similar conclusions reached by a number of other investigators, as recently reported by Peacock. Peacock states {open_quotes}Taken as a whole, the results of studies outlined ... indicate that VDR alleles, cannot account for the major part of the heritable component of bone density as indicated by Morrison etmore » al.{close_quotes}. The majority of the 17 studies cited in this editorial could not confirm an association between the VDR genotype and the bone phenotype. Surely one cannot criticize this combined work as representing an artifact because of a too small sample size. We do not dispute the suggestion by Nguyen et al. that large sample sizes are required to analyze small biological effects. This is evident in both Peacock`s summary and in their own bone density studies. We did not design our study with a larger sample size because, based on the work of Morrison et al., we had hypothesized a large biological effect; large sample sizes are only needed for small biological effects. 4 refs.« less
The cost of large numbers of hypothesis tests on power, effect size and sample size.
Lazzeroni, L C; Ray, A
2012-01-01
Advances in high-throughput biology and computer science are driving an exponential increase in the number of hypothesis tests in genomics and other scientific disciplines. Studies using current genotyping platforms frequently include a million or more tests. In addition to the monetary cost, this increase imposes a statistical cost owing to the multiple testing corrections needed to avoid large numbers of false-positive results. To safeguard against the resulting loss of power, some have suggested sample sizes on the order of tens of thousands that can be impractical for many diseases or may lower the quality of phenotypic measurements. This study examines the relationship between the number of tests on the one hand and power, detectable effect size or required sample size on the other. We show that once the number of tests is large, power can be maintained at a constant level, with comparatively small increases in the effect size or sample size. For example at the 0.05 significance level, a 13% increase in sample size is needed to maintain 80% power for ten million tests compared with one million tests, whereas a 70% increase in sample size is needed for 10 tests compared with a single test. Relative costs are less when measured by increases in the detectable effect size. We provide an interactive Excel calculator to compute power, effect size or sample size when comparing study designs or genome platforms involving different numbers of hypothesis tests. The results are reassuring in an era of extreme multiple testing.
Hancock, Bruno C; Ketterhagen, William R
2011-10-14
Discrete element model (DEM) simulations of the discharge of powders from hoppers under gravity were analyzed to provide estimates of dosage form content uniformity during the manufacture of solid dosage forms (tablets and capsules). For a system that exhibits moderate segregation the effects of sample size, number, and location within the batch were determined. The various sampling approaches were compared to current best-practices for sampling described in the Product Quality Research Institute (PQRI) Blend Uniformity Working Group (BUWG) guidelines. Sampling uniformly across the discharge process gave the most accurate results with respect to identifying segregation trends. Sigmoidal sampling (as recommended in the PQRI BUWG guidelines) tended to overestimate potential segregation issues, whereas truncated sampling (common in industrial practice) tended to underestimate them. The size of the sample had a major effect on the absolute potency RSD. The number of sampling locations (10 vs. 20) had very little effect on the trends in the data, and the number of samples analyzed at each location (1 vs. 3 vs. 7) had only a small effect for the sampling conditions examined. The results of this work provide greater understanding of the effect of different sampling approaches on the measured content uniformity of real dosage forms, and can help to guide the choice of appropriate sampling protocols. Copyright © 2011 Elsevier B.V. All rights reserved.
Klinkenberg, Don; Thomas, Ekelijn; Artavia, Francisco F Calvo; Bouma, Annemarie
2011-08-01
Design of surveillance programs to detect infections could benefit from more insight into sampling schemes. We address the effect of sampling schemes for Salmonella Enteritidis surveillance in laying hens. Based on experimental estimates for the transmission rate in flocks, and the characteristics of an egg immunological test, we have simulated outbreaks with various sampling schemes, and with the current boot swab program with a 15-week sampling interval. Declaring a flock infected based on a single positive egg was not possible because test specificity was too low. Thus, a threshold number of positive eggs was defined to declare a flock infected, and, for small sample sizes, eggs from previous samplings had to be included in a cumulative sample to guarantee a minimum flock level specificity. Effectiveness of surveillance was measured by the proportion of outbreaks detected, and by the number of contaminated table eggs brought on the market. The boot swab program detected 90% of the outbreaks, with 75% fewer contaminated eggs compared to no surveillance, whereas the baseline egg program (30 eggs each 15 weeks) detected 86%, with 73% fewer contaminated eggs. We conclude that a larger sample size results in more detected outbreaks, whereas a smaller sampling interval decreases the number of contaminated eggs. Decreasing sample size and interval simultaneously reduces the number of contaminated eggs, but not indefinitely: the advantage of more frequent sampling is counterbalanced by the cumulative sample including less recently laid eggs. Apparently, optimizing surveillance has its limits when test specificity is taken into account. © 2011 Society for Risk Analysis.
Competitive intelligence information management and innovation in small technology-based companies
NASA Astrophysics Data System (ADS)
Tanev, Stoyan
2007-05-01
In this article we examine how (i) company type and (ii) the competitive intelligence information used by small technology-based companies affect their innovation performance. The focus is on the specific information types used and not on the information sources. Information topics are classified in four groups - customers (10), company (9), competitor (11) and industry (12). The sample consists of 45 small new technology-based companies, specialized suppliers, and service companies from a variety of sectors - software, photonics, telecommunications, biomedical engineering and biotech, traditional manufacturing etc. The results suggest that the total number of intelligence information topics companies use to make decisions about innovation is not associated with the number of their new products, processes, services and patents. Therefore the companies in our sample do not seem to have the resources, processes or value systems required to use different competitive intelligence information when making decisions on innovation or may rely more on their own internal logic than on external information. Companies are classified using a Pavitt-like taxonomy. Service companies are considered as a separate company type. This allows for explicitly studying both, the innovative role of new services in product driven companies, and the role of new product development in service companies.
Approximation of Failure Probability Using Conditional Sampling
NASA Technical Reports Server (NTRS)
Giesy. Daniel P.; Crespo, Luis G.; Kenney, Sean P.
2008-01-01
In analyzing systems which depend on uncertain parameters, one technique is to partition the uncertain parameter domain into a failure set and its complement, and judge the quality of the system by estimating the probability of failure. If this is done by a sampling technique such as Monte Carlo and the probability of failure is small, accurate approximation can require so many sample points that the computational expense is prohibitive. Previous work of the authors has shown how to bound the failure event by sets of such simple geometry that their probabilities can be calculated analytically. In this paper, it is shown how to make use of these failure bounding sets and conditional sampling within them to substantially reduce the computational burden of approximating failure probability. It is also shown how the use of these sampling techniques improves the confidence intervals for the failure probability estimate for a given number of sample points and how they reduce the number of sample point analyses needed to achieve a given level of confidence.
Resampling methods in Microsoft Excel® for estimating reference intervals
Theodorsson, Elvar
2015-01-01
Computer- intensive resampling/bootstrap methods are feasible when calculating reference intervals from non-Gaussian or small reference samples. Microsoft Excel® in version 2010 or later includes natural functions, which lend themselves well to this purpose including recommended interpolation procedures for estimating 2.5 and 97.5 percentiles. The purpose of this paper is to introduce the reader to resampling estimation techniques in general and in using Microsoft Excel® 2010 for the purpose of estimating reference intervals in particular. Parametric methods are preferable to resampling methods when the distributions of observations in the reference samples is Gaussian or can transformed to that distribution even when the number of reference samples is less than 120. Resampling methods are appropriate when the distribution of data from the reference samples is non-Gaussian and in case the number of reference individuals and corresponding samples are in the order of 40. At least 500-1000 random samples with replacement should be taken from the results of measurement of the reference samples. PMID:26527366
Resampling methods in Microsoft Excel® for estimating reference intervals.
Theodorsson, Elvar
2015-01-01
Computer-intensive resampling/bootstrap methods are feasible when calculating reference intervals from non-Gaussian or small reference samples. Microsoft Excel® in version 2010 or later includes natural functions, which lend themselves well to this purpose including recommended interpolation procedures for estimating 2.5 and 97.5 percentiles. The purpose of this paper is to introduce the reader to resampling estimation techniques in general and in using Microsoft Excel® 2010 for the purpose of estimating reference intervals in particular. Parametric methods are preferable to resampling methods when the distributions of observations in the reference samples is Gaussian or can transformed to that distribution even when the number of reference samples is less than 120. Resampling methods are appropriate when the distribution of data from the reference samples is non-Gaussian and in case the number of reference individuals and corresponding samples are in the order of 40. At least 500-1000 random samples with replacement should be taken from the results of measurement of the reference samples.
Automatic extraction and processing of small RNAs on a multi-well/multi-channel (M&M) chip.
Zhong, Runtao; Flack, Kenneth; Zhong, Wenwan
2012-12-07
The study of the regulatory roles in small RNAs can be accelerated by techniques that permit simple, low-cost, and rapid extraction of small RNAs from a small number of cells. In order to ensure highly specific and sensitive detection, the extracted RNAs should be free of the background nucleic acids and present stably in a small volume. To meet these criteria, we designed a multi-well/multi-channel (M&M) chip to carry out automatic and selective isolation of small RNAs via solid-phase extraction (SPE), followed by reverse-transcription (RT) to convert them to the more stable cDNAs in a final volume of 2 μL. Droplets containing buffers for RNA binding, washing, and elution were trapped in microwells, which were connected by one channel, and suspended in mineral oil. The silica magnetic particles (SMPs) for SPE were moved along the channel from well to well, i.e. in between droplets, by a fixed magnet and a translation stage, allowing the nucleic acid fragments to bind to the SMPs, be washed, and then be eluted for RT reaction within 15 minutes. RNAs shorter than 63 nt were selectively enriched from cell lysates, with recovery comparable to that of a commercial kit. Physical separation of the droplets on our M&M chip allowed the usage of multiple channels for parallel processing of multiple samples. It also permitted smooth integration with on-chip RT-PCR, which simultaneously detected the target microRNA, mir-191, expressed in fewer than 10 cancer cells. Our results have demonstrated that the M&M chip device is a valuable and cost-saving platform for studying small RNA expression patterns in a limited number of cells with reasonable sample throughput.
The heterogeneity statistic I(2) can be biased in small meta-analyses.
von Hippel, Paul T
2015-04-14
Estimated effects vary across studies, partly because of random sampling error and partly because of heterogeneity. In meta-analysis, the fraction of variance that is due to heterogeneity is estimated by the statistic I(2). We calculate the bias of I(2), focusing on the situation where the number of studies in the meta-analysis is small. Small meta-analyses are common; in the Cochrane Library, the median number of studies per meta-analysis is 7 or fewer. We use Mathematica software to calculate the expectation and bias of I(2). I(2) has a substantial bias when the number of studies is small. The bias is positive when the true fraction of heterogeneity is small, but the bias is typically negative when the true fraction of heterogeneity is large. For example, with 7 studies and no true heterogeneity, I(2) will overestimate heterogeneity by an average of 12 percentage points, but with 7 studies and 80 percent true heterogeneity, I(2) can underestimate heterogeneity by an average of 28 percentage points. Biases of 12-28 percentage points are not trivial when one considers that, in the Cochrane Library, the median I(2) estimate is 21 percent. The point estimate I(2) should be interpreted cautiously when a meta-analysis has few studies. In small meta-analyses, confidence intervals should supplement or replace the biased point estimate I(2).
NASA Astrophysics Data System (ADS)
Qiang, Wei
2011-12-01
We describe a sampling scheme for the two-dimensional (2D) solid state NMR experiments, which can be readily applied to the sensitivity-limited samples. The sampling scheme utilizes continuous, non-uniform sampling profile for the indirect dimension, i.e. the acquisition number decreases as a function of the evolution time ( t1) in the indirect dimension. For a beta amyloid (Aβ) fibril sample, we observed overall 40-50% signal enhancement by measuring the cross peak volume, while the cross peak linewidths remained comparable to the linewidths obtained by regular sampling and processing strategies. Both the linear and Gaussian decay functions for the acquisition numbers result in similar percentage of increment in signal. In addition, we demonstrated that this sampling approach can be applied with different dipolar recoupling approaches such as radiofrequency assisted diffusion (RAD) and finite-pulse radio-frequency-driven recoupling (fpRFDR). This sampling scheme is especially suitable for the sensitivity-limited samples which require long signal averaging for each t1 point, for instance the biological membrane proteins where only a small fraction of the sample is isotopically labeled.
Waage, Astrid S.; Vardund, Traute; Lund, Vidar; Kapperud, Georg
1999-01-01
A rapid and sensitive assay was developed for detection of small numbers of Campylobacter jejuni and Campylobacter coli cells in environmental water, sewage, and food samples. Water and sewage samples were filtered, and the filters were enriched overnight in a nonselective medium. The enrichment cultures were prepared for PCR by a rapid and simple procedure consisting of centrifugation, proteinase K treatment, and boiling. A seminested PCR based on specific amplification of the intergenic sequence between the two Campylobacter flagellin genes, flaA and flaB, was performed, and the PCR products were visualized by agarose gel electrophoresis. The assay allowed us to detect 3 to 15 CFU of C. jejuni per 100 ml in water samples containing a background flora consisting of up to 8,700 heterotrophic organisms per ml and 10,000 CFU of coliform bacteria per 100 ml. Dilution of the enriched cultures 1:10 with sterile broth prior to the PCR was sometimes necessary to obtain positive results. The assay was also conducted with food samples analyzed with or without overnight enrichment. As few as ≤3 CFU per g of food could be detected with samples subjected to overnight enrichment, while variable results were obtained for samples analyzed without prior enrichment. This rapid and sensitive nested PCR assay provides a useful tool for specific detection of C. jejuni or C. coli in drinking water, as well as environmental water, sewage, and food samples containing high levels of background organisms. PMID:10103261
2013-01-01
Background Gene expression data could likely be a momentous help in the progress of proficient cancer diagnoses and classification platforms. Lately, many researchers analyze gene expression data using diverse computational intelligence methods, for selecting a small subset of informative genes from the data for cancer classification. Many computational methods face difficulties in selecting small subsets due to the small number of samples compared to the huge number of genes (high-dimension), irrelevant genes, and noisy genes. Methods We propose an enhanced binary particle swarm optimization to perform the selection of small subsets of informative genes which is significant for cancer classification. Particle speed, rule, and modified sigmoid function are introduced in this proposed method to increase the probability of the bits in a particle’s position to be zero. The method was empirically applied to a suite of ten well-known benchmark gene expression data sets. Results The performance of the proposed method proved to be superior to other previous related works, including the conventional version of binary particle swarm optimization (BPSO) in terms of classification accuracy and the number of selected genes. The proposed method also requires lower computational time compared to BPSO. PMID:23617960
Preparation and physical characterization of pure beta-carotene.
Laughlin, Robert G; Bunke, Gregory M; Eads, Charles D; Laidig, William D; Shelley, John C
2002-05-01
Pure all-trans beta-carotene has been prepared on the 10's of grams scale by isothermal Fractional Dissolution (FD) of commercial laboratory samples in tetrahydrofuran (THF). beta-Carotene purified in this way is black, with a faint brownish tinge. The electronic spectra of black samples extend into the near infrared, with end-absorption past 750 nm. Black samples react directly with dioxygen under mild conditions to yield the familiar orange or red powders. Pure beta-carotene rigorously obeys Beer's Law in octane over the entire UV-Vis spectral range, while commercial laboratory samples and recrystallized samples do not. NMR self-diffusion coefficient data demonstrate that beta-carotene exists as simple molecular solutions in octane and toluene. The anomalously high crystallinity of beta-carotene can be attributed (from analysis using molecular mechanics) to the facts that: (1) the number of theoretically possible conformers of beta-carotene is extremely small, and (2) only a small fraction of these (ca. 12%, or 127) may actually exist in fluid phases.
Gu, Weidong; Medalla, Felicita; Hoekstra, Robert M
2018-02-01
The National Antimicrobial Resistance Monitoring System (NARMS) at the Centers for Disease Control and Prevention tracks resistance among Salmonella infections. The annual number of Salmonella isolates of a particular serotype from states may be small, making direct estimation of resistance proportions unreliable. We developed a Bayesian hierarchical model to improve estimation by borrowing strength from relevant sampling units. We illustrate the models with different specifications of spatio-temporal interaction using 2004-2013 NARMS data for ceftriaxone-resistant Salmonella serotype Heidelberg. Our results show that Bayesian estimates of resistance proportions were smoother than observed values, and the difference between predicted and observed proportions was inversely related to the number of submitted isolates. The model with interaction allowed for tracking of annual changes in resistance proportions at the state level. We demonstrated that Bayesian hierarchical models provide a useful tool to examine spatio-temporal patterns of small sample size such as those found in NARMS. Published by Elsevier Ltd.
Matching algorithm of missile tail flame based on back-propagation neural network
NASA Astrophysics Data System (ADS)
Huang, Da; Huang, Shucai; Tang, Yidong; Zhao, Wei; Cao, Wenhuan
2018-02-01
This work presents a spectral matching algorithm of missile plume detection that based on neural network. The radiation value of the characteristic spectrum of the missile tail flame is taken as the input of the network. The network's structure including the number of nodes and layers is determined according to the number of characteristic spectral bands and missile types. We can get the network weight matrixes and threshold vectors through training the network using training samples, and we can determine the performance of the network through testing the network using the test samples. A small amount of data cause the network has the advantages of simple structure and practicality. Network structure composed of weight matrix and threshold vector can complete task of spectrum matching without large database support. Network can achieve real-time requirements with a small quantity of data. Experiment results show that the algorithm has the ability to match the precise spectrum and strong robustness.
Direct comparison of nanoindentation and tensile test results on reactor-irradiated materials
Krumweide, David L; Yamamoto, Takuya; Saleh, Tarik A.; ...
2018-03-13
Nanoindentation testing has been used for decades to assess materials on a local scale and to obtain fundamental mechanical property parameters. Nuclear materials research often faces the challenge of testing rather small samples due to the hazardous nature, limited space in reactors, and shallow ion-irradiated zones, fostering the need for small-scale mechanical testing (SSMT). As such, correlating the results from SSMT to bulk properties is particularly of interest. Here, this study compares macroscopic tensile test data (yield and flow stresses) to nanoindentation data (hardness) obtained on a number of different neutron-irradiated materials in order to understand the scaling behavior onmore » radiation-damaged samples.« less
Direct comparison of nanoindentation and tensile test results on reactor-irradiated materials
DOE Office of Scientific and Technical Information (OSTI.GOV)
Krumweide, David L; Yamamoto, Takuya; Saleh, Tarik A.
Nanoindentation testing has been used for decades to assess materials on a local scale and to obtain fundamental mechanical property parameters. Nuclear materials research often faces the challenge of testing rather small samples due to the hazardous nature, limited space in reactors, and shallow ion-irradiated zones, fostering the need for small-scale mechanical testing (SSMT). As such, correlating the results from SSMT to bulk properties is particularly of interest. Here, this study compares macroscopic tensile test data (yield and flow stresses) to nanoindentation data (hardness) obtained on a number of different neutron-irradiated materials in order to understand the scaling behavior onmore » radiation-damaged samples.« less
Bayesian road safety analysis: incorporation of past evidence and effect of hyper-prior choice.
Miranda-Moreno, Luis F; Heydari, Shahram; Lord, Dominique; Fu, Liping
2013-09-01
This paper aims to address two related issues when applying hierarchical Bayesian models for road safety analysis, namely: (a) how to incorporate available information from previous studies or past experiences in the (hyper) prior distributions for model parameters and (b) what are the potential benefits of incorporating past evidence on the results of a road safety analysis when working with scarce accident data (i.e., when calibrating models with crash datasets characterized by a very low average number of accidents and a small number of sites). A simulation framework was developed to evaluate the performance of alternative hyper-priors including informative and non-informative Gamma, Pareto, as well as Uniform distributions. Based on this simulation framework, different data scenarios (i.e., number of observations and years of data) were defined and tested using crash data collected at 3-legged rural intersections in California and crash data collected for rural 4-lane highway segments in Texas. This study shows how the accuracy of model parameter estimates (inverse dispersion parameter) is considerably improved when incorporating past evidence, in particular when working with the small number of observations and crash data with low mean. The results also illustrates that when the sample size (more than 100 sites) and the number of years of crash data is relatively large, neither the incorporation of past experience nor the choice of the hyper-prior distribution may affect the final results of a traffic safety analysis. As a potential solution to the problem of low sample mean and small sample size, this paper suggests some practical guidance on how to incorporate past evidence into informative hyper-priors. By combining evidence from past studies and data available, the model parameter estimates can significantly be improved. The effect of prior choice seems to be less important on the hotspot identification. The results show the benefits of incorporating prior information when working with limited crash data in road safety studies. Copyright © 2013 National Safety Council and Elsevier Ltd. All rights reserved.
Capece, Paula I; Aliaga-Rossel, Enzo; Jansen, Patrick A
2013-03-01
Tapirs are known as effective dispersers of large-seeded tree species, but their role in dispersing small-seeded plant species has yet to be established. Tapir feces have been reported to contain large numbers of small seeds, but whether these are viable has rarely been evaluated. We determined the abundance and viability of small seeds in feces of Central American tapir (Tapirus bairdii) on Barro Colorado Island, Panama. A total of 72 fecal samples were collected opportunistically from 4 tapir latrine sites. Seeds were manually extracted from feces and classified by size. Seed viability was estimated by opening each seed and examining for the presence of at least 1 intact firm white endosperm. In total, we obtained 8166 seeds of at least 16 plant species. Small-seeded species dominated, with 96% of all seeds found measuring <5 mm. The canopy tree Laetia procera was the most abundant species in the samples. Of all small seeds found, 69% contained an intact endosperm and appeared viable. This suggests that small seeds, like large seeds, often pass through the digestive tract of T. bairdii intact. Thus, tapirs potentially serve as effective dispersers of a wide range of small-seeded plant species. © 2012 Wiley Publishing Asia Pty Ltd, ISZS and IOZ/CAS.
NASA Astrophysics Data System (ADS)
Glazner, Allen F.; Sadler, Peter M.
2016-12-01
The duration of a geologic interval, such as the time over which a given volume of magma accumulated to form a pluton, or the lifespan of a large igneous province, is commonly determined from a relatively small number of geochronologic determinations (e.g., 4-10) within that interval. Such sample sets can underestimate the true length of the interval by a significant amount. For example, the average interval determined from a sample of size n = 5, drawn from a uniform random distribution, will underestimate the true interval by 50%. Even for n = 10, the average sample only captures ˜80% of the interval. If the underlying distribution is known then a correction factor can be determined from theory or Monte Carlo analysis; for a uniform random distribution, this factor is
Opportunities and Challenges of Linking Scientific Core Samples to the Geoscience Data Ecosystem
NASA Astrophysics Data System (ADS)
Noren, A. J.
2016-12-01
Core samples generated in scientific drilling and coring are critical for the advancement of the Earth Sciences. The scientific themes enabled by analysis of these samples are diverse, and include plate tectonics, ocean circulation, Earth-life system interactions (paleoclimate, paleobiology, paleoanthropology), Critical Zone processes, geothermal systems, deep biosphere, and many others, and substantial resources are invested in their collection and analysis. Linking core samples to researchers, datasets, publications, and funding agencies through registration of globally unique identifiers such as International Geo Sample Numbers (IGSNs) offers great potential for advancing several frontiers. These include maximizing sample discoverability, access, reuse, and return on investment; a means for credit to researchers; and documentation of project outputs to funding agencies. Thousands of kilometers of core samples and billions of derivative subsamples have been generated through thousands of investigators' projects, yet the vast majority of these samples are curated at only a small number of facilities. These numbers, combined with the substantial similarity in sample types, make core samples a compelling target for IGSN implementation. However, differences between core sample communities and other geoscience disciplines continue to create barriers to implementation. Core samples involve parent-child relationships spanning 8 or more generations, an exponential increase in sample numbers between levels in the hierarchy, concepts related to depth/position in the sample, requirements for associating data derived from core scanning and lithologic description with data derived from subsample analysis, and publications based on tens of thousands of co-registered scan data points and thousands of analyses of subsamples. These characteristics require specialized resources for accurate and consistent assignment of IGSNs, and a community of practice to establish norms, workflows, and infrastructure to support implementation.
Wang, Shuaiqun; Aorigele; Kong, Wei; Zeng, Weiming; Hong, Xiaomin
2016-01-01
Gene expression data composed of thousands of genes play an important role in classification platforms and disease diagnosis. Hence, it is vital to select a small subset of salient features over a large number of gene expression data. Lately, many researchers devote themselves to feature selection using diverse computational intelligence methods. However, in the progress of selecting informative genes, many computational methods face difficulties in selecting small subsets for cancer classification due to the huge number of genes (high dimension) compared to the small number of samples, noisy genes, and irrelevant genes. In this paper, we propose a new hybrid algorithm HICATS incorporating imperialist competition algorithm (ICA) which performs global search and tabu search (TS) that conducts fine-tuned search. In order to verify the performance of the proposed algorithm HICATS, we have tested it on 10 well-known benchmark gene expression classification datasets with dimensions varying from 2308 to 12600. The performance of our proposed method proved to be superior to other related works including the conventional version of binary optimization algorithm in terms of classification accuracy and the number of selected genes.
Aorigele; Zeng, Weiming; Hong, Xiaomin
2016-01-01
Gene expression data composed of thousands of genes play an important role in classification platforms and disease diagnosis. Hence, it is vital to select a small subset of salient features over a large number of gene expression data. Lately, many researchers devote themselves to feature selection using diverse computational intelligence methods. However, in the progress of selecting informative genes, many computational methods face difficulties in selecting small subsets for cancer classification due to the huge number of genes (high dimension) compared to the small number of samples, noisy genes, and irrelevant genes. In this paper, we propose a new hybrid algorithm HICATS incorporating imperialist competition algorithm (ICA) which performs global search and tabu search (TS) that conducts fine-tuned search. In order to verify the performance of the proposed algorithm HICATS, we have tested it on 10 well-known benchmark gene expression classification datasets with dimensions varying from 2308 to 12600. The performance of our proposed method proved to be superior to other related works including the conventional version of binary optimization algorithm in terms of classification accuracy and the number of selected genes. PMID:27579323
Wickenberg-Bolin, Ulrika; Göransson, Hanna; Fryknäs, Mårten; Gustafsson, Mats G; Isaksson, Anders
2006-03-13
Supervised learning for classification of cancer employs a set of design examples to learn how to discriminate between tumors. In practice it is crucial to confirm that the classifier is robust with good generalization performance to new examples, or at least that it performs better than random guessing. A suggested alternative is to obtain a confidence interval of the error rate using repeated design and test sets selected from available examples. However, it is known that even in the ideal situation of repeated designs and tests with completely novel samples in each cycle, a small test set size leads to a large bias in the estimate of the true variance between design sets. Therefore different methods for small sample performance estimation such as a recently proposed procedure called Repeated Random Sampling (RSS) is also expected to result in heavily biased estimates, which in turn translates into biased confidence intervals. Here we explore such biases and develop a refined algorithm called Repeated Independent Design and Test (RIDT). Our simulations reveal that repeated designs and tests based on resampling in a fixed bag of samples yield a biased variance estimate. We also demonstrate that it is possible to obtain an improved variance estimate by means of a procedure that explicitly models how this bias depends on the number of samples used for testing. For the special case of repeated designs and tests using new samples for each design and test, we present an exact analytical expression for how the expected value of the bias decreases with the size of the test set. We show that via modeling and subsequent reduction of the small sample bias, it is possible to obtain an improved estimate of the variance of classifier performance between design sets. However, the uncertainty of the variance estimate is large in the simulations performed indicating that the method in its present form cannot be directly applied to small data sets.
Relationships between Perron-Frobenius eigenvalue and measurements of loops in networks
NASA Astrophysics Data System (ADS)
Chen, Lei; Kou, Yingxin; Li, Zhanwu; Xu, An; Chang, Yizhe
2018-07-01
The Perron-Frobenius eigenvalue (PFE) is widely used as measurement of the number of loops in networks, but what exactly the relationship between the PFE and the number of loops in networks is has not been researched yet, is it strictly monotonically increasing? And what are the relationships between the PFE and other measurements of loops in networks? Such as the average loop degree of nodes, and the distribution of loop ranks. We make researches on these questions based on samples of ER random network, NW small-world network and BA scale-free network, and the results confirm that, both the number of loops in network and the average loop degree of nodes of all samples do increase with the increase of the PFE in general trend, but neither of them are strictly monotonically increasing, so the PFE is capable to be used as a rough estimative measurement of the number of loops in networks and the average loop degree of nodes. Furthermore, we find that a majority of the loop ranks of all samples obey Weibull distribution, of which the scale parameter A and the shape parameter B have approximate power-law relationships with the PFE of the samples.
Exhaled particles as markers of small airway inflammation in subjects with asthma.
Larsson, Per; Lärstad, Mona; Bake, Björn; Hammar, Oscar; Bredberg, Anna; Almstrand, Ann-Charlotte; Mirgorodskaya, Ekaterina; Olin, Anna-Carin
2017-09-01
Exhaled breath contains suspended particles of respiratory tract lining fluid from the small airways. The particles are formed when closed airways open during inhalation. We have developed a method called Particles in Exhaled air (PExA ® ) to measure and sample these particles in the exhaled aerosol. Here, we use the PExA ® method to study the effects of birch pollen exposure on the small airways of individuals with asthma and birch pollen allergy. We hypothesized that birch pollen-induced inflammation could change the concentrations of surfactant protein A and albumin in the respiratory tract lining fluid of the small airways and influence the amount of exhaled particles. The amount of exhaled particles was reduced after birch pollen exposure in subjects with asthma and birch pollen allergy, but no significant effect on the concentrations of surfactant protein A and albumin in exhaled particles was found. The reduction in the number of exhaled particles may be due to inflammation in the small airways, which would reduce their diameter and potentially reduce the number of small airways that open and close during inhalation and exhalation. © 2015 The Authors. Clinical Physiology and Functional Imaging published by John Wiley & Sons Ltd.
Manabe, Sho; Morimoto, Chie; Hamano, Yuya; Fujimoto, Shuntaro
2017-01-01
In criminal investigations, forensic scientists need to evaluate DNA mixtures. The estimation of the number of contributors and evaluation of the contribution of a person of interest (POI) from these samples are challenging. In this study, we developed a new open-source software “Kongoh” for interpreting DNA mixture based on a quantitative continuous model. The model uses quantitative information of peak heights in the DNA profile and considers the effect of artifacts and allelic drop-out. By using this software, the likelihoods of 1–4 persons’ contributions are calculated, and the most optimal number of contributors is automatically determined; this differs from other open-source software. Therefore, we can eliminate the need to manually determine the number of contributors before the analysis. Kongoh also considers allele- or locus-specific effects of biological parameters based on the experimental data. We then validated Kongoh by calculating the likelihood ratio (LR) of a POI’s contribution in true contributors and non-contributors by using 2–4 person mixtures analyzed through a 15 short tandem repeat typing system. Most LR values obtained from Kongoh during true-contributor testing strongly supported the POI’s contribution even for small amounts or degraded DNA samples. Kongoh correctly rejected a false hypothesis in the non-contributor testing, generated reproducible LR values, and demonstrated higher accuracy of the estimated number of contributors than another software based on the quantitative continuous model. Therefore, Kongoh is useful in accurately interpreting DNA evidence like mixtures and small amounts or degraded DNA samples. PMID:29149210
Manabe, Sho; Morimoto, Chie; Hamano, Yuya; Fujimoto, Shuntaro; Tamaki, Keiji
2017-01-01
In criminal investigations, forensic scientists need to evaluate DNA mixtures. The estimation of the number of contributors and evaluation of the contribution of a person of interest (POI) from these samples are challenging. In this study, we developed a new open-source software "Kongoh" for interpreting DNA mixture based on a quantitative continuous model. The model uses quantitative information of peak heights in the DNA profile and considers the effect of artifacts and allelic drop-out. By using this software, the likelihoods of 1-4 persons' contributions are calculated, and the most optimal number of contributors is automatically determined; this differs from other open-source software. Therefore, we can eliminate the need to manually determine the number of contributors before the analysis. Kongoh also considers allele- or locus-specific effects of biological parameters based on the experimental data. We then validated Kongoh by calculating the likelihood ratio (LR) of a POI's contribution in true contributors and non-contributors by using 2-4 person mixtures analyzed through a 15 short tandem repeat typing system. Most LR values obtained from Kongoh during true-contributor testing strongly supported the POI's contribution even for small amounts or degraded DNA samples. Kongoh correctly rejected a false hypothesis in the non-contributor testing, generated reproducible LR values, and demonstrated higher accuracy of the estimated number of contributors than another software based on the quantitative continuous model. Therefore, Kongoh is useful in accurately interpreting DNA evidence like mixtures and small amounts or degraded DNA samples.
Ecologists are often faced with problem of small sample size, correlated and large number of predictors, and high noise-to-signal relationships. This necessitates excluding important variables from the model when applying standard multiple or multivariate regression analyses. In ...
Ide, Jun'ichiro; Chiwa, Masaaki; Higashi, Naoko; Maruno, Ryoko; Mori, Yasushi; Otsuki, Kyoichi
2012-08-01
This study sought to determine the lowest number of storm events required for adequate estimation of annual nutrient loads from a forested watershed using the regression equation between cumulative load (∑L) and cumulative stream discharge (∑Q). Hydrological surveys were conducted for 4 years, and stream water was sampled sequentially at 15-60-min intervals during 24 h in 20 events, as well as weekly in a small forested watershed. The bootstrap sampling technique was used to determine the regression (∑L-∑Q) equations of dissolved nitrogen (DN) and phosphorus (DP), particulate nitrogen (PN) and phosphorus (PP), dissolved inorganic nitrogen (DIN), and suspended solid (SS) for each dataset of ∑L and ∑Q. For dissolved nutrients (DN, DP, DIN), the coefficient of variance (CV) in 100 replicates of 4-year average annual load estimates was below 20% with datasets composed of five storm events. For particulate nutrients (PN, PP, SS), the CV exceeded 20%, even with datasets composed of more than ten storm events. The differences in the number of storm events required for precise load estimates between dissolved and particulate nutrients were attributed to the goodness of fit of the ∑L-∑Q equations. Bootstrap simulation based on flow-stratified sampling resulted in fewer storm events than the simulation based on random sampling and showed that only three storm events were required to give a CV below 20% for dissolved nutrients. These results indicate that a sampling design considering discharge levels reduces the frequency of laborious chemical analyses of water samples required throughout the year.
Training of polyp staging systems using mixed imaging modalities.
Wimmer, Georg; Gadermayr, Michael; Kwitt, Roland; Häfner, Michael; Tamaki, Toru; Yoshida, Shigeto; Tanaka, Shinji; Merhof, Dorit; Uhl, Andreas
2018-05-04
In medical image data sets, the number of images is usually quite small. The small number of training samples does not allow to properly train classifiers which leads to massive overfitting to the training data. In this work, we investigate whether increasing the number of training samples by merging datasets from different imaging modalities can be effectively applied to improve predictive performance. Further, we investigate if the extracted features from the employed image representations differ between different imaging modalities and if domain adaption helps to overcome these differences. We employ twelve feature extraction methods to differentiate between non-neoplastic and neoplastic lesions. Experiments are performed using four different classifier training strategies, each with a different combination of training data. The specifically designed setup for these experiments enables a fair comparison between the four training strategies. Combining high definition with high magnification training data and chromoscopic with non-chromoscopic training data partly improved the results. The usage of domain adaptation has only a small effect on the results compared to just using non-adapted training data. Merging datasets from different imaging modalities turned out to be partially beneficial for the case of combining high definition endoscopic data with high magnification endoscopic data and for combining chromoscopic with non-chromoscopic data. NBI and chromoendoscopy on the other hand are mostly too different with respect to the extracted features to combine images of these two modalities for classifier training. Copyright © 2018 Elsevier Ltd. All rights reserved.
The Very Small Scale Clustering of SDSS-II and SDSS-III Galaxies
NASA Astrophysics Data System (ADS)
Piscionere, Jennifer
2015-01-01
We measure the angular clustering of galaxies from the Sloan Digital Sky Survey Data Release 7 in order to probe the spatial distribution of satellite galaxies within their dark matter halos. Specifically, we measure the angular correlation function on very small scales (7 - 320‧‧) in a range of luminosity threshold samples (absolute r-band magnitudes of -18 up to -21) that are constructed from the subset of SDSS that has been spectroscopically observed more than once (the so-called plate overlap region). We choose to measure angular clustering in this reduced survey footprint in order to minimize the effects of fiber collision incompleteness, which are otherwise substantial on these small scales. We model our clustering measurements using a fully numerical halo model that populates dark matter halos in N-body simulations to create realistic mock galaxy catalogs. The model has free parameters that specify both the number and spatial distribution of galaxies within their host halos. We adopt a flexible density profile for the spatial distribution of satellite galaxies that is similar to the dark matter Navarro-Frenk-White (NFW) profile, except that the inner slope is allowed to vary. We find that the angular clustering of our most luminous samples (Mr < -20 and -21) suggests that luminous satellite galaxies have substantially steeper inner density profiles than NFW. Lower luminosity samples are less constraining, however, and are consistent with satellite galaxies having shallow density profiles. Our results confirm the findings of Watson et al. (2012) while using different clustering measurements and modeling methodology. With the new SDSS-III Baryon Oscillation Spectroscopic Survey (BOSS; Dawson et al., 2013), we can measure how the same class of galaxy evolves over time. The BOSS CMASS sample is of roughly constant stellar mass and number density out to z ˜ 0.6. The clustering of these samples appears to evolve very little with redshift, and each of the samples exhibit flattening of wp at roughly the same comoving distance of 100kpc.
Pedroza, Claudia; Truong, Van Thi Thanh
2017-11-02
Analyses of multicenter studies often need to account for center clustering to ensure valid inference. For binary outcomes, it is particularly challenging to properly adjust for center when the number of centers or total sample size is small, or when there are few events per center. Our objective was to evaluate the performance of generalized estimating equation (GEE) log-binomial and Poisson models, generalized linear mixed models (GLMMs) assuming binomial and Poisson distributions, and a Bayesian binomial GLMM to account for center effect in these scenarios. We conducted a simulation study with few centers (≤30) and 50 or fewer subjects per center, using both a randomized controlled trial and an observational study design to estimate relative risk. We compared the GEE and GLMM models with a log-binomial model without adjustment for clustering in terms of bias, root mean square error (RMSE), and coverage. For the Bayesian GLMM, we used informative neutral priors that are skeptical of large treatment effects that are almost never observed in studies of medical interventions. All frequentist methods exhibited little bias, and the RMSE was very similar across the models. The binomial GLMM had poor convergence rates, ranging from 27% to 85%, but performed well otherwise. The results show that both GEE models need to use small sample corrections for robust SEs to achieve proper coverage of 95% CIs. The Bayesian GLMM had similar convergence rates but resulted in slightly more biased estimates for the smallest sample sizes. However, it had the smallest RMSE and good coverage across all scenarios. These results were very similar for both study designs. For the analyses of multicenter studies with a binary outcome and few centers, we recommend adjustment for center with either a GEE log-binomial or Poisson model with appropriate small sample corrections or a Bayesian binomial GLMM with informative priors.
NASA Technical Reports Server (NTRS)
Wingard, Charles Doug; Munafo, Paul M. (Technical Monitor)
2002-01-01
Protein crystals are grown in microgravity experiments inside the Space Shuttle during orbit. Such crystals are basically grown in a five-component system containing a salt, buffer, polymer, organic and water. During these experiments, a number of different polymeric containment materials must be compatible with up to hundreds of different PCG solutions in various concentrations for durations up to 180 days. When such compatibility experiments are performed at NASA/MSFC (Marshall Space Flight Center) simultaneously on containment material samples immersed in various solutions in vials, the samples are rather small out of necessity. DMA4 modulus was often used as the primary screening parameter for such small samples as a pass/fail criterion for incompatibility issues. In particular, the TA Instruments DMA 2980 film tension clamp was used to test rubber O-rings as small in I.D. as 0.091 in. by cutting through the cross-section at one place, then clamping the stretched linear cord stock at each end. The film tension clamp was also used to successfully test short length samples of medical/surgical grade tubing with an O.D. of 0.125 in.
Hu, Jianhua; Wright, Fred A
2007-03-01
The identification of the genes that are differentially expressed in two-sample microarray experiments remains a difficult problem when the number of arrays is very small. We discuss the implications of using ordinary t-statistics and examine other commonly used variants. For oligonucleotide arrays with multiple probes per gene, we introduce a simple model relating the mean and variance of expression, possibly with gene-specific random effects. Parameter estimates from the model have natural shrinkage properties that guard against inappropriately small variance estimates, and the model is used to obtain a differential expression statistic. A limiting value to the positive false discovery rate (pFDR) for ordinary t-tests provides motivation for our use of the data structure to improve variance estimates. Our approach performs well compared to other proposed approaches in terms of the false discovery rate.
Taniguchi, Hidetaka; Sato, Hiroshi; Shirakawa, Tomohiro
2018-05-09
Human learners can generalize a new concept from a small number of samples. In contrast, conventional machine learning methods require large amounts of data to address the same types of problems. Humans have cognitive biases that promote fast learning. Here, we developed a method to reduce the gap between human beings and machines in this type of inference by utilizing cognitive biases. We implemented a human cognitive model into machine learning algorithms and compared their performance with the currently most popular methods, naïve Bayes, support vector machine, neural networks, logistic regression and random forests. We focused on the task of spam classification, which has been studied for a long time in the field of machine learning and often requires a large amount of data to obtain high accuracy. Our models achieved superior performance with small and biased samples in comparison with other representative machine learning methods.
The Preliminary Examination of Organics in the Returned Stardust Samples from Comet Wild 2
NASA Technical Reports Server (NTRS)
Sandford, S. A.; Aleon, J.; Alexander, C.; Butterworth, A.; Clemett, S. J.; Cody, G.; Cooper, G.; Dworkin, J. P.; Flynn, G. J.; Gilles, M. K.
2006-01-01
The primary objective of STARDUST is to collect coma samples from comet 8lP/Wild 2. These samples were collected by impact onto aerogel tiles on Jan 2, 2004 when the spacecraft flew through the comet's coma at a relative velocity of about 6.1 km/sec. Measurements of dust impacts on the front of the spacecraft suggest that the aerogel particle collector was impacted by 2800 +/- 500 particles larger than 15 micron in diameter. Following recovery of the Sample Return Capsule (SRC) on Jan 15, 2006, the aerogel collector trays will be removed in a clean room at JSC. After documentation of the collection, selected aerogel tiles will be removed and aerogel and cometary samples will be extracted for study. A number of different extraction techniques will be used, each optimized for the analytical technique that is to be used. The STARDUST Mission will carry out a 6 month preliminary examination (PE) of a small portion of the returned samples. The examination of the samples will be made by a number of subteams that will concentrate on specific aspects of the samples. One of these is the Organics PE Team (see the author list above for team members). These team members will use a number of analytical techniques to produce a preliminary characterization of the abundance and nature of the organics (if any) in the returned samples.
Evolutionary Trends and the Salience Bias (with Apologies to Oil Tankers, Karl Marx, and Others).
ERIC Educational Resources Information Center
McShea, Daniel W.
1994-01-01
Examines evolutionary trends, specifically trends in size, complexity, and fitness. Notes that documentation of these trends consists of either long lists of cases, or descriptions of a small number of salient cases. Proposes the use of random samples to avoid this "saliency bias." (SR)
ERIC Educational Resources Information Center
Begeny, John C.; Krouse, Hailey E.; Brown, Kristina G.; Mann, Courtney M.
2011-01-01
Teacher judgments about students' academic abilities are important for instructional decision making and potential special education entitlement decisions. However, the small number of studies evaluating teachers' judgments are limited methodologically (e.g., sample size, procedural sophistication) and have yet to answer important questions…
Detection of fragments from internal insects in wheat samples using a laboratory entoleter
USDA-ARS?s Scientific Manuscript database
A simple, rapid method was developed for estimating the number of insect fragments in flour caused by hidden, internal-feeding insects in whole grains during storage. The method uses a small mechanical rotary device (entoleter), which accelerates whole wheat kernels to high speeds and projects them...
Evidence of a sewer vapor transport pathway at the USEPA vapor intrusion research duplex
The role of sewer lines as preferential pathways for vapor intrusion is poorly understood. Although the importance of sewer lines for volatile organic compound (VOC) transport has been documented at a small number of sites with vapor intrusion, sewer lines are not routinely sampl...
Probability surveys of stream and river resources (hereafter referred to as streams) provide reliable estimates of stream condition when the areas for the estimates have sufficient number of sample sites. Monitoring programs are frequently asked to provide estimates for areas th...
Téllez-Sosa, Juan; Rodríguez, Mario Henry; Gómez-Barreto, Rosa E.; Valdovinos-Torres, Humberto; Hidalgo, Ana Cecilia; Cruz-Hervert, Pablo; Luna, René Santos; Carrillo-Valenzo, Erik; Ramos, Celso; García-García, Lourdes; Martínez-Barnetche, Jesús
2013-01-01
Background Influenza viruses display a high mutation rate and complex evolutionary patterns. Next-generation sequencing (NGS) has been widely used for qualitative and semi-quantitative assessment of genetic diversity in complex biological samples. The “deep sequencing” approach, enabled by the enormous throughput of current NGS platforms, allows the identification of rare genetic viral variants in targeted genetic regions, but is usually limited to a small number of samples. Methodology and Principal Findings We designed a proof-of-principle study to test whether redistributing sequencing throughput from a high depth-small sample number towards a low depth-large sample number approach is feasible and contributes to influenza epidemiological surveillance. Using 454-Roche sequencing, we sequenced at a rather low depth, a 307 bp amplicon of the neuraminidase gene of the Influenza A(H1N1) pandemic (A(H1N1)pdm) virus from cDNA amplicons pooled in 48 barcoded libraries obtained from nasal swab samples of infected patients (n = 299) taken from May to November, 2009 pandemic period in Mexico. This approach revealed that during the transition from the first (May-July) to second wave (September-November) of the pandemic, the initial genetic variants were replaced by the N248D mutation in the NA gene, and enabled the establishment of temporal and geographic associations with genetic diversity and the identification of mutations associated with oseltamivir resistance. Conclusions NGS sequencing of a short amplicon from the NA gene at low sequencing depth allowed genetic screening of a large number of samples, providing insights to viral genetic diversity dynamics and the identification of genetic variants associated with oseltamivir resistance. Further research is needed to explain the observed replacement of the genetic variants seen during the second wave. As sequencing throughput rises and library multiplexing and automation improves, we foresee that the approach presented here can be scaled up for global genetic surveillance of influenza and other infectious diseases. PMID:23843978
Numerical modelling of multimode fibre-optic communication lines
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sidelnikov, O S; Fedoruk, M P; Sygletos, S
The results of numerical modelling of nonlinear propagation of an optical signal in multimode fibres with a small differential group delay are presented. It is found that the dependence of the error vector magnitude (EVM) on the differential group delay can be reduced by increasing the number of ADC samples per symbol in the numerical implementation of the differential group delay compensation algorithm in the receiver. The possibility of using multimode fibres with a small differential group delay for data transmission in modern digital communication systems is demonstrated. It is shown that with increasing number of modes the strong couplingmore » regime provides a lower EVM level than the weak coupling one. (fibre-optic communication lines)« less
GRADE guidelines: 5. Rating the quality of evidence--publication bias.
Guyatt, Gordon H; Oxman, Andrew D; Montori, Victor; Vist, Gunn; Kunz, Regina; Brozek, Jan; Alonso-Coello, Pablo; Djulbegovic, Ben; Atkins, David; Falck-Ytter, Yngve; Williams, John W; Meerpohl, Joerg; Norris, Susan L; Akl, Elie A; Schünemann, Holger J
2011-12-01
In the GRADE approach, randomized trials start as high-quality evidence and observational studies as low-quality evidence, but both can be rated down if a body of evidence is associated with a high risk of publication bias. Even when individual studies included in best-evidence summaries have a low risk of bias, publication bias can result in substantial overestimates of effect. Authors should suspect publication bias when available evidence comes from a number of small studies, most of which have been commercially funded. A number of approaches based on examination of the pattern of data are available to help assess publication bias. The most popular of these is the funnel plot; all, however, have substantial limitations. Publication bias is likely frequent, and caution in the face of early results, particularly with small sample size and number of events, is warranted. Copyright © 2011 Elsevier Inc. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hardcastle, Nicholas; Bayliss, Adam; Wong, Jeannie Hsiu Ding
2012-08-15
Purpose: A recent field safety notice from TomoTherapy detailed the underdosing of small, off-axis targets when receiving high doses per fraction. This is due to angular undersampling in the dose calculation gantry angles. This study evaluates a correction method to reduce the underdosing, to be implemented in the current version (v4.1) of the TomoTherapy treatment planning software. Methods: The correction method, termed 'Super Sampling' involved the tripling of the number of gantry angles from which the dose is calculated during optimization and dose calculation. Radiochromic film was used to measure the dose to small targets at various off-axis distances receivingmore » a minimum of 21 Gy in one fraction. Measurements were also performed for single small targets at the center of the Lucy phantom, using radiochromic film and the dose magnifying glass (DMG). Results: Without super sampling, the peak dose deficit increased from 0% to 18% for a 10 mm target and 0% to 30% for a 5 mm target as off-axis target distances increased from 0 to 16.5 cm. When super sampling was turned on, the dose deficit trend was removed and all peak doses were within 5% of the planned dose. For measurements in the Lucy phantom at 9.7 cm off-axis, the positional and dose magnitude accuracy using super sampling was verified using radiochromic film and the DMG. Conclusions: A correction method implemented in the TomoTherapy treatment planning system which triples the angular sampling of the gantry angles used during optimization and dose calculation removes the underdosing for targets as small as 5 mm diameter, up to 16.5 cm off-axis receiving up to 21 Gy.« less
Wu, Baolin
2006-02-15
Differential gene expression detection and sample classification using microarray data have received much research interest recently. Owing to the large number of genes p and small number of samples n (p > n), microarray data analysis poses big challenges for statistical analysis. An obvious problem owing to the 'large p small n' is over-fitting. Just by chance, we are likely to find some non-differentially expressed genes that can classify the samples very well. The idea of shrinkage is to regularize the model parameters to reduce the effects of noise and produce reliable inferences. Shrinkage has been successfully applied in the microarray data analysis. The SAM statistics proposed by Tusher et al. and the 'nearest shrunken centroid' proposed by Tibshirani et al. are ad hoc shrinkage methods. Both methods are simple, intuitive and prove to be useful in empirical studies. Recently Wu proposed the penalized t/F-statistics with shrinkage by formally using the (1) penalized linear regression models for two-class microarray data, showing good performance. In this paper we systematically discussed the use of penalized regression models for analyzing microarray data. We generalize the two-class penalized t/F-statistics proposed by Wu to multi-class microarray data. We formally derive the ad hoc shrunken centroid used by Tibshirani et al. using the (1) penalized regression models. And we show that the penalized linear regression models provide a rigorous and unified statistical framework for sample classification and differential gene expression detection.
NASA Astrophysics Data System (ADS)
Shayanfar, Mohsen Ali; Barkhordari, Mohammad Ali; Roudak, Mohammad Amin
2017-06-01
Monte Carlo simulation (MCS) is a useful tool for computation of probability of failure in reliability analysis. However, the large number of required random samples makes it time-consuming. Response surface method (RSM) is another common method in reliability analysis. Although RSM is widely used for its simplicity, it cannot be trusted in highly nonlinear problems due to its linear nature. In this paper, a new efficient algorithm, employing the combination of importance sampling, as a class of MCS, and RSM is proposed. In the proposed algorithm, analysis starts with importance sampling concepts and using a represented two-step updating rule of design point. This part finishes after a small number of samples are generated. Then RSM starts to work using Bucher experimental design, with the last design point and a represented effective length as the center point and radius of Bucher's approach, respectively. Through illustrative numerical examples, simplicity and efficiency of the proposed algorithm and the effectiveness of the represented rules are shown.
Lo, Andy; Weiner, Joel H; Li, Liang
2013-09-17
Due to limited sample amounts, instrument time considerations, and reagent costs, only a small number of replicate experiments are typically performed for quantitative proteome analyses. Generation of reproducible data that can be readily assessed for consistency within a small number of datasets is critical for accurate quantification. We report our investigation of a strategy using reciprocal isotope labeling of two comparative samples as a tool for determining proteome changes. Reciprocal labeling was evaluated to determine the internal consistency of quantified proteome changes from Escherichia coli grown under aerobic and anaerobic conditions. Qualitatively, the peptide overlap between replicate analyses of the same sample and reverse labeled samples were found to be within 8%. Quantitatively, reciprocal analyses showed only a slight increase in average overall inconsistency when compared with replicate analyses (1.29 vs. 1.24-fold difference). Most importantly, reverse labeling was successfully used to identify spurious values resulting from incorrect peptide identifications and poor peak fitting. After removal of 5% of the peptide data with low reproducibility, a total of 275 differentially expressed proteins (>1.50-fold difference) were consistently identified and were then subjected to bioinformatics analysis. General considerations and guidelines for reciprocal labeling experimental design and biological significance of obtained results are discussed. Copyright © 2013 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Rotem, Asaf; Garraway, Levi; Su, Mei-Ju; Basu, Anindita; Regev, Aviv; Struhl, Kevin
2017-02-01
Three-dimensional growth conditions reflect the natural environment of cancer cells and are crucial to be performed at drug screens. We developed a 3D assay for cellular transformation that involves growth in low attachment (GILA) conditions and is strongly correlated with the 50-year old benchmark assay-soft agar. Using GILA, we performed high-throughput screens for drugs and genes that selectively inhibit or increase transformation, but not proliferation. This phenotypic approach is complementary to our genetic approach that utilizes single-cell RNA-sequencing of a patient sample to identify putative oncogenes that confer sensitivity to drugs designed to specifically inhibit the identified oncoprotein. Currently, we are dealing with a big challenge in our field- the limited number of cells that might be extracted from a biopsy. Small patient-derived samples are hard to test in the traditional multiwell plate and it will be helpful to minimize the culture area and the experimental system. We managed to design a suitable microfluidic device for limited number of cells and perform the assay using image analysis. We aim to test drugs on tumor cells, outside of the patient body- and recommend on the ideal treatment that is tailored to the individual. This device will help to minimize biopsy-sampling volumes and minimize interventions in the patient's tumor.
System for sampling and monitoring microscopic organisms and substances
Au, Frederick H. F.; Beckert, Werner F.
1976-01-01
A technique and apparatus used therewith for determining the uptake of plutonium and other contaminants by soil microorganisms which, in turn, gives a measure of the plutonium and/or other contaminants available to the biosphere at that particular time. A measured quantity of uncontaminated spores of a selected mold is added to a moistened sample of the soil to be tested. The mixture is allowed to sit a predetermined number of days under specified temperature conditions. An agar layer is then applied to the top of the sample. After three or more days, when spores of the mold growing in the sample have formed, the spores are collected by a miniature vacuum collection apparatus operated under preselected vacuum conditions, which collect only the spores with essentially no contamination by mycelial fragments or culture medium. After collection, the fungal spores are dried and analyzed for the plutonium and/or other contaminants. The apparatus is also suitable for collection of pollen, small insects, dust and other small particles, material from thin-layer chromatography plates, etc.
Integrated Blood Barcode Chips
Fan, Rong; Vermesh, Ophir; Srivastava, Alok; Yen, Brian K.H.; Qin, Lidong; Ahmad, Habib; Kwong, Gabriel A.; Liu, Chao-Chao; Gould, Juliane; Hood, Leroy; Heath, James R.
2008-01-01
Blood comprises the largest version of the human proteome1. Changes of plasma protein profiles can reflect physiological or pathological conditions associated with many human diseases, making blood the most important fluid for clinical diagnostics2-4. Nevertheless, only a handful of plasma proteins are utilized in routine clinical tests. This is due to a host of reasons, including the intrinsic complexity of the plasma proteome1, the heterogeneity of human diseases and the fast kinetics associated with protein degradation in sampled blood5. Simple technologies that can sensitively sample large numbers of proteins over broad concentration ranges, from small amounts of blood, and within minutes of sample collection, would assist in solving these problems. Herein, we report on an integrated microfluidic system, called the Integrated Blood Barcode Chip (IBBC). It enables on-chip blood separation and the rapid measurement of a panel of plasma proteins from small quantities of blood samples including a fingerprick of whole blood. This platform holds potential for inexpensive, non-invasive, and informative clinical diagnoses, particularly, for point-of-care. PMID:19029914
Small D-type asteroids in the NEO population: new targets for space missions
NASA Astrophysics Data System (ADS)
Barucci, Maria Antonietta; Perna, D.; Popescu, M.; Fornasier, S.; Doressoundiram, A.; Lantz, C.; Merlin, F.; Fulchignoni, M.; Dotto, E.; Kanuchova, S.
2018-06-01
In the framework of the Near Earth Objects (NEOs) observational campaign carried out within the NEOShield-2 project, we identify nine new small D-type asteroids with estimated diameter less than 600 m. The link with meteorites for this class of asteroids is weak and the best fit obtained is with the Tagish Lake meteorite for seven of them. D-type asteroids are believed to contain the most pristine material of the Solar system and could have delivered the pre-biotic material to the Earth. Our results double the known sample of the D-types in the NEO population and triple the candidates of this class for a sample-return mission (at very low ΔV). Our finding increases considerably the number of targets for sample-return mission. A sample-return mission to a D-type asteroid will provide a major progress in understanding the early history of the Solar system and to investigate the origin of life on the Earth.
A new small-angle X-ray scattering set-up on the crystallography beamline I711 at MAX-lab.
Knaapila, M; Svensson, C; Barauskas, J; Zackrisson, M; Nielsen, S S; Toft, K N; Vestergaard, B; Arleth, L; Olsson, U; Pedersen, J S; Cerenius, Y
2009-07-01
A small-angle X-ray scattering (SAXS) set-up has recently been developed at beamline I711 at the MAX II storage ring in Lund (Sweden). An overview of the required modifications is presented here together with a number of application examples. The accessible q range in a SAXS experiment is 0.009-0.3 A(-1) for the standard set-up but depends on the sample-to-detector distance, detector offset, beamstop size and wavelength. The SAXS camera has been designed to have a low background and has three collinear slit sets for collimating the incident beam. The standard beam size is about 0.37 mm x 0.37 mm (full width at half-maximum) at the sample position, with a flux of 4 x 10(10) photons s(-1) and lambda = 1.1 A. The vacuum is of the order of 0.05 mbar in the unbroken beam path from the first slits until the exit window in front of the detector. A large sample chamber with a number of lead-throughs allows different sample environments to be mounted. This station is used for measurements on weakly scattering proteins in solutions and also for colloids, polymers and other nanoscale structures. A special application supported by the beamline is the effort to establish a micro-fluidic sample environment for structural analysis of samples that are only available in limited quantities. Overall, this work demonstrates how a cost-effective SAXS station can be constructed on a multipurpose beamline.
De Jong, G D; Hoback, W W
2006-06-01
Carrion insect succession studies have historically used repeated sampling of single or a few carcasses to produce data, either weighing the carcasses, removing a qualitative subsample of the fauna present, or both, on every visit over the course of decomposition and succession. This study, conducted in a set of related experimental hypotheses with two trials in a single season, investigated the effect that repeated sampling has on insect succession, determined by the number of taxa collected on each visit and by community composition. Each trial lasted at least 21 days, with daily visits on the first 14 days. Rat carcasses used in this study were all placed in the field on the same day, but then either sampled qualitatively on every visit (similar to most succession studies) or ignored until a given day of succession, when they were sampled qualitatively (a subsample) and then destructively sampled in their entirety. Carcasses sampled on every visit were in two groups: those from which only a sample of the fauna was taken and those from which a sample of fauna was taken and the carcass was weighed for biomass determination. Of the carcasses visited only once, the number of taxa in subsamples was compared to the actual number of taxa present when the carcass was destructively sampled to determine if the subsamples adequately represented the total carcass fauna. Data from the qualitative subsamples of those carcasses visited only once were also compared to data collected from carcasses that were sampled on every visit to investigate the effect of the repeated sampling. A total of 39 taxa were collected from carcasses during the study and the component taxa are discussed individually in relation to their role in succession. Number of taxa differed on only one visit between the qualitative subsamples and the actual number of taxa present, primarily because the organisms missed by the qualitative sampling were cryptic (hidden deep within body cavities) or rare (only represented by very few specimens). There were no differences discovered between number of taxa in qualitative subsamples from carcasses sampled repeatedly (with or without biomass determinations) and those sampled only a single time. Community composition differed considerably in later stages of decomposition, with disparate communities due primarily to small numbers of rare taxa. These results indicate that the methods used historically for community composition determination in experimental forensic entomology are generally adequate.
Bremsstrahlung-Based Imaging and Assays of Radioactive, Mixed and Hazardous Waste
NASA Astrophysics Data System (ADS)
Kwofie, J.; Wells, D. P.; Selim, F. A.; Harmon, F.; Duttagupta, S. P.; Jones, J. L.; White, T.; Roney, T.
2003-08-01
A new nondestructive accelerator based x-ray fluorescence (AXRF) approach has been developed to identify heavy metals in large-volume samples. Such samples are an important part of the process and waste streams of U.S Department of Energy sites, as well as other industries such as mining and milling. Distributions of heavy metal impurities in these process and waste samples can range from homogeneous to highly inhomogeneous, and non-destructive assays and imaging that can address both are urgently needed. Our approach is based on using high-energy, pulsed bremsstrahlung beams (3-6.5 MeV) from small electron accelerators to produce K-shell atomic fluorescence x-rays. In addition we exploit pair-production, Compton scattering and x-ray transmission measurements from these beams to probe locations of high density and high atomic number. The excellent penetrability of these beams allows assays and images for soil-like samples at least 15 g/cm2 thick, with elemental impurities of atomic number greater than approximately 50. Fluorescence yield of a variety of targets was measured as a function of impurity atomic number, impurity homogeneity, and sample thickness. We report on actual and potential detection limits of heavy metal impurities in a soil matrix for a variety of samples, and on the potential for imaging, using AXRF and these related probes.
Verspoor, E; Knox, D; Marshall, S
2016-12-01
An eclectic set of tissues and existing data, including purposely collected samples, spanning 1997-2006, was used in an ad hoc assessment of hybridization and introgression of farmed wild Atlantic salmon Salmo salar in the small Loch na Thull (LnT) catchment in north-west Scotland. The catchment is in an area of marine farm production and contains freshwater smolt rearing cages. The LnT S. salar stock was found to be genetically distinctive from stocks in neighbouring rivers and, despite regular reports of feral farm S. salar, there was no evidence of physical or genetic mixing. This cannot be completely ruled out, however, and low level mixing with other local wild stocks has been suggested. The LnT population appeared underpinned by relatively smaller effective number of breeders (N eb ) and showed relatively low levels of genetic diversity, consistent with a small effective population size. Small sample sizes, an incomplete farm baseline and the use of non-diagnostic molecular markers, constrain the power of the analysis but the findings strongly support the LnT catchment having a genetically distinct wild S. salar population little affected by interbreeding with feral farm escapes. © 2016 The Fisheries Society of the British Isles.
Ruiz-Toledo, Jovani; Vandame, Rémy; Castro-Chan, Ricardo Alberto; Penilla-Navarro, Rosa Patricia; Gómez, Jaime; Sánchez, Daniel
2018-05-10
In this paper, we show the results of investigating the presence of organochlorine pesticides in honey and pollen samples from managed colonies of the honey bee, Apis mellifera L. and of the stingless bee Scaptotrigona mexicana Guérin. Three colonies of each species were moved into each of two sites. Three samples of pollen and three samples of honey were collected from each colony: the first collection occurred at the beginning of the study and the following ones at every six months during a year. Thus the total number of samples collected was 36 for honey (18 for A. mellifera and 18 for S. mexicana ) and 36 for pollen (18 for A. mellifera and 18 for S. mexicana ). We found that 88.44% and 93.33% of honey samples, and 22.22% and 100% of pollen samples of S. mexicana and A. mellifera , respectively, resulted positive to at least one organochlorine. The most abundant pesticides were Heptaclor (44% of the samples), γ-HCH (36%), DDT (19%), Endrin (18%) and DDE (11%). Despite the short foraging range of S. mexicana , the number of pesticides quantified in the honey samples was similar to that of A. mellifera . Paradoxically we found a small number of organochlorines in pollen samples of S. mexicana in comparison to A. mellifera , perhaps indicating a low abundance of pollen sources within the foraging range of this species.
The topology of large-scale structure. III - Analysis of observations
NASA Astrophysics Data System (ADS)
Gott, J. Richard, III; Miller, John; Thuan, Trinh X.; Schneider, Stephen E.; Weinberg, David H.; Gammie, Charles; Polk, Kevin; Vogeley, Michael; Jeffrey, Scott; Bhavsar, Suketu P.; Melott, Adrian L.; Giovanelli, Riccardo; Hayes, Martha P.; Tully, R. Brent; Hamilton, Andrew J. S.
1989-05-01
A recently developed algorithm for quantitatively measuring the topology of large-scale structures in the universe was applied to a number of important observational data sets. The data sets included an Abell (1958) cluster sample out to Vmax = 22,600 km/sec, the Giovanelli and Haynes (1985) sample out to Vmax = 11,800 km/sec, the CfA sample out to Vmax = 5000 km/sec, the Thuan and Schneider (1988) dwarf sample out to Vmax = 3000 km/sec, and the Tully (1987) sample out to Vmax = 3000 km/sec. It was found that, when the topology is studied on smoothing scales significantly larger than the correlation length (i.e., smoothing length, lambda, not below 1200 km/sec), the topology is spongelike and is consistent with the standard model in which the structure seen today has grown from small fluctuations caused by random noise in the early universe. When the topology is studied on the scale of lambda of about 600 km/sec, a small shift is observed in the genus curve in the direction of a 'meatball' topology.
The topology of large-scale structure. III - Analysis of observations. [in universe
NASA Technical Reports Server (NTRS)
Gott, J. Richard, III; Weinberg, David H.; Miller, John; Thuan, Trinh X.; Schneider, Stephen E.
1989-01-01
A recently developed algorithm for quantitatively measuring the topology of large-scale structures in the universe was applied to a number of important observational data sets. The data sets included an Abell (1958) cluster sample out to Vmax = 22,600 km/sec, the Giovanelli and Haynes (1985) sample out to Vmax = 11,800 km/sec, the CfA sample out to Vmax = 5000 km/sec, the Thuan and Schneider (1988) dwarf sample out to Vmax = 3000 km/sec, and the Tully (1987) sample out to Vmax = 3000 km/sec. It was found that, when the topology is studied on smoothing scales significantly larger than the correlation length (i.e., smoothing length, lambda, not below 1200 km/sec), the topology is spongelike and is consistent with the standard model in which the structure seen today has grown from small fluctuations caused by random noise in the early universe. When the topology is studied on the scale of lambda of about 600 km/sec, a small shift is observed in the genus curve in the direction of a 'meatball' topology.
Brand-specific consumption of alcohol among underage youth in the United States.
Siegel, Michael; DeJong, William; Naimi, Timothy S; Fortunato, Erin K; Albers, Alison B; Heeren, Timothy; Rosenbloom, David L; Ross, Craig; Ostroff, Joshua; Rodkin, Sergei; King, Charles; Borzekowski, Dina L G; Rimal, Rajiv N; Padon, Alisa A; Eck, Raimee H; Jernigan, David H
2013-07-01
Little is known about brand-specific alcohol consumption among underage youth, as existing information is collected at the level of alcoholic beverage type. This study identifies the alcohol brands consumed by a nationally representative sample of underage youth in the United States. We obtained a national sample of 1,032 underage youth, aged 13 to 20, using a pre-recruited Internet panel maintained by Knowledge Networks. Youth aged 18 to 20 were recruited directly from the panel via email invitation. Teens aged 13 to 17 were identified by asking adult panelists to identify a member of their household. The survey assessed the past 30-day consumption of 898 brands of alcohol among 16 alcoholic beverage types, including the frequency and amount of each brand consumed in the past 30 days. Market share for a given brand was calculated by dividing the total number of drinks for that brand in the past 30 days across the entire sample by the total number of drinks for all identified brands. The alcohol brands with highest prevalence of past 30-day consumption were Bud Light (27.9%, 95% confidence interval [CI] 23.3 to 32.4%), Smirnoff malt beverages (17.0%, 95% CI 12.9 to 21.1%), and Budweiser (14.6%, 95% CI 11.0 to 18.3%). Brand market share was concentrated in a relatively small number of brands, with the top 25 brands accounting for nearly half of all market shares. Underage youth alcohol consumption, although spread out over several alcoholic beverage types, is concentrated among a relatively small number of alcohol brands. This finding has important implications for alcohol research, practice, and policy. Copyright © 2013 by the Research Society on Alcoholism.
Brand-Specific Consumption of Alcohol among Underage Youth in the United States
Siegel, Michael; DeJong, William; Naimi, Timothy S.; Fortunato, Erin K.; Albers, Alison B.; Heeren, Timothy; Rosenbloom, David L.; Ross, Craig; Ostroff, Joshua; Rodkin, Sergei; King, Charles; Borzekowski, Dina L.G.; Rimal, Rajiv N.; Padon, Alisa A.; Eck, Raimee H.; Jernigan, David H.
2013-01-01
Background Little is known about brand-specific alcohol consumption among underage youth, as existing information is collected at the level of alcoholic beverage type. This study identifies the alcohol brands consumed by a nationally representative sample of underage youth in the U.S. Methods We obtained a national sample of 1,032 underage youth, ages 13–20, using a pre-recruited internet panel maintained by Knowledge Networks. Youth ages 18–20 were recruited directly from the panel via email invitation. Teens ages 13–17 were identified by asking adult panelists to identify a member of their household. The survey assessed the past 30-day consumption of 898 brands of alcohol among 16 alcoholic beverage types, including the frequency and amount of each brand consumed in the past 30 days. Market share for a given brand was calculated by dividing the total number of drinks for that brand in the past 30 days across the entire sample by the total number of drinks for all identified brands. Results The alcohol brands with highest prevalence of past 30-day consumption were Bud Light (27.9%, 95% confidence interval [CI] 23.3%–32.4%), Smirnoff malt beverages (17.0%, 95% CI 12.9%–21.1%), and Budweiser (14.6%, 95% CI 11.0%–18.3%). Brand market share was concentrated in a relatively small number of brands, with the top 25 brands accounting for nearly half of all market share. Conclusions Underage youth alcohol consumption, although spread out over several alcoholic beverage types, is concentrated among a relatively small number of alcohol brands. This finding has important implications for alcohol research, practice, and policy. PMID:23398328
Herrick, Robert F; Stewart, James H; Allen, Joseph G
2016-02-01
PCBs in building materials such as caulks and sealants are a largely unrecognized source of contamination in the building environment. Schools are of particular interest, as the period of extensive school construction (about 1950 to 1980) coincides with the time of greatest use of PCBs as plasticizers in building materials. In the USA, we estimate that the number of schools with PCB in building caulk ranges from 12,960 to 25,920 based upon the number of schools built in the time of PCB use and the proportion of buildings found to contain PCB caulk and sealants. Field and laboratory studies have demonstrated that PCBs from both interior and exterior caulking can be the source of elevated PCB air concentrations in these buildings, at levels that exceed health-based PCB exposure guidelines for building occupants. Air sampling in buildings containing PCB caulk has shown that the airborne PCB concentrations can be highly variable, even in repeat samples collected within a room. Sampling and data analysis strategies that recognize this variability can provide the basis for informed decision making about compliance with health-based exposure limits, even in cases where small numbers of samples are taken. The health risks posed by PCB exposures, particularly among children, mandate precautionary approaches to managing PCBs in building materials.
Cryopreservation of Circulating Tumor Cells for Enumeration and Characterization.
Nejlund, Sarah; Smith, Julie; Kraan, Jaco; Stender, Henrik; Van, Mai N; Langkjer, Sven T; Nielsen, Mikkel T; Sölétormos, György; Hillig, Thore
2016-08-01
A blood sample containing circulating tumor cells (CTCs) may serve as a surrogate for metastasis in invasive cancer. Cryopreservation will provide new opportunities in management of clinical samples in the laboratory and allow collection of samples over time for future analysis of existing and upcoming cancer biomarkers. Blood samples from healthy volunteers were spiked with high (∼500) and low (∼50) number of tumor cells from culture. The samples were stored at -80C with cryopreservative dimethyl sulfoxide mixed with Roswell Park Memorial Institute 1640 medium. Flow cytometry tested if cryopreservation affected specific biomarkers regularly used to detect CTCs, i.e. cytokeratin (CK) and epithelial cell adhesion molecule (EpCAM) and white blood cell specific lymphocyte common antigen (CD45). After various time intervals (up to 6 months), samples were thawed and tumor cell recovery (enumeration) was examined. Clinical samples may differ from cell line studies, so the cryopreservation protocol was tested on 17 patients with invasive breast cancer and tumor cell recovery was examined. Two blood samples were drawn from each patient. Biomarkers, CK, CD45, and EpCAM, were not affected by the freezing and thawing procedures. Cryopreserved samples (n = 2) spiked with a high number of tumor cells (∼500) had a ∼90% recovery compared with the spiked fresh samples. In samples spiked with lower numbers of tumor cells (median = 43 in n = 5 samples), the recovery was 63% after cryopreservation (median 27 tumor cells), p = 0.03. With an even lower number of spiked tumor cells (median = 3 in n = 8 samples), the recovery rate of tumor cells after cryopreservation did not seem to be affected (median = 8), p = 0.09. Time of cryopreservation did not affect recovery. When testing the effect of cryopreservation on enumeration in clinical samples, no difference was observed in the number of CTCs between the fresh and the cryopreserved samples based on n = 17 pairs, p = 0.83; however, the variation was large. This large variation was confirmed by clinically paired fresh samples (n = 64 pairs), where 95% of the samples (<30 CTCs) vary in number up to ±15 CTCs, p = 0.18. A small loss of CTCs after cryopreservation may be expected; however, cryopreservation of CTCs for biomarker characterization for clinical applications seems promising.
The Groove of Growth: How Early Gains in Math Ability Influence Adolescent Achievement
ERIC Educational Resources Information Center
Watts, Tyler W.; Duncan, Greg J.; Siegler, Robert S.; Davis-Kean, Pamela E.
2014-01-01
A number of studies, both small scale and of nationally-representative student samples, have reported substantial associations between school entry math ability and later elementary school achievement. However, questions remain regarding the persistence of the association between early growth in math ability and later math achievement due to the…
Dysgenic Fertility, Intelligence and Family Size in Libya
ERIC Educational Resources Information Center
Al-Shahomee, Alsedig Abdalgadr; Lynn, Richard; Abdalla, Saleh El-ghmary
2013-01-01
The Standard Progressive Matrices (SPM) was administered to a sample of 592 16 year old school students in Libya. There was a small negative correlation of -0.14 between SPM scores and the number of siblings, indicating only marginal dysgenic fertility. Supplementary material giving the data is given online. (Contains 1 table.)
Mossbauer analysis of Luna 16 lunar surface material
NASA Technical Reports Server (NTRS)
Nady, D. L.; Cher, L.; Kulcsar, K.
1974-01-01
Samples of Apollo 11 lunar surface material were studied by the Mossbauer effect. Owing to the small number of other resonant isotopes, all measurements were made with Fe-57 nuclei. The principal constituents of the material were as follows: Iron containing silicates (olivine, pyroxene, and so on), ilmenite (FeTiO3), and metallic iron.
Stanwix House Vocabulary Study: Plurals. Investigative Report Number 7227.
ERIC Educational Resources Information Center
Wood, Penelope; Bond, Jack H.
Research analyzed: 1) the treatment of plurals in the Functional Basic Word List for Special Pupils (Stanwix House List), and 2) plural recognition by educable mentally handicapped (EMH) students. Review of the Stanwix House List revealed that plurals were treated as discrete vocabulary items. Using a small sample of EMH subjects, the researchers…
NASA Astrophysics Data System (ADS)
Karatay, Ahmet; Küçüköz, Betül; Pekdemir, Sami; Onses, Mustafa Serdar; Elmali, Ayhan
2017-11-01
We have investigated the plasmonic properties of heterostructures that consist of gold nanosphere (NSs) with average diameters of 60 nm, 40 nm and 20 nm on poly (ethylene glycol) (PEG) brushes by using ultrafast pump-probe spectroscopy experiments. Gold NSs start to behave like gold nanorods with increasing number of immobilization cycles due to the close proximity. Gold NSs immobilized by 3 and 5 deposition cycles show longitudinal modes of plasmon bands at long wavelengths which are characteristic behaviors for gold nanorods. Increasing the number of immobilization cycle also increase relaxation times of samples due to the close proximity. Linear absorption spectra and scanning electron microscopy images show that there are close packing assemblies for heterostructures containing 20 nm gold NSs as the small particle. Ultrafast electron transfer (<100 fs) occurs between transverse and longitudinal modes by exciting the samples at both 520 nm and 650 nm. Further, experimental results indicate that, heterostructures with the small particles have faster relaxation times than other heterostructures due to closed packing of 20 nm gold NSs.
Performance evaluation of DNA copy number segmentation methods.
Pierre-Jean, Morgane; Rigaill, Guillem; Neuvial, Pierre
2015-07-01
A number of bioinformatic or biostatistical methods are available for analyzing DNA copy number profiles measured from microarray or sequencing technologies. In the absence of rich enough gold standard data sets, the performance of these methods is generally assessed using unrealistic simulation studies, or based on small real data analyses. To make an objective and reproducible performance assessment, we have designed and implemented a framework to generate realistic DNA copy number profiles of cancer samples with known truth. These profiles are generated by resampling publicly available SNP microarray data from genomic regions with known copy-number state. The original data have been extracted from dilutions series of tumor cell lines with matched blood samples at several concentrations. Therefore, the signal-to-noise ratio of the generated profiles can be controlled through the (known) percentage of tumor cells in the sample. This article describes this framework and its application to a comparison study between methods for segmenting DNA copy number profiles from SNP microarrays. This study indicates that no single method is uniformly better than all others. It also helps identifying pros and cons of the compared methods as a function of biologically informative parameters, such as the fraction of tumor cells in the sample and the proportion of heterozygous markers. This comparison study may be reproduced using the open source and cross-platform R package jointseg, which implements the proposed data generation and evaluation framework: http://r-forge.r-project.org/R/?group_id=1562. © The Author 2014. Published by Oxford University Press.
Chen, Qixuan; Li, Jingguang
2014-05-01
Many recent studies have examined the association between number acuity, which is the ability to rapidly and non-symbolically estimate the quantity of items appearing in a scene, and symbolic math performance. However, various contradictory results have been reported. To comprehensively evaluate the association between number acuity and symbolic math performance, we conduct a meta-analysis to synthesize the results observed in previous studies. First, a meta-analysis of cross-sectional studies (36 samples, N = 4705) revealed a significant positive correlation between these skills (r = 0.20, 95% CI = [0.14, 0.26]); the association remained after considering other potential moderators (e.g., whether general cognitive abilities were controlled). Moreover, a meta-analysis of longitudinal studies revealed 1) that number acuity may prospectively predict later math performance (r = 0.24, 95% CI = [0.11, 0.37]; 6 samples) and 2) that number acuity is retrospectively correlated to early math performance as well (r = 0.17, 95% CI = [0.07, 0.26]; 5 samples). In summary, these pieces of evidence demonstrate a moderate but statistically significant association between number acuity and math performance. Based on the estimated effect sizes, power analyses were conducted, which suggested that many previous studies were underpowered due to small sample sizes. This may account for the disparity between findings in the literature, at least in part. Finally, the theoretical and practical implications of our meta-analytic findings are presented, and future research questions are discussed. Copyright © 2014 Elsevier B.V. All rights reserved.
How many stakes are required to measure the mass balance of a glacier?
Fountain, A.G.; Vecchia, A.
1999-01-01
Glacier mass balance is estimated for South Cascade Glacier and Maclure Glacier using a one-dimensional regression of mass balance with altitude as an alternative to the traditional approach of contouring mass balance values. One attractive feature of regression is that it can be applied to sparse data sets where contouring is not possible and can provide an objective error of the resulting estimate. Regression methods yielded mass balance values equivalent to contouring methods. The effect of the number of mass balance measurements on the final value for the glacier showed that sample sizes as small as five stakes provided reasonable estimates, although the error estimates were greater than for larger sample sizes. Different spatial patterns of measurement locations showed no appreciable influence on the final value as long as different surface altitudes were intermittently sampled over the altitude range of the glacier. Two different regression equations were examined, a quadratic, and a piecewise linear spline, and comparison of results showed little sensitivity to the type of equation. These results point to the dominant effect of the gradient of mass balance with altitude of alpine glaciers compared to transverse variations. The number of mass balance measurements required to determine the glacier balance appears to be scale invariant for small glaciers and five to ten stakes are sufficient.
Sizing for the apparel industry using statistical analysis - a Brazilian case study
NASA Astrophysics Data System (ADS)
Capelassi, C. H.; Carvalho, M. A.; El Kattel, C.; Xu, B.
2017-10-01
The study of the body measurements of Brazilian women used the Kinect Body Imaging system for 3D body scanning. The result of the study aims to meet the needs of the apparel industry for accurate measurements. Data was statistically treated using the IBM SPSS 23 system, with 95% confidence (P<0,05) for the inferential analysis, with the purpose of grouping the measurements in sizes, so that a smaller number of sizes can cover a greater number of people. The sample consisted of 101 volunteers aged between 19 and 62 years. A cluster analysis was performed to identify the main body shapes of the sample. The results were divided between the top and bottom body portions; For the top portion, were used the measurements of the abdomen, waist and bust circumferences, as well as the height; For the bottom portion, were used the measurements of the hip circumference and the height. Three sizing systems were developed for the researched sample from the Abdomen-to-Height Ratio - AHR (top portion): Small (AHR < 0,52), Medium (AHR: 0,52-0,58), Large (AHR > 0,58) and from the Hip-to-Height Ratio - HHR (bottom portion): Small (HHR < 0,62), Medium (HHR: 0,62-0,68), Large (HHR > 0,68).
Anderson, Eric C; Ng, Thomas C
2016-02-01
We develop a computational framework for addressing pedigree inference problems using small numbers (80-400) of single nucleotide polymorphisms (SNPs). Our approach relaxes the assumptions, which are commonly made, that sampling is complete with respect to the pedigree and that there is no genotyping error. It relies on representing the inferred pedigree as a factor graph and invoking the Sum-Product algorithm to compute and store quantities that allow the joint probability of the data to be rapidly computed under a large class of rearrangements of the pedigree structure. This allows efficient MCMC sampling over the space of pedigrees, and, hence, Bayesian inference of pedigree structure. In this paper we restrict ourselves to inference of pedigrees without loops using SNPs assumed to be unlinked. We present the methodology in general for multigenerational inference, and we illustrate the method by applying it to the inference of full sibling groups in a large sample (n=1157) of Chinook salmon typed at 95 SNPs. The results show that our method provides a better point estimate and estimate of uncertainty than the currently best-available maximum-likelihood sibling reconstruction method. Extensions of this work to more complex scenarios are briefly discussed. Published by Elsevier Inc.
Naugle, Alecia Larew; Barlow, Kristina E; Eblen, Denise R; Teter, Vanessa; Umholtz, Robert
2006-11-01
The U.S. Food Safety and Inspection Service (FSIS) tests sets of samples of selected raw meat and poultry products for Salmonella to ensure that federally inspected establishments meet performance standards defined in the pathogen reduction-hazard analysis and critical control point system (PR-HACCP) final rule. In the present report, sample set results are described and associations between set failure and set and establishment characteristics are identified for 4,607 sample sets collected from 1998 through 2003. Sample sets were obtained from seven product classes: broiler chicken carcasses (n = 1,010), cow and bull carcasses (n = 240), market hog carcasses (n = 560), steer and heifer carcasses (n = 123), ground beef (n = 2,527), ground chicken (n = 31), and ground turkey (n = 116). Of these 4,607 sample sets, 92% (4,255) were collected as part of random testing efforts (A sets), and 93% (4,166) passed. However, the percentage of positive samples relative to the maximum number of positive results allowable in a set increased over time for broilers but decreased or stayed the same for the other product classes. Three factors associated with set failure were identified: establishment size, product class, and year. Set failures were more likely early in the testing program (relative to 2003). Small and very small establishments were more likely to fail than large ones. Set failure was less likely in ground beef than in other product classes. Despite an overall decline in set failures through 2003, these results highlight the need for continued vigilance to reduce Salmonella contamination in broiler chicken and continued implementation of programs designed to assist small and very small establishments with PR-HACCP compliance issues.
Tan, Ling; Hu, Yerong; Tao, Yongguang; Wang, Bin; Xiao, Jun; Tang, Zhenjie; Lu, Ting
2018-01-01
Background To identify whether RET is a potential target for NSCLC treatment, we examined the status of the RET gene in 631 early and mid stage NSCLC cases from south central China. Methods RET expression was identified by Western blot. RET‐positive expression samples were verified by immunohistochemistry. RET gene mutation, copy number variation, and rearrangement were analyzed by DNA Sanger sequencing, TaqMan copy number assays, and reverse transcription‐PCR. ALK and ROS1 expression levels were tested by Western blot and EGFR mutation using Sanger sequencing. Results The RET‐positive rate was 2.5% (16/631). RET‐positive expression was related to poorer tumor differentiation (P < 0.05). In the 16 RET‐positive samples, only two samples of moderately and poorly differentiated lung adenocarcinomas displayed RET rearrangement, both in RET‐KIF5B fusion partners. Neither ALK nor ROS1 translocation was found. The EGFR mutation rate in RET‐positive samples was significantly lower than in RET‐negative samples (P < 0.05). Conclusion RET‐positive expression in early and mid stage NSCLC cases from south central China is relatively low and is related to poorer tumor differentiation. RET gene alterations (copy number gain and rearrangement) exist in all RET‐positive samples. RET‐positive expression is a relatively independent factor in NSCLC patients, which indicates that the RET gene may be a novel target site for personalized treatment of NSCLC. PMID:29473341
Crabtree, Nathaniel M; Moore, Jason H; Bowyer, John F; George, Nysia I
2017-01-01
A computational evolution system (CES) is a knowledge discovery engine that can identify subtle, synergistic relationships in large datasets. Pareto optimization allows CESs to balance accuracy with model complexity when evolving classifiers. Using Pareto optimization, a CES is able to identify a very small number of features while maintaining high classification accuracy. A CES can be designed for various types of data, and the user can exploit expert knowledge about the classification problem in order to improve discrimination between classes. These characteristics give CES an advantage over other classification and feature selection algorithms, particularly when the goal is to identify a small number of highly relevant, non-redundant biomarkers. Previously, CESs have been developed only for binary class datasets. In this study, we developed a multi-class CES. The multi-class CES was compared to three common feature selection and classification algorithms: support vector machine (SVM), random k-nearest neighbor (RKNN), and random forest (RF). The algorithms were evaluated on three distinct multi-class RNA sequencing datasets. The comparison criteria were run-time, classification accuracy, number of selected features, and stability of selected feature set (as measured by the Tanimoto distance). The performance of each algorithm was data-dependent. CES performed best on the dataset with the smallest sample size, indicating that CES has a unique advantage since the accuracy of most classification methods suffer when sample size is small. The multi-class extension of CES increases the appeal of its application to complex, multi-class datasets in order to identify important biomarkers and features.
Photometric microlensing and stellar mass determination
NASA Astrophysics Data System (ADS)
Samadi, R.
Microlensing was suggested for stellar mass determination of nearby stars by several authors (e.g. Paczynski 1995 and Miralada-Escude 1996). There are two aspects in gravitational microlensing: photometry and astrometry. Here only the photometric aspect -i.e. magnification of a background source by a stellar lens- will be considered. The first study in this domain was done by Paczyński (1995). An attempt to investigate some observational constraints (Alard et al. 1996), takes its origin in this study. It will be shown here that blending of the source by the lens, not only induces a degeneracy which has been pointed out by Wozniak and Paczynski (1997) and can be removed by measuring the flux of the source outside the microlensing event, but also strongly reduces the photometric cross section as well as the duration of microlensing events. The expected number of events decreases strongly with blending. Blending effects can be reduced by selecting faint lens candidates. Unfortunately however, it is difficult to monitor a sufficient number of lens candidates per night, and restricting oneself to nearby and high proper motion objects would lead to a very small number of lens candidates. In the case of short duration events, the light curve must be sampled at short time intervals. When sampling daily, the condition is very strong and the expected rate of events very small. Observing from space would increase significantly the rate, but this gain is still insufficient. We conclude that the event rate will be very small and that the project does not seem to be feasible at least under current observing conditions and even in the near future.
Frenzel, Steven A.; Dorava, Joseph M.
1999-01-01
Five streams in the Cook Inlet Basin, Alaska, were sampled in 1998 to provide the National Park Service with baseline information on water quality. Four of these streams drain National Park Service land: Costello and Colorado Creeks in Denali National Park and Preserve, Johnson River in Lake Clark National Park and Preserve, and Kamishak River in Katmai National Park and Preserve. The fifth site was on the Talkeetna River, outside of national park boundaries. Samples of stream water, streambed sediments, and fish tissues were collected for chemical analyses. Biological and geomorphic information was also collected at each site. Nutrient concentrations in stream water were low and commonly were less than analytical detection limits. Analyses of fish tissues for 28 organochlorine compounds at Talkeetna River and Costello Creek produced just one detection. Hexachlorobenzene was detected at a concentration of 5.70 micrograms per kilogram in slimy sculpin from the Talkeetna River. Streambed sediment samples from the Talkeetna River had three organochlorine compounds at detectable levels; hexachlorobenzene was measured at 13 micrograms per kilogram and two other compounds were below the minimum reporting levels. At Colorado Creek, Johnson River, and Kamishak River, where fish samples were not collected, no organochlorine compounds were detected in streambed sediment samples. Several semivolatile organic compounds were detected at Colorado Creek and Costello Creek. Only one compound, dibenzothiophene, detected at Costello Creek at a concentration of 85 micrograms per kilogram was above the minimum reporting limit. No semivolatile organic compounds were detected at the Talkeetna, Kamishak, or Johnson Rivers. Trace elements were detected in both fish tissues and streambed sediments. Macroinvertebrate and fish samples contained few taxa at all sites. Total numbers of macroinvertebrate taxa ranged from 19 at the Johnson River to 38 at the Talkeetna River. Diptera were the most abundant and diverse order of macroinvertebrates at all sites. Total numbers of diptera taxa ranged from 8 at the Kamishak River to 19 at the Talkeetna River. Fish communities were represented by a maximum of nine taxa at the Talkeetna River and were absent at Colorado Creek. The Johnson River sampling site produced small numbers of juvenile Dolly Varden, and Costello Creek produced small numbers of both juvenile Dolly Varden and slimy sculpin.
Asteroid families - An initial search
NASA Technical Reports Server (NTRS)
Williams, James G.
1992-01-01
A stereo examination was conducted for clusters in three-dimensional proper element space within a sample of both numbered and faint Palomar-Leiden Survey (PLS) asteroids. The clusters were then objectively filtered for small Poisson probability of chance occurrence; 104 were accepted as families with 4- to 12-member populations, and are interpreted as impact-generated. Structure is common in the well-populated families: the better-sampled families are accordingly discussed in terms of their geometry and taxonomy. Some families are very rich in faint PLS members.
NASA Technical Reports Server (NTRS)
Vangenderen, J. L. (Principal Investigator); Lock, B. F.
1976-01-01
The author has identified the following significant results. Scope of the preprocessing techniques was restricted to standard material from the EROS Data Center accompanied by some enlarging procedures and the use of the diazo process. Investigation has shown that the most appropriate sampling strategy for this study is the stratified random technique. A viable sampling procedure, together with a method for determining minimum number of sample points in order to test results of any interpretation are presented.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wasson, J.T.; Ouyang, Xinwei; Wang, Jianmin
1989-03-01
The authors report concentrations of 14 elements in the metal of 38 iron meteorites and a pallasite. The meteorites are classified based on these data and on structural observations. Three samples are paired with previously classified irons; thus, these additional 35 irons raise the number of well-classified, independent iron meteorites to 598. One Yamato iron contains 342 mg/g Ni, the second highest Ni content in an IAB iron after Oktibbeha County. Two small irons from Western Australia appear to be metal nodules from mesosiderites. Several of the new irons are from Antarctica. Of 24 independent irons from Antarctica, 8 aremore » ungrouped. The fraction, 0.333, is much higher than the fraction 0.161 among all 598 classified irons. Statistical tests show that it is highly improbably ({approximately}2.9% probability) that the Antarctic population is a random sample of the larger population. The difference is probably related to the fact that the median mass of Antarctic irons is about two orders of magnitude smaller than that of non-Antarctic irons. It is doubtful that the difference results from fragmentation patterns yielding different size distributions favoring smaller masses among ungrouped irons. More likely is the possibility that smaller meteoroids tend to sample a larger number of asteroidal source regions, perhaps because small meteoroids tend to have higher ejection velocities or because small meteoroids have random-walked a greater increment of orbital semimajor axis away from that of the parent body.« less
NASA Astrophysics Data System (ADS)
Voss, Sebastian; Zimmermann, Beate; Zimmermann, Alexander
2016-04-01
In the last three decades, an increasing number of studies analyzed spatial patterns in throughfall to investigate the consequences of rainfall redistribution for biogeochemical and hydrological processes in forests. In the majority of cases, variograms were used to characterize the spatial properties of the throughfall data. The estimation of the variogram from sample data requires an appropriate sampling scheme: most importantly, a large sample and an appropriate layout of sampling locations that often has to serve both variogram estimation and geostatistical prediction. While some recommendations on these aspects exist, they focus on Gaussian data and high ratios of the variogram range to the extent of the study area. However, many hydrological data, and throughfall data in particular, do not follow a Gaussian distribution. In this study, we examined the effect of extent, sample size, sampling design, and calculation methods on variogram estimation of throughfall data. For our investigation, we first generated non-Gaussian random fields based on throughfall data with heavy outliers. Subsequently, we sampled the fields with three extents (plots with edge lengths of 25 m, 50 m, and 100 m), four common sampling designs (two grid-based layouts, transect and random sampling), and five sample sizes (50, 100, 150, 200, 400). We then estimated the variogram parameters by method-of-moments and residual maximum likelihood. Our key findings are threefold. First, the choice of the extent has a substantial influence on the estimation of the variogram. A comparatively small ratio of the extent to the correlation length is beneficial for variogram estimation. Second, a combination of a minimum sample size of 150, a design that ensures the sampling of small distances and variogram estimation by residual maximum likelihood offers a good compromise between accuracy and efficiency. Third, studies relying on method-of-moments based variogram estimation may have to employ at least 200 sampling points for reliable variogram estimates. These suggested sample sizes exceed the numbers recommended by studies dealing with Gaussian data by up to 100 %. Given that most previous throughfall studies relied on method-of-moments variogram estimation and sample sizes << 200, our current knowledge about throughfall spatial variability stands on shaky ground.
NASA Astrophysics Data System (ADS)
Voss, Sebastian; Zimmermann, Beate; Zimmermann, Alexander
2016-09-01
In the last decades, an increasing number of studies analyzed spatial patterns in throughfall by means of variograms. The estimation of the variogram from sample data requires an appropriate sampling scheme: most importantly, a large sample and a layout of sampling locations that often has to serve both variogram estimation and geostatistical prediction. While some recommendations on these aspects exist, they focus on Gaussian data and high ratios of the variogram range to the extent of the study area. However, many hydrological data, and throughfall data in particular, do not follow a Gaussian distribution. In this study, we examined the effect of extent, sample size, sampling design, and calculation method on variogram estimation of throughfall data. For our investigation, we first generated non-Gaussian random fields based on throughfall data with large outliers. Subsequently, we sampled the fields with three extents (plots with edge lengths of 25 m, 50 m, and 100 m), four common sampling designs (two grid-based layouts, transect and random sampling) and five sample sizes (50, 100, 150, 200, 400). We then estimated the variogram parameters by method-of-moments (non-robust and robust estimators) and residual maximum likelihood. Our key findings are threefold. First, the choice of the extent has a substantial influence on the estimation of the variogram. A comparatively small ratio of the extent to the correlation length is beneficial for variogram estimation. Second, a combination of a minimum sample size of 150, a design that ensures the sampling of small distances and variogram estimation by residual maximum likelihood offers a good compromise between accuracy and efficiency. Third, studies relying on method-of-moments based variogram estimation may have to employ at least 200 sampling points for reliable variogram estimates. These suggested sample sizes exceed the number recommended by studies dealing with Gaussian data by up to 100 %. Given that most previous throughfall studies relied on method-of-moments variogram estimation and sample sizes ≪200, currently available data are prone to large uncertainties.
Thompson, Jennifer A; Fielding, Katherine; Hargreaves, James; Copas, Andrew
2017-12-01
Background/Aims We sought to optimise the design of stepped wedge trials with an equal allocation of clusters to sequences and explored sample size comparisons with alternative trial designs. Methods We developed a new expression for the design effect for a stepped wedge trial, assuming that observations are equally correlated within clusters and an equal number of observations in each period between sequences switching to the intervention. We minimised the design effect with respect to (1) the fraction of observations before the first and after the final sequence switches (the periods with all clusters in the control or intervention condition, respectively) and (2) the number of sequences. We compared the design effect of this optimised stepped wedge trial to the design effects of a parallel cluster-randomised trial, a cluster-randomised trial with baseline observations, and a hybrid trial design (a mixture of cluster-randomised trial and stepped wedge trial) with the same total cluster size for all designs. Results We found that a stepped wedge trial with an equal allocation to sequences is optimised by obtaining all observations after the first sequence switches and before the final sequence switches to the intervention; this means that the first sequence remains in the control condition and the last sequence remains in the intervention condition for the duration of the trial. With this design, the optimal number of sequences is [Formula: see text], where [Formula: see text] is the cluster-mean correlation, [Formula: see text] is the intracluster correlation coefficient, and m is the total cluster size. The optimal number of sequences is small when the intracluster correlation coefficient and cluster size are small and large when the intracluster correlation coefficient or cluster size is large. A cluster-randomised trial remains more efficient than the optimised stepped wedge trial when the intracluster correlation coefficient or cluster size is small. A cluster-randomised trial with baseline observations always requires a larger sample size than the optimised stepped wedge trial. The hybrid design can always give an equally or more efficient design, but will be at most 5% more efficient. We provide a strategy for selecting a design if the optimal number of sequences is unfeasible. For a non-optimal number of sequences, the sample size may be reduced by allowing a proportion of observations before the first or after the final sequence has switched. Conclusion The standard stepped wedge trial is inefficient. To reduce sample sizes when a hybrid design is unfeasible, stepped wedge trial designs should have no observations before the first sequence switches or after the final sequence switches.
Importance of benthic prey for fishes in coral reef-associated sediments
DeFelice, R.C.; Parrish, J.D.
2003-01-01
The importance of open, sandy substrate adjacent to coral reefs as habitat and a food source for fishes has been little studied in most shallow tropical waters in the Pacific, including Hawai'i. In this study, in Hanalei Bay, Hiwai'i, we identified and quantified the major invertebrate fauna (larger than 0.5 mm) in the well-characterized sands adjoining the shallow fringing reefs. Concurrently, we identified the fish species that seemed to make substantial use of these sand habitats, estimated their density there, sampled their gut contents to examine trophic links with the sand habitat, and made other observations and collections to determine the times, locations, and types of activity there. A variety of (mostly small) polychaeres were dominant in the sediments at most sampling stations, along with many small crustaceans (e.g., amphipods, isopods, ostracods, and small shrimps) and fair numbers of mollusks (especially bivalves) and small echinoids. Fish guts examined contained ???77% of the total number of benthic taxa collected, including nearly all those just listed. However, fish consumption was selective, and the larger shrimps, crabs, and small cryptic fishes were dominant in the diets of most of the numerous predator taxa. Diets of benthic-feeding fishes showed relatively low specific overlap. The fish fauna in this area included substrate-indifferent pelagics, species with various degrees of reef relatedness, reef-restricted species, and (at the other extreme) permanent cryptic sand dwellers. Data on occurrence and movements of fishes indicated that a band of sandy substrate several tens of meters wide next to the reef was an active area for fishes, and activity was considerably different at different times of day and for fish of different ages. These results imply an important trophic role for the benthos in these near-reef habitats in support of reef-associated fishes.
Han, Yanxi; Li, Jinming
2017-10-26
In this era of precision medicine, molecular biology is becoming increasingly significant for the diagnosis and therapeutic management of non-small cell lung cancer. The specimen as the primary element of the whole testing flow is particularly important for maintaining the accuracy of gene alteration testing. Presently, the main sample types applied in routine diagnosis are tissue and cytology biopsies. Liquid biopsies are considered as the most promising alternatives when tissue and cytology samples are not available. Each sample type possesses its own strengths and weaknesses, pertaining to the disparity of sampling, preparation and preservation procedures, the heterogeneity of inter- or intratumors, the tumor cellularity (percentage and number of tumor cells) of specimens, etc., and none of them can individually be a "one size to fit all". Therefore, in this review, we summarized the strengths and weaknesses of different sample types that are widely used in clinical practice, offered solutions to reduce the negative impact of the samples and proposed an optimized strategy for choice of samples during the entire diagnostic course. We hope to provide valuable information to laboratories for choosing optimal clinical specimens to achieve comprehensive functional genomic landscapes and formulate individually tailored treatment plans for NSCLC patients that are in advanced stages.
Yin, Weiwei; Garimalla, Swetha; Moreno, Alberto; Galinski, Mary R; Styczynski, Mark P
2015-08-28
There are increasing efforts to bring high-throughput systems biology techniques to bear on complex animal model systems, often with a goal of learning about underlying regulatory network structures (e.g., gene regulatory networks). However, complex animal model systems typically have significant limitations on cohort sizes, number of samples, and the ability to perform follow-up and validation experiments. These constraints are particularly problematic for many current network learning approaches, which require large numbers of samples and may predict many more regulatory relationships than actually exist. Here, we test the idea that by leveraging the accuracy and efficiency of classifiers, we can construct high-quality networks that capture important interactions between variables in datasets with few samples. We start from a previously-developed tree-like Bayesian classifier and generalize its network learning approach to allow for arbitrary depth and complexity of tree-like networks. Using four diverse sample networks, we demonstrate that this approach performs consistently better at low sample sizes than the Sparse Candidate Algorithm, a representative approach for comparison because it is known to generate Bayesian networks with high positive predictive value. We develop and demonstrate a resampling-based approach to enable the identification of a viable root for the learned tree-like network, important for cases where the root of a network is not known a priori. We also develop and demonstrate an integrated resampling-based approach to the reduction of variable space for the learning of the network. Finally, we demonstrate the utility of this approach via the analysis of a transcriptional dataset of a malaria challenge in a non-human primate model system, Macaca mulatta, suggesting the potential to capture indicators of the earliest stages of cellular differentiation during leukopoiesis. We demonstrate that by starting from effective and efficient approaches for creating classifiers, we can identify interesting tree-like network structures with significant ability to capture the relationships in the training data. This approach represents a promising strategy for inferring networks with high positive predictive value under the constraint of small numbers of samples, meeting a need that will only continue to grow as more high-throughput studies are applied to complex model systems.
Data-driven sensor placement from coherent fluid structures
NASA Astrophysics Data System (ADS)
Manohar, Krithika; Kaiser, Eurika; Brunton, Bingni W.; Kutz, J. Nathan; Brunton, Steven L.
2017-11-01
Optimal sensor placement is a central challenge in the prediction, estimation and control of fluid flows. We reinterpret sensor placement as optimizing discrete samples of coherent fluid structures for full state reconstruction. This permits a drastic reduction in the number of sensors required for faithful reconstruction, since complex fluid interactions can often be described by a small number of coherent structures. Our work optimizes point sensors using the pivoted matrix QR factorization to sample coherent structures directly computed from flow data. We apply this sampling technique in conjunction with various data-driven modal identification methods, including the proper orthogonal decomposition (POD) and dynamic mode decomposition (DMD). In contrast to POD-based sensors, DMD demonstrably enables the optimization of sensors for prediction in systems exhibiting multiple scales of dynamics. Finally, reconstruction accuracy from pivot sensors is shown to be competitive with sensors obtained using traditional computationally prohibitive optimization methods.
Krawczyk, Paweł Adam; Ramlau, Rodryg Adam; Szumiło, Justyna; Kozielski, Jerzy; Kalinka-Warzocha, Ewa; Bryl, Maciej; Knopik-Dąbrowicz, Alina; Spychalski, Łukasz; Szczęsna, Aleksandra; Rydzik, Ewelina; Milanowski, Janusz
2013-01-01
Introduction ALK gene rearrangement is observed in a small subset (3–7%) of non-small cell lung cancer (NSCLC) patients. The efficacy of crizotinib was shown in lung cancer patients harbouring ALK rearrangement. Nowadays, the analysis of ALK gene rearrangement is added to molecular examination of predictive factors. Aim of the study The frequency of ALK gene rearrangement as well as the type of its irregularity was analysed by fluorescence in situ hybridisation (FISH) in tissue samples from NSCLC patients. Material and methods The ALK gene rearrangement was analysed in 71 samples including 53 histological and 18 cytological samples. The analysis could be performed in 56 cases (78.87%), significantly more frequently in histological than in cytological materials. The encountered problem with ALK rearrangement diagnosis resulted from the scarcity of tumour cells in cytological samples, high background fluorescence noises and fragmentation of cell nuclei. Results The normal ALK copy number without gene rearrangement was observed in 26 (36.62%) patients ALK gene polysomy without gene rearrangement was observed in 25 (35.21%) samples while in 3 (4.23%) samples ALK gene amplification was found. ALK gene rearrangement was observed in 2 (2.82%) samples from males, while in the first case the rearrangement coexisted with ALK amplification. In the second case, signet-ring tumour cells were found during histopathological examination and this patient was successfully treated with crizotinib with partial remission lasting 16 months. Conclusions FISH is a useful technique for ALK gene rearrangement analysis which allows us to specify the type of gene irregularities. ALK gene examination could be performed in histological as well as cytological (cellblocks) samples, but obtaining a reliable result in cytological samples depends on the cellularity of examined materials. PMID:24592134
Optimal tumor sampling for immunostaining of biomarkers in breast carcinoma
2011-01-01
Introduction Biomarkers, such as Estrogen Receptor, are used to determine therapy and prognosis in breast carcinoma. Immunostaining assays of biomarker expression have a high rate of inaccuracy; for example, estimates are as high as 20% for Estrogen Receptor. Biomarkers have been shown to be heterogeneously expressed in breast tumors and this heterogeneity may contribute to the inaccuracy of immunostaining assays. Currently, no evidence-based standards exist for the amount of tumor that must be sampled in order to correct for biomarker heterogeneity. The aim of this study was to determine the optimal number of 20X fields that are necessary to estimate a representative measurement of expression in a whole tissue section for selected biomarkers: ER, HER-2, AKT, ERK, S6K1, GAPDH, Cytokeratin, and MAP-Tau. Methods Two collections of whole tissue sections of breast carcinoma were immunostained for biomarkers. Expression was quantified using the Automated Quantitative Analysis (AQUA) method of quantitative immunofluorescence. Simulated sampling of various numbers of fields (ranging from one to thirty five) was performed for each marker. The optimal number was selected for each marker via resampling techniques and minimization of prediction error over an independent test set. Results The optimal number of 20X fields varied by biomarker, ranging between three to fourteen fields. More heterogeneous markers, such as MAP-Tau protein, required a larger sample of 20X fields to produce representative measurement. Conclusions The optimal number of 20X fields that must be sampled to produce a representative measurement of biomarker expression varies by marker with more heterogeneous markers requiring a larger number. The clinical implication of these findings is that breast biopsies consisting of a small number of fields may be inadequate to represent whole tumor biomarker expression for many markers. Additionally, for biomarkers newly introduced into clinical use, especially if therapeutic response is dictated by level of expression, the optimal size of tissue sample must be determined on a marker-by-marker basis. PMID:21592345
NASA Astrophysics Data System (ADS)
Hertanto, B. S.; Nurmalasari, C. D. A.; Nuhriawangsa, A. M. P.; Cahyadi, M.; Kartikasari, L. R.
2018-01-01
The aim of this study was to determine the physical and microbiological quality of chicken meat produced by the different type of enterprise slaughterhouse in Karanganyar District. The number of 20 poultry slaughterhouses was determined by convenience sampling method. The samples of chicken meat were randomly collected from medium enterprise poultry slaughterhouses (n=12) and small enterprise poultry slaughterhouses (n=8). A survey was carried out among poultry slaughterhouses in Karanganyar District. All the samples were subjected to physical quality consisted of pH test, texture, and color, while microbiological quality consisted of total plate count, microbial detection of Escherichia coli and Salmonella. The data were analyzed using descriptive quantitative analysis. The study showed that chicken meat in 6 small enterprise slaughterhouses and 11 medium enterprise slaughterhouses had normal pH of 5.81 - 6.3. Color and texture of chicken meats had relatively normal in both small and medium enterprise slaughterhouses. The total plate count of chicken meat showed in both small and medium enterprise slaughterhouses was <1x106 CFU/gr. The test of bacterial contamination showed that 3 of small and medium enterprise slaughterhouses were positively contaminated by Escherichia coli of >1x101 CFU/gr, and Salmonella was detected in 1 medium enterprise slaughterhouse. The overall results of the study suggest that the potential risk of chicken meat contamination depends on the processing of chicken meat in poultry slaughterhouses.
Ritchie, Nicholas W M; Newbury, Dale E; Lindstrom, Abigail P
2011-12-01
Artifacts are the nemesis of trace element analysis in electron-excited energy dispersive X-ray spectrometry. Peaks that result from nonideal behavior in the detector or sample can fool even an experienced microanalyst into believing that they have trace amounts of an element that is not present. Many artifacts, such as the Si escape peak, absorption edges, and coincidence peaks, can be traced to the detector. Others, such as secondary fluorescence peaks and scatter peaks, can be traced to the sample. We have identified a new sample-dependent artifact that we attribute to Compton scattering of energetic X-rays generated in a small feature and subsequently scattered from a low atomic number matrix. It seems likely that this artifact has not previously been reported because it only occurs under specific conditions and represents a relatively small signal. However, with the advent of silicon drift detectors and their utility for trace element analysis, we anticipate that more people will observe it and possibly misidentify it. Though small, the artifact is not inconsequential. Under some conditions, it is possible to mistakenly identify the Compton scatter artifact as approximately 1% of an element that is not present.
Gao, Yi Qin
2008-04-07
Here, we introduce a simple self-adaptive computational method to enhance the sampling in energy, configuration, and trajectory spaces. The method makes use of two strategies. It first uses a non-Boltzmann distribution method to enhance the sampling in the phase space, in particular, in the configuration space. The application of this method leads to a broad energy distribution in a large energy range and a quickly converged sampling of molecular configurations. In the second stage of simulations, the configuration space of the system is divided into a number of small regions according to preselected collective coordinates. An enhanced sampling of reactive transition paths is then performed in a self-adaptive fashion to accelerate kinetics calculations.
Finding a Needle in a Haystack: Distinguishing Mexican Maize Landraces Using a Small Number of SNPs
Caldu-Primo, Jose L.; Mastretta-Yanes, Alicia; Wegier, Ana; Piñero, Daniel
2017-01-01
In Mexico's territory, the center of origin and domestication of maize (Zea mays), there is a large phenotypic diversity of this crop. This diversity has been classified into “landraces.” Previous studies have reported that genomic variation in Mexican maize is better explained by environmental factors, particularly those related with altitude, than by landrace. Still, landraces are extensively used by agronomists, who recognize them as stable and discriminatory categories for the classification of samples. In order to investigate the genomic foundation of maize landraces, we analyzed genomic data (35,909 SNPs from Illumina MaizeSNP50 BeadChip) obtained from 50 samples representing five maize landraces (Comiteco, Conejo, Tehua, Zapalote Grande, and Zapalote Chico), and searched for markers suitable for landrace assignment. Landrace clusters could not be identified taking all the genomic information, but they become manifest taking only a subset of SNPs with high FST among landraces. Discriminant analysis of principal components was conducted to classify samples using SNP data. Two classification analyses were done, first classifying samples by landrace and then by altitude category. Through this classification method, we identified 20 landrace-informative SNPs and 14 altitude-informative SNPs, with only 6 SNPs in common for both analyses. These results show that Mexican maize phenotypic diversity can be classified in landraces using a small number of genomic markers, given the fact that landrace genomic diversity is influenced by environmental factors as well as artificial selection due to bio-cultural practices. PMID:28458682
Refractory materials in comet samples
NASA Astrophysics Data System (ADS)
Joswiak, D. J.; Brownlee, D. E.; Nguyen, A. N.; Messenger, S.
2017-08-01
Transmission electron microscope examination of more than 250 fragments, >1 μm from comet Wild 2 and a giant cluster interplanetary dust particle (GCP) of probable cometary origin has revealed four new calcium-aluminum-rich inclusions (CAIs), an amoeboid olivine aggregate (AOA), and an additional AOA or Al-rich chondrule (ARC) object. All of the CAIs have concentric mineral structures and are composed of spinel + anorthite cores surrounded by Al,Ti clinopyroxenes and are similar to two previous CAIs discovered in Wild 2. All of the cometary refractory objects are of moderate refractory character. The mineral assemblages, textures, and bulk compositions of the comet CAIs are similar to nodules in fine-grained, spinel-rich inclusions (FGIs) found in primitive chondrites and like the nodules may be nebular condensates that were altered via solid-gas reactions in the solar nebula. Oxygen isotopes collected on one Wild 2 CAI also match FGIs. The lack of the most refractory inclusions in the comet samples may reflect the higher abundances of small moderately refractory CAI nodules that were produced in the nebula and the small sample sizes collected. In the comet samples, approximately 2-3% of all fragments larger than 1 μm, by number, are CAIs and nearly 50% of all bulbous Stardust tracks contain at least one CAI. We estimate that 0.5 volume % of Wild 2 material and 1 volume % of GCP is in the form of CAIs. ARCs and AOAs account for <1% of the Wild 2 and GCP grains by number.
Small Demodex populations colonize most parts of the skin of healthy dogs.
Ravera, Iván; Altet, Laura; Francino, Olga; Sánchez, Armand; Roldán, Wendy; Villanueva, Sergio; Bardagí, Mar; Ferrer, Lluís
2013-02-01
It is unproven that all dogs harbour Demodex mites in their skin. In fact, several microscopic studies have failed to demonstrate mites in healthy dogs. Demodex canis is a normal inhabitant of the skin of most, if not all, dogs. This hypothesis was tested using a sensitive real-time PCR to detect Demodex DNA in the skin of dogs. One hundred dogs living in a humane society shelter, 20 privately owned and healthy dogs and eight dogs receiving immunosuppressive or antineoplastic therapy. Hair samples (250-300 hairs with their hair bulbs) were taken from five or 20 skin locations. A real-time PCR that amplifies a 166 bp sequence of the D. canis chitin synthase gene was used. The percentage of positive dogs increased with the number of sampling points. When a large canine population was sampled at five cutaneous locations, 18% of dogs were positive for Demodex DNA. When 20 skin locations were sampled, all dogs tested positive for mite DNA. Our study indicates that Demodex colonization of the skin is present in all dogs, independent of age, sex, breed or coat. Nevertheless, the population of mites in a healthy dog appears to be small. Demodex DNA was amplified from all 20 cutaneous points investigated, without statistically significant differences. Using a real-time PCR technique, Demodex mites, albeit in very low numbers, were found to be normal inhabitants of haired areas of the skin of healthy dogs. © 2013 The Authors. Veterinary Dermatology © 2013 ESVD and ACVD.
The coverage of a random sample from a biological community.
Engen, S
1975-03-01
A taxonomic group will frequently have a large number of species with small abundances. When a sample is drawn at random from this group, one is therefore faced with the problem that a large proportion of the species will not be discovered. A general definition of quantitative measures of "sample coverage" is proposed, and the problem of statistical inference is considered for two special cases, (1) the actual total relative abundance of those species that are represented in the sample, and (2) their relative contribution to the information index of diversity. The analysis is based on a extended version of the negative binomial species frequency model. The results are tabulated.
Evaluating methods for monitoring populations of Mexican spotted owls: A case study
Jospeh L. Ganey; Gary C. White; David C. Bowden; Alan B. Franklin
2004-01-01
Monitoring population status of rare or elusive species presents special challenges. Understanding population trends requires separating signal (true and important changes in abundance) from noise (normal temporal and sampling variation; e.g., Block et al. 2001). This is particularly difficult when small numbers or elusive habits make it difficult to obtain precise...
Estimating forest attribute parameters for small areas using nearest neighbors techniques
Ronald E. McRoberts
2012-01-01
Nearest neighbors techniques have become extremely popular, particularly for use with forest inventory data. With these techniques, a population unit prediction is calculated as a linear combination of observations for a selected number of population units in a sample that are most similar, or nearest, in a space of ancillary variables to the population unit requiring...
NASA Astrophysics Data System (ADS)
Polyakov, A. F.; Strat'ev, V. K.; Tret'yakov, A. F.; Shekhter, Yu. L.
2010-06-01
Heat transfer from six samples of porous reticular material to cooling gas (air) at small Reynolds numbers is experimentally studied. The specific features pertinent to heat transfer essentially affected by longitudinal heat conductivity along gas flow are analyzed. The experimental results are generalized in the form of dimensionless empirical relations.
[Distribution form of nutrients in roof runoff].
Wang, Biao; Li, Tian; Meng, Ying-Ying; Ren, Zhong-Jia; Cao, Bing-Yin
2008-11-01
6 rainfall-runoff events were examined from a concrete roof by a traffic artery in Shanghai to characterize the particle-bound and dissolved nutrients in roof runoff and to get more knowledge about roof runoff pollutants and lay the theoretic foundation for the selection of roof runoff purification method. Results indicated that event mean concentration (EMC) of TN ranges between 4.208 mg/L and 8.427 mg/L compared to 0.078-0.185 mg/L for TP. DN and PP are respectively the primary forms of TN and TP. During the runoff, the number ratio of small particles is gradually increased. The dynamic behavior of TP is similar to TSS, but the TN-TSS relationship in the course of runoff is not obvious. The increase of both PN/TSS and the number ration of small particles showed that nitrogen (microgramme per gram particle weight) attached to small particles is more than large particles. Regression analysis between TSS and TP & PP of samples in the early 10 min of runoff results in a high R2, but the relationship between TSS and TP & PP of samples in the entire runoff is not as close as the early 10 min, which reflects that the phosphorus quantity attached to the particle changes from the perspective of the entire course of runoff. First flush of TP is stronger than TN because the load discharge of dissolved nutrients is more stable than particle nutrients. In addition, more nitrogen absorbed by the unit mass small particles to a certain degree weakens the first flush of TN.
Exploring the Energy Landscapes of Protein Folding Simulations with Bayesian Computation
Burkoff, Nikolas S.; Várnai, Csilla; Wells, Stephen A.; Wild, David L.
2012-01-01
Nested sampling is a Bayesian sampling technique developed to explore probability distributions localized in an exponentially small area of the parameter space. The algorithm provides both posterior samples and an estimate of the evidence (marginal likelihood) of the model. The nested sampling algorithm also provides an efficient way to calculate free energies and the expectation value of thermodynamic observables at any temperature, through a simple post processing of the output. Previous applications of the algorithm have yielded large efficiency gains over other sampling techniques, including parallel tempering. In this article, we describe a parallel implementation of the nested sampling algorithm and its application to the problem of protein folding in a Gō-like force field of empirical potentials that were designed to stabilize secondary structure elements in room-temperature simulations. We demonstrate the method by conducting folding simulations on a number of small proteins that are commonly used for testing protein-folding procedures. A topological analysis of the posterior samples is performed to produce energy landscape charts, which give a high-level description of the potential energy surface for the protein folding simulations. These charts provide qualitative insights into both the folding process and the nature of the model and force field used. PMID:22385859
Exploring the energy landscapes of protein folding simulations with Bayesian computation.
Burkoff, Nikolas S; Várnai, Csilla; Wells, Stephen A; Wild, David L
2012-02-22
Nested sampling is a Bayesian sampling technique developed to explore probability distributions localized in an exponentially small area of the parameter space. The algorithm provides both posterior samples and an estimate of the evidence (marginal likelihood) of the model. The nested sampling algorithm also provides an efficient way to calculate free energies and the expectation value of thermodynamic observables at any temperature, through a simple post processing of the output. Previous applications of the algorithm have yielded large efficiency gains over other sampling techniques, including parallel tempering. In this article, we describe a parallel implementation of the nested sampling algorithm and its application to the problem of protein folding in a Gō-like force field of empirical potentials that were designed to stabilize secondary structure elements in room-temperature simulations. We demonstrate the method by conducting folding simulations on a number of small proteins that are commonly used for testing protein-folding procedures. A topological analysis of the posterior samples is performed to produce energy landscape charts, which give a high-level description of the potential energy surface for the protein folding simulations. These charts provide qualitative insights into both the folding process and the nature of the model and force field used. Copyright © 2012 Biophysical Society. Published by Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Jenk, Theo Manuel; Rubino, Mauro; Etheridge, David; Ciobanu, Viorela Gabriela; Blunier, Thomas
2016-08-01
Palaeoatmospheric records of carbon dioxide and its stable carbon isotope composition (δ13C) obtained from polar ice cores provide important constraints on the natural variability of the carbon cycle. However, the measurements are both analytically challenging and time-consuming; thus only data exist from a limited number of sampling sites and time periods. Additional analytical resources with high analytical precision and throughput are thus desirable to extend the existing datasets. Moreover, consistent measurements derived by independent laboratories and a variety of analytical systems help to further increase confidence in the global CO2 palaeo-reconstructions. Here, we describe our new set-up for simultaneous measurements of atmospheric CO2 mixing ratios and atmospheric δ13C and δ18O-CO2 in air extracted from ice core samples. The centrepiece of the system is a newly designed needle cracker for the mechanical release of air entrapped in ice core samples of 8-13 g operated at -45 °C. The small sample size allows for high resolution and replicate sampling schemes. In our method, CO2 is cryogenically and chromatographically separated from the bulk air and its isotopic composition subsequently determined by continuous flow isotope ratio mass spectrometry (IRMS). In combination with thermal conductivity measurement of the bulk air, the CO2 mixing ratio is calculated. The analytical precision determined from standard air sample measurements over ice is ±1.9 ppm for CO2 and ±0.09 ‰ for δ13C. In a laboratory intercomparison study with CSIRO (Aspendale, Australia), good agreement between CO2 and δ13C results is found for Law Dome ice core samples. Replicate analysis of these samples resulted in a pooled standard deviation of 2.0 ppm for CO2 and 0.11 ‰ for δ13C. These numbers are good, though they are rather conservative estimates of the overall analytical precision achieved for single ice sample measurements. Facilitated by the small sample requirement, replicate measurements are feasible, allowing the method precision to be improved potentially. Further, new analytical approaches are introduced for the accurate correction of the procedural blank and for a consistent detection of measurement outliers, which is based on δ18O-CO2 and the exchange of oxygen between CO2 and the surrounding ice (H2O).
Kanda, Kojun; Pflug, James M; Sproul, John S; Dasenko, Mark A; Maddison, David R
2015-01-01
In this paper we explore high-throughput Illumina sequencing of nuclear protein-coding, ribosomal, and mitochondrial genes in small, dried insects stored in natural history collections. We sequenced one tenebrionid beetle and 12 carabid beetles ranging in size from 3.7 to 9.7 mm in length that have been stored in various museums for 4 to 84 years. Although we chose a number of old, small specimens for which we expected low sequence recovery, we successfully recovered at least some low-copy nuclear protein-coding genes from all specimens. For example, in one 56-year-old beetle, 4.4 mm in length, our de novo assembly recovered about 63% of approximately 41,900 nucleotides in a target suite of 67 nuclear protein-coding gene fragments, and 70% using a reference-based assembly. Even in the least successfully sequenced carabid specimen, reference-based assembly yielded fragments that were at least 50% of the target length for 34 of 67 nuclear protein-coding gene fragments. Exploration of alternative references for reference-based assembly revealed few signs of bias created by the reference. For all specimens we recovered almost complete copies of ribosomal and mitochondrial genes. We verified the general accuracy of the sequences through comparisons with sequences obtained from PCR and Sanger sequencing, including of conspecific, fresh specimens, and through phylogenetic analysis that tested the placement of sequences in predicted regions. A few possible inaccuracies in the sequences were detected, but these rarely affected the phylogenetic placement of the samples. Although our sample sizes are low, an exploratory regression study suggests that the dominant factor in predicting success at recovering nuclear protein-coding genes is a high number of Illumina reads, with success at PCR of COI and killing by immersion in ethanol being secondary factors; in analyses of only high-read samples, the primary significant explanatory variable was body length, with small beetles being more successfully sequenced.
Dasenko, Mark A.
2015-01-01
In this paper we explore high-throughput Illumina sequencing of nuclear protein-coding, ribosomal, and mitochondrial genes in small, dried insects stored in natural history collections. We sequenced one tenebrionid beetle and 12 carabid beetles ranging in size from 3.7 to 9.7 mm in length that have been stored in various museums for 4 to 84 years. Although we chose a number of old, small specimens for which we expected low sequence recovery, we successfully recovered at least some low-copy nuclear protein-coding genes from all specimens. For example, in one 56-year-old beetle, 4.4 mm in length, our de novo assembly recovered about 63% of approximately 41,900 nucleotides in a target suite of 67 nuclear protein-coding gene fragments, and 70% using a reference-based assembly. Even in the least successfully sequenced carabid specimen, reference-based assembly yielded fragments that were at least 50% of the target length for 34 of 67 nuclear protein-coding gene fragments. Exploration of alternative references for reference-based assembly revealed few signs of bias created by the reference. For all specimens we recovered almost complete copies of ribosomal and mitochondrial genes. We verified the general accuracy of the sequences through comparisons with sequences obtained from PCR and Sanger sequencing, including of conspecific, fresh specimens, and through phylogenetic analysis that tested the placement of sequences in predicted regions. A few possible inaccuracies in the sequences were detected, but these rarely affected the phylogenetic placement of the samples. Although our sample sizes are low, an exploratory regression study suggests that the dominant factor in predicting success at recovering nuclear protein-coding genes is a high number of Illumina reads, with success at PCR of COI and killing by immersion in ethanol being secondary factors; in analyses of only high-read samples, the primary significant explanatory variable was body length, with small beetles being more successfully sequenced. PMID:26716693
NASA Technical Reports Server (NTRS)
Thronson, Harley A., Jr.; Latter, William B.; Black, John H.; Bally, John; Hacking, Perry
1987-01-01
A large sample of evolved carbon-rich and oxygen-rich objects has been studied using data from the IRAS Point Source Catalog. The number density of infrared-emitting 'carbon' stars shows no variation with Galactocentric radius, while the evolved 'oxygen' star volume density can be well fitted by a given law. A law is given for the number of carbon stars; a total is found in the Galaxy of 48,000 highly evolved oxygen stars. The mass-return rate for all evolved stars is found to be 0.35 solar mass/yr, with a small percentage contribution from carbon stars. The mass-loss rates for both types of stars are dominated by the small number of objects with the smallest rates. A mean lifetime of about 200,000 yr is obtained for both carbon and oxygen stars. Main-sequence stars in the mass range of three to five solar masses are the probable precursors of the carbon stars.
Collective behavior of large-scale neural networks with GPU acceleration.
Qu, Jingyi; Wang, Rubin
2017-12-01
In this paper, the collective behaviors of a small-world neuronal network motivated by the anatomy of a mammalian cortex based on both Izhikevich model and Rulkov model are studied. The Izhikevich model can not only reproduce the rich behaviors of biological neurons but also has only two equations and one nonlinear term. Rulkov model is in the form of difference equations that generate a sequence of membrane potential samples in discrete moments of time to improve computational efficiency. These two models are suitable for the construction of large scale neural networks. By varying some key parameters, such as the connection probability and the number of nearest neighbor of each node, the coupled neurons will exhibit types of temporal and spatial characteristics. It is demonstrated that the implementation of GPU can achieve more and more acceleration than CPU with the increasing of neuron number and iterations. These two small-world network models and GPU acceleration give us a new opportunity to reproduce the real biological network containing a large number of neurons.
Tokar, Tomas; Pastrello, Chiara; Ramnarine, Varune R.; Zhu, Chang-Qi; Craddock, Kenneth J.; Pikor, Larrisa A.; Vucic, Emily A.; Vary, Simon; Shepherd, Frances A.; Tsao, Ming-Sound; Lam, Wan L.; Jurisica, Igor
2018-01-01
In many cancers, significantly down- or upregulated genes are found within chromosomal regions with DNA copy number alteration opposite to the expression changes. Generally, this paradox has been overlooked as noise, but can potentially be a consequence of interference of epigenetic regulatory mechanisms, including microRNA-mediated control of mRNA levels. To explore potential associations between microRNAs and paradoxes in non-small-cell lung cancer (NSCLC) we curated and analyzed lung adenocarcinoma (LUAD) data, comprising gene expressions, copy number aberrations (CNAs) and microRNA expressions. We integrated data from 1,062 tumor samples and 241 normal lung samples, including newly-generated array comparative genomic hybridization (aCGH) data from 63 LUAD samples. We identified 85 “paradoxical” genes whose differential expression consistently contrasted with aberrations of their copy numbers. Paradoxical status of 70 out of 85 genes was validated on sample-wise basis using The Cancer Genome Atlas (TCGA) LUAD data. Of these, 41 genes are prognostic and form a clinically relevant signature, which we validated on three independent datasets. By meta-analysis of results from 9 LUAD microRNA expression studies we identified 24 consistently-deregulated microRNAs. Using TCGA-LUAD data we showed that deregulation of 19 of these microRNAs explains differential expression of the paradoxical genes. Our results show that deregulation of paradoxical genes is crucial in LUAD and their expression pattern is maintained epigenetically, defying gene copy number status. PMID:29507679
Resolving occlusion and segmentation errors in multiple video object tracking
NASA Astrophysics Data System (ADS)
Cheng, Hsu-Yung; Hwang, Jenq-Neng
2009-02-01
In this work, we propose a method to integrate the Kalman filter and adaptive particle sampling for multiple video object tracking. The proposed framework is able to detect occlusion and segmentation error cases and perform adaptive particle sampling for accurate measurement selection. Compared with traditional particle filter based tracking methods, the proposed method generates particles only when necessary. With the concept of adaptive particle sampling, we can avoid degeneracy problem because the sampling position and range are dynamically determined by parameters that are updated by Kalman filters. There is no need to spend time on processing particles with very small weights. The adaptive appearance for the occluded object refers to the prediction results of Kalman filters to determine the region that should be updated and avoids the problem of using inadequate information to update the appearance under occlusion cases. The experimental results have shown that a small number of particles are sufficient to achieve high positioning and scaling accuracy. Also, the employment of adaptive appearance substantially improves the positioning and scaling accuracy on the tracking results.
NASA Technical Reports Server (NTRS)
Hudson, Nicolas; Lin, Ying; Barengoltz, Jack
2010-01-01
A method for evaluating the probability of a Viable Earth Microorganism (VEM) contaminating a sample during the sample acquisition and handling (SAH) process of a potential future Mars Sample Return mission is developed. A scenario where multiple core samples would be acquired using a rotary percussive coring tool, deployed from an arm on a MER class rover is analyzed. The analysis is conducted in a structured way by decomposing sample acquisition and handling process into a series of discrete time steps, and breaking the physical system into a set of relevant components. At each discrete time step, two key functions are defined: The probability of a VEM being released from each component, and the transport matrix, which represents the probability of VEM transport from one component to another. By defining the expected the number of VEMs on each component at the start of the sampling process, these decompositions allow the expected number of VEMs on each component at each sampling step to be represented as a Markov chain. This formalism provides a rigorous mathematical framework in which to analyze the probability of a VEM entering the sample chain, as well as making the analysis tractable by breaking the process down into small analyzable steps.
Neurons from the adult human dentate nucleus: neural networks in the neuron classification.
Grbatinić, Ivan; Marić, Dušica L; Milošević, Nebojša T
2015-04-07
Topological (central vs. border neuron type) and morphological classification of adult human dentate nucleus neurons according to their quantified histomorphological properties using neural networks on real and virtual neuron samples. In the real sample 53.1% and 14.1% of central and border neurons, respectively, are classified correctly with total of 32.8% of misclassified neurons. The most important result present 62.2% of misclassified neurons in border neurons group which is even greater than number of correctly classified neurons (37.8%) in that group, showing obvious failure of network to classify neurons correctly based on computational parameters used in our study. On the virtual sample 97.3% of misclassified neurons in border neurons group which is much greater than number of correctly classified neurons (2.7%) in that group, again confirms obvious failure of network to classify neurons correctly. Statistical analysis shows that there is no statistically significant difference in between central and border neurons for each measured parameter (p>0.05). Total of 96.74% neurons are morphologically classified correctly by neural networks and each one belongs to one of the four histomorphological types: (a) neurons with small soma and short dendrites, (b) neurons with small soma and long dendrites, (c) neuron with large soma and short dendrites, (d) neurons with large soma and long dendrites. Statistical analysis supports these results (p<0.05). Human dentate nucleus neurons can be classified in four neuron types according to their quantitative histomorphological properties. These neuron types consist of two neuron sets, small and large ones with respect to their perykarions with subtypes differing in dendrite length i.e. neurons with short vs. long dendrites. Besides confirmation of neuron classification on small and large ones, already shown in literature, we found two new subtypes i.e. neurons with small soma and long dendrites and with large soma and short dendrites. These neurons are most probably equally distributed throughout the dentate nucleus as no significant difference in their topological distribution is observed. Copyright © 2015 Elsevier Ltd. All rights reserved.
Designing a two-rank acceptance sampling plan for quality inspection of geospatial data products
NASA Astrophysics Data System (ADS)
Tong, Xiaohua; Wang, Zhenhua; Xie, Huan; Liang, Dan; Jiang, Zuoqin; Li, Jinchao; Li, Jun
2011-10-01
To address the disadvantages of classical sampling plans designed for traditional industrial products, we originally propose a two-rank acceptance sampling plan (TRASP) for the inspection of geospatial data outputs based on the acceptance quality level (AQL). The first rank sampling plan is to inspect the lot consisting of map sheets, and the second is to inspect the lot consisting of features in an individual map sheet. The TRASP design is formulated as an optimization problem with respect to sample size and acceptance number, which covers two lot size cases. The first case is for a small lot size with nonconformities being modeled by a hypergeometric distribution function, and the second is for a larger lot size with nonconformities being modeled by a Poisson distribution function. The proposed TRASP is illustrated through two empirical case studies. Our analysis demonstrates that: (1) the proposed TRASP provides a general approach for quality inspection of geospatial data outputs consisting of non-uniform items and (2) the proposed acceptance sampling plan based on TRASP performs better than other classical sampling plans. It overcomes the drawbacks of percent sampling, i.e., "strictness for large lot size, toleration for small lot size," and those of a national standard used specifically for industrial outputs, i.e., "lots with different sizes corresponding to the same sampling plan."
Method for collecting spores from a mold
Au, Frederick H. F.; Beckert, Werner F.
1977-01-01
A technique and apparatus used therewith for determining the uptake of plutonium and other contaminants by soil microorganisms which, in turn, gives a measure of the plutonium and/or other contaminants available to the biosphere at that particular time. A measured quantity of uncontaminated spores of a selected mold is added to a moistened sample of the soil to be tested. The mixture is allowed to sit a predetermined number of days under specified temperature conditions. An agar layer is then applied to the top of the sample. After three or more days, when spores of the mold growing in the sample have formed, the spores are collected by a miniature vacuum collection apparatus operated under preselected vacuum conditions, which collect only the spores with essentially no contamination by mycelial fragments or culture medium. After collection, the fungal spores are dried and analyzed for the plutonium and/or other contaminants. The apparatus is also suitable for collection of pollen, small insects, dust and other small particles, material from thin-layer chromatography plates, etc.
Willard, M D; Simpson, R B; Fossum, T W; Cohen, N D; Delles, E K; Kolp, D L; Carey, D P; Reinhart, G A
1994-04-15
Sixteen German Shepherd Dogs were found, via quantitative microbial culture of intestinal fluid samples, to have small intestinal bacterial overgrowth (IBO) over an 11-month period. All dogs were deficient in serum IgA. Consistent clinical signs suggestive of an alimentary tract disorder were not observed. Serum cobalamin determinations were not helpful in detecting IBO. Serum folate concentrations had variable sensitivity and specificity for detecting dogs from which we could culture > or = 1 x 10(5) bacterial/ml from intestinal fluid samples in the nonfed state. Histologic and intestinal mucosal cytologic examinations were not useful in detecting IBO. Substantial within-dog and between-dog variation was found in the numbers and species of bacteria in the intestines. The difficulty in diagnosing IBO, the variability in organisms found in individual dogs on repeated sampling, the likelihood that intestinal fluid microbial cultures failed to diagnose IBO in some dogs, and the potential of IBO to be clinically inapparent were the most important findings in this study.
DNA extraction and amplification from contemporary Polynesian bark-cloth.
Moncada, Ximena; Payacán, Claudia; Arriaza, Francisco; Lobos, Sergio; Seelenfreund, Daniela; Seelenfreund, Andrea
2013-01-01
Paper mulberry has been used for thousands of years in Asia and Oceania for making paper and bark-cloth, respectively. Museums around the world hold valuable collections of Polynesian bark-cloth. Genetic analysis of the plant fibers from which the textiles were made may answer a number of questions of interest related to provenance, authenticity or species used in the manufacture of these textiles. Recovery of nucleic acids from paper mulberry bark-cloth has not been reported before. We describe a simple method for the extraction of PCR-amplifiable DNA from small samples of contemporary Polynesian bark-cloth (tapa) using two types of nuclear markers. We report the amplification of about 300 bp sequences of the ITS1 region and of a microsatellite marker. Sufficient DNA was retrieved from all bark-cloth samples to permit successful PCR amplification. This method shows a means of obtaining useful genetic information from modern bark-cloth samples and opens perspectives for the analyses of small fragments derived from ethnographic materials.
DNA Extraction and Amplification from Contemporary Polynesian Bark-Cloth
Moncada, Ximena; Payacán, Claudia; Arriaza, Francisco; Lobos, Sergio; Seelenfreund, Daniela; Seelenfreund, Andrea
2013-01-01
Background Paper mulberry has been used for thousands of years in Asia and Oceania for making paper and bark-cloth, respectively. Museums around the world hold valuable collections of Polynesian bark-cloth. Genetic analysis of the plant fibers from which the textiles were made may answer a number of questions of interest related to provenance, authenticity or species used in the manufacture of these textiles. Recovery of nucleic acids from paper mulberry bark-cloth has not been reported before. Methodology We describe a simple method for the extraction of PCR-amplifiable DNA from small samples of contemporary Polynesian bark-cloth (tapa) using two types of nuclear markers. We report the amplification of about 300 bp sequences of the ITS1 region and of a microsatellite marker. Conclusions Sufficient DNA was retrieved from all bark-cloth samples to permit successful PCR amplification. This method shows a means of obtaining useful genetic information from modern bark-cloth samples and opens perspectives for the analyses of small fragments derived from ethnographic materials. PMID:23437166
Prognostic relevance of lymph node ratio and total lymph node count for small bowel adenocarcinoma.
Tran, Thuy B; Qadan, Motaz; Dua, Monica M; Norton, Jeffrey A; Poultsides, George A; Visser, Brendan C
2015-08-01
Nodal metastasis is a known prognostic factor for small bowel adenocarcinoma. The goals of this study were to evaluate the number of lymph nodes (LNs) that should be retrieved and the impact of lymph node ratio (LNR) on survival. Surveillance, Epidemiology, and End Results was queried to identify patients with small bowel adenocarcinoma who underwent resection from 1988 to 2010. Survival was calculated with the Kaplan-Meier method. Multivariate analysis identified predictors of survival. A total of 2,772 patients underwent resection with at least one node retrieved, and this sample included equal numbers of duodenal (n = 1,387) and jejunoileal (n = 1,386) adenocarcinomas. There were 1,371 patients with no nodal metastasis (N0, 49.4%), 928 N1 (33.5%), and 474 N2 (17.1%). The median numbers of LNs examined for duodenal and jejunoileal cancers were 9 and 8, respectively. Cut-point analysis demonstrated that harvesting at least 9 for jejunoileal and 5 LN for duodenal cancers resulted in the greatest survival difference. Increasing LNR at both sites was associated with decreased overall median survival (LNR = 0, 71 months; LNR 0-0.02, 35 months; LNR 0.21-0.4, 25 months; and LNR >0.4, 16 months; P < .001). Multivariate analysis confirmed number of LNs examined, T-stage, LN positivity, and LNR were independent predictors of survival. LNR has a profound impact on survival in patients with small bowel adenocarcinoma. To achieve adequate staging, we recommend retrieving a minimum of 5 LN for duodenal and 9 LN for jejunoileal adenocarcinomas. Copyright © 2015 Elsevier Inc. All rights reserved.
Probabilistic generation of random networks taking into account information on motifs occurrence.
Bois, Frederic Y; Gayraud, Ghislaine
2015-01-01
Because of the huge number of graphs possible even with a small number of nodes, inference on network structure is known to be a challenging problem. Generating large random directed graphs with prescribed probabilities of occurrences of some meaningful patterns (motifs) is also difficult. We show how to generate such random graphs according to a formal probabilistic representation, using fast Markov chain Monte Carlo methods to sample them. As an illustration, we generate realistic graphs with several hundred nodes mimicking a gene transcription interaction network in Escherichia coli.
Probabilistic Generation of Random Networks Taking into Account Information on Motifs Occurrence
Bois, Frederic Y.
2015-01-01
Abstract Because of the huge number of graphs possible even with a small number of nodes, inference on network structure is known to be a challenging problem. Generating large random directed graphs with prescribed probabilities of occurrences of some meaningful patterns (motifs) is also difficult. We show how to generate such random graphs according to a formal probabilistic representation, using fast Markov chain Monte Carlo methods to sample them. As an illustration, we generate realistic graphs with several hundred nodes mimicking a gene transcription interaction network in Escherichia coli. PMID:25493547
Detecting insect pollinator declines on regional and global scales
Lubuhn, Gretchen; Droege, Sam; Connor, Edward F.; Gemmill-Herren, Barbara; Potts, Simon G.; Minckley, Robert L.; Griswold, Terry; Jean, Robert; Kula, Emanuel; Roubik, David W.; Cane, Jim; Wright, Karen W.; Frankie, Gordon; Parker, Frank
2013-01-01
Recently there has been considerable concern about declines in bee communities in agricultural and natural habitats. The value of pollination to agriculture, provided primarily by bees, is >$200 billion/year worldwide, and in natural ecosystems it is thought to be even greater. However, no monitoring program exists to accurately detect declines in abundance of insect pollinators; thus, it is difficult to quantify the status of bee communities or estimate the extent of declines. We used data from 11 multiyear studies of bee communities to devise a program to monitor pollinators at regional, national, or international scales. In these studies, 7 different methods for sampling bees were used and bees were sampled on 3 different continents. We estimated that a monitoring program with 200-250 sampling locations each sampled twice over 5 years would provide sufficient power to detect small (2-5%) annual declines in the number of species and in total abundance and would cost U.S.$2,000,000. To detect declines as small as 1% annually over the same period would require >300 sampling locations. Given the role of pollinators in food security and ecosystem function, we recommend establishment of integrated regional and international monitoring programs to detect changes in pollinator communities.
NMR methods for metabolomics of mammalian cell culture bioreactors.
Aranibar, Nelly; Reily, Michael D
2014-01-01
Metabolomics has become an important tool for measuring pools of small molecules in mammalian cell cultures expressing therapeutic proteins. NMR spectroscopy has played an important role, largely because it requires minimal sample preparation, does not require chromatographic separation, and is quantitative. The concentrations of large numbers of small molecules in the extracellular media or within the cells themselves can be measured directly on the culture supernatant and on the supernatant of the lysed cells, respectively, and correlated with endpoints such as titer, cell viability, or glycosylation patterns. The observed changes can be used to generate hypotheses by which these parameters can be optimized. This chapter focuses on the sample preparation, data acquisition, and analysis to get the most out of NMR metabolomics data from CHO cell cultures but could easily be extended to other in vitro culture systems.
Ras activation by SOS: Allosteric regulation by altered fluctuation dynamics
Iversen, Lars; Tu, Hsiung-Lin; Lin, Wan-Chen; Christensen, Sune M.; Abel, Steven M.; Iwig, Jeff; Wu, Hung-Jen; Gureasko, Jodi; Rhodes, Christopher; Petit, Rebecca S.; Hansen, Scott D.; Thill, Peter; Yu, Cheng-Han; Stamou, Dimitrios; Chakraborty, Arup K.; Kuriyan, John; Groves, Jay T.
2014-01-01
Activation of the small guanosine triphosphatase H-Ras by the exchange factor Son of Sevenless (SOS) is an important hub for signal transduction. Multiple layers of regulation, through protein and membrane interactions, govern activity of SOS. We characterized the specific activity of individual SOS molecules catalyzing nucleotide exchange in H-Ras. Single-molecule kinetic traces revealed that SOS samples a broad distribution of turnover rates through stochastic fluctuations between distinct, long-lived (more than 100 seconds), functional states. The expected allosteric activation of SOS by Ras–guanosine triphosphate (GTP) was conspicuously absent in the mean rate. However, fluctuations into highly active states were modulated by Ras-GTP. This reveals a mechanism in which functional output may be determined by the dynamical spectrum of rates sampled by a small number of enzymes, rather than the ensemble average. PMID:24994643
Efficacy of deep biopsy for subepithelial lesions in the upper gastrointestinal tract.
Vaicekauskas, Rolandas; Stanaitis, Juozas; Valantinas, Jonas
2016-01-01
Accurate diagnosis of subepithelial lesions (SELs) in the gastrointestinal tract depends on a variety of methods: endoscopy, endoscopic ultrasound and different types of biopsy. Making an error-free diagnosis is vital for the subsequent application of an appropriate treatment. To evaluate the efficacy of deep biopsy via the endoscopic submucosal dissection (ESD) technique for SELs in the upper gastrointestinal tract. It was a case series study. Deep biopsy via the ESD technique was completed in 38 patients between November 2012 and October 2014. Thirty-eight SELs in the upper gastrointestinal tract of varying size (very small ≤ 1 cm, small 1-2 cm and large ≥ 2 cm) by means of the ESD technique after an incision with an electrosurgical knife of the overlying layers and revealing a small part of the lesion were biopsied under direct endoscopic view. Deep biopsy via the ESD technique was diagnostic in 28 of 38 patients (73.3%; 95% CI: 59.7-89.7%). The diagnostic yield for SELs with a clear endophytic shape increased to 91.3%. An evident endophytic appearance of a subepithelial lesion, the mean number of biopsied samples (6.65 ±1.36) and the total size in length of all samples per case (19.88 ±8.07 mm) were the main criteria influencing the positiveness of deep biopsy in the diagnostic group compared to the nondiagnostic one (p = 0.001; p = 0.025; p = 0.008). Deep biopsy via the ESD technique is an effective and safe method for the diagnosis of SELs especially with a clear endophytic appearance in a large number of biopsied samples.
Majumdar, Subhabrata; Basak, Subhash C
2018-04-26
Proper validation is an important aspect of QSAR modelling. External validation is one of the widely used validation methods in QSAR where the model is built on a subset of the data and validated on the rest of the samples. However, its effectiveness for datasets with a small number of samples but large number of predictors remains suspect. Calculating hundreds or thousands of molecular descriptors using currently available software has become the norm in QSAR research, owing to computational advances in the past few decades. Thus, for n chemical compounds and p descriptors calculated for each molecule, the typical chemometric dataset today has high value of p but small n (i.e. n < p). Motivated by the evidence of inadequacies of external validation in estimating the true predictive capability of a statistical model in recent literature, this paper performs an extensive and comparative study of this method with several other validation techniques. We compared four validation methods: leave-one-out, K-fold, external and multi-split validation, using statistical models built using the LASSO regression, which simultaneously performs variable selection and modelling. We used 300 simulated datasets and one real dataset of 95 congeneric amine mutagens for this evaluation. External validation metrics have high variation among different random splits of the data, hence are not recommended for predictive QSAR models. LOO has the overall best performance among all validation methods applied in our scenario. Results from external validation are too unstable for the datasets we analyzed. Based on our findings, we recommend using the LOO procedure for validating QSAR predictive models built on high-dimensional small-sample data. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
A universal TaqMan-based RT-PCR protocol for cost-efficient detection of small noncoding RNA.
Jung, Ulrike; Jiang, Xiaoou; Kaufmann, Stefan H E; Patzel, Volker
2013-12-01
Several methods for the detection of RNA have been developed over time. For small RNA detection, a stem-loop reverse primer-based protocol relying on TaqMan RT-PCR has been described. This protocol requires an individual specific TaqMan probe for each target RNA and, hence, is highly cost-intensive for experiments with small sample sizes or large numbers of different samples. We describe a universal TaqMan-based probe protocol which can be used to detect any target sequence and demonstrate its applicability for the detection of endogenous as well as artificial eukaryotic and bacterial small RNAs. While the specific and the universal probe-based protocol showed the same sensitivity, the absolute sensitivity of detection was found to be more than 100-fold lower for both than previously reported. In subsequent experiments, we found previously unknown limitations intrinsic to the method affecting its feasibility in determination of mature template RISC incorporation as well as in multiplexing. Both protocols were equally specific in discriminating between correct and incorrect small RNA targets or between mature miRNA and its unprocessed RNA precursor, indicating the stem-loop RT-primer, but not the TaqMan probe, triggers target specificity. The presented universal TaqMan-based RT-PCR protocol represents a cost-efficient method for the detection of small RNAs.
NASA Technical Reports Server (NTRS)
Kinard, J. T.
1975-01-01
The development of a procedure for obtaining data related to wear metal determinations in used lubricants is discussed. The procedure makes it possible to obtain rapid, simultaneous determinations of a number of wear metals at levels of parts per thousand to low parts per billion using a small amount of sample. The electrode assembly and instrumentation used in the process are described. Samples of data obtained from tests conducted under controlled conditions are tabulated.
Speil, Sidney
1974-01-01
The problems of quantitating chrysotile in water by fiber count techniques are reviewed briefly and the use of mass quantitation is suggested as a preferable measure. Chrysotile fiber has been found in almost every sample of natural water examined, but generally transmission electron miscroscopy (TEM) is required because of the small diameters involved. The extreme extrapolation required in mathematically converting a few fibers or fiber fragments under the TEM to the fiber content of a liquid sample casts considerable doubt on the validity of numbers used to compare chrysotile contents of different liquids. PMID:4470930
Image subsampling and point scoring approaches for large-scale marine benthic monitoring programs
NASA Astrophysics Data System (ADS)
Perkins, Nicholas R.; Foster, Scott D.; Hill, Nicole A.; Barrett, Neville S.
2016-07-01
Benthic imagery is an effective tool for quantitative description of ecologically and economically important benthic habitats and biota. The recent development of autonomous underwater vehicles (AUVs) allows surveying of spatial scales that were previously unfeasible. However, an AUV collects a large number of images, the scoring of which is time and labour intensive. There is a need to optimise the way that subsamples of imagery are chosen and scored to gain meaningful inferences for ecological monitoring studies. We examine the trade-off between the number of images selected within transects and the number of random points scored within images on the percent cover of target biota, the typical output of such monitoring programs. We also investigate the efficacy of various image selection approaches, such as systematic or random, on the bias and precision of cover estimates. We use simulated biotas that have varying size, abundance and distributional patterns. We find that a relatively small sampling effort is required to minimise bias. An increased precision for groups that are likely to be the focus of monitoring programs is best gained through increasing the number of images sampled rather than the number of points scored within images. For rare species, sampling using point count approaches is unlikely to provide sufficient precision, and alternative sampling approaches may need to be employed. The approach by which images are selected (simple random sampling, regularly spaced etc.) had no discernible effect on mean and variance estimates, regardless of the distributional pattern of biota. Field validation of our findings is provided through Monte Carlo resampling analysis of a previously scored benthic survey from temperate waters. We show that point count sampling approaches are capable of providing relatively precise cover estimates for candidate groups that are not overly rare. The amount of sampling required, in terms of both the number of images and number of points, varies with the abundance, size and distributional pattern of target biota. Therefore, we advocate either the incorporation of prior knowledge or the use of baseline surveys to establish key properties of intended target biota in the initial stages of monitoring programs.
Valenti, Daniela; de Bari, Lidia; De Filippis, Bianca; Ricceri, Laura; Vacca, Rosa Anna
2014-01-01
Studies of mitochondrial bioenergetics in brain pathophysiology are often precluded by the need to isolate mitochondria immediately after tissue dissection from a large number of brain biopsies for comparative studies. Here we present a procedure of cryopreservation of small brain areas from which mitochondrial enriched fractions (crude mitochondria) with high oxidative phosphorylation efficiency can be isolated. Small mouse brain areas were frozen and stored in a solution containing glycerol as cryoprotectant. Crude mitochondria were isolated by differential centrifugation from both cryopreserved and freshly explanted brain samples and were compared with respect to their ability to generate membrane potential and produce ATP. Intactness of outer and inner mitochondrial membranes was verified by polarographic ascorbate and cytochrome c tests and spectrophotometric assay of citrate synthase activity. Preservation of structural integrity and oxidative phosphorylation efficiency was successfully obtained in crude mitochondria isolated from different areas of cryopreserved mouse brain samples. Long-term cryopreservation of small brain areas from which intact and phosphorylating mitochondria can be isolated for the study of mitochondrial bioenergetics will significantly expand the study of mitochondrial defects in neurological pathologies, allowing large comparative studies and favoring interlaboratory and interdisciplinary analyses. Copyright © 2013 Elsevier Inc. All rights reserved.
Using GEO Optical Observations to Infer Orbit Populations
NASA Technical Reports Server (NTRS)
Matney, Mark; Africano, John
2002-01-01
NASA's Orbital Debris measurements program has a goal to characterize the small debris environment in the geosynchronous Earth-orbit (GEO) region using optical telescopes ("small" refers to objects too small to catalog and track with current systems). Traditionally, observations of GEO and near-GEO objects involve following the object with the telescope long enough to obtain an orbit. When observing very dim objects with small field-of-view telescopes, though, the observations are generally too short to obtain accurate orbital elements. However, it is possible to use such observations to statistically characterize the small object environment. A telescope pointed at a particular spot could potentially see objects in a number of different orbits. Inevitably, when looking at one region for certain types of orbits, there are objects in other types of orbits that cannot be seen. Observation campaigns are designed with these limitations in mind and are set up to span a number of regions of the sky, making it possible to sample all potential orbits under consideration. Each orbit is not seen with the same probability, however, so there are observation biases intrinsic to any observation campaign. Fortunately, it is possible to remove such biases and reconstruct a meaningful estimate of the statistical orbit populations of small objects in GEO. This information, in turn, can be used to investigate the nature of debris sources and to characterize the risk to GEO spacecraft. This paper describes these statistical tools and presents estimates of small object GEO populations.
Gabbett, Tim J; Abernethy, Bruce; Jenkins, David G
2012-02-01
The purpose of this study was to investigate the effect of changes in field size on the physiological and skill demands of small-sided games in elite junior and senior rugby league players. Sixteen elite senior rugby league players ([mean ± SE] age, 23.6 ± 0.5 years) and 16 elite junior rugby league players ([mean ± SE] age, 17.3 ± 0.3 years) participated in this study. On day 1, 2 teams played an 8-minute small-sided game on a small field (10-m width × 40-m length), whereas the remaining 2 teams played the small-sided game on a larger sized field (40-m width × 70-m length). On day 2, the groups were crossed over. Movement was recorded by a global positioning system unit sampling at 5 Hz. Games were filmed to count the number of possessions and the number and quality of disposals. The games played on a larger field resulted in a greater (p < 0.05) total distance covered, and distances covered in moderate, high, and very-high velocity movement intensities. Senior players covered more distance at moderate, high, and very-high intensities, and less distance at low and very-low intensities during small-sided games than junior players. Although increasing field size had no significant influence (p > 0.05) over the duration of recovery periods for junior players, larger field size significantly reduced (p < 0.05) the amount of short-, moderate-, and long-duration recovery periods in senior players. No significant between-group differences (p > 0.05) were detected for games played on a small or large field for the number or quality of skill involvements. These results suggest that increases in field size serve to increase the physiological demands of small-sided games but have minimal influence over the volume or quality of skill executions in elite rugby league players.
How to justify small-refinery info/control system modernization
DOE Office of Scientific and Technical Information (OSTI.GOV)
Haskins, D.E.
1993-05-01
Information and control systems modernization can be justified by successful implementation of advanced process control (APC) in nearly all refineries, even the small ones. However, the small refineries require special solutions to meet the challenges of limited resources in both finance and manpower. Based on a number of case studies, a typical small refinery as it operates today is described. A sample information and control system modernization plan is described and the typical cost and benefits show how the project cost can be justified. Business objectives of an HPI plant are to satisfy customers by providing specific products, to satisfymore » the owners by maximizing profits and to satisfy the public by being safe and environmentally correct. Managers have always tried to meet these objectives with functions for the total plant.« less
Quantification of Protozoa and Viruses from Small Water Volumes
Bonilla, J. Alfredo; Bonilla, Tonya D.; Abdelzaher, Amir M.; Scott, Troy M.; Lukasik, Jerzy; Solo-Gabriele, Helena M.; Palmer, Carol J.
2015-01-01
Large sample volumes are traditionally required for the analysis of waterborne pathogens. The need for large volumes greatly limits the number of samples that can be processed. The goals of this study were to compare extraction and detection procedures for quantifying protozoan parasites and viruses from small volumes of marine water. The intent was to evaluate a logistically simpler method of sample collection and processing that would facilitate direct pathogen measures as part of routine monitoring programs. Samples were collected simultaneously using a bilayer device with protozoa capture by size (top filter) and viruses capture by charge (bottom filter). Protozoan detection technologies utilized for recovery of Cryptosporidium spp. and Giardia spp. were qPCR and the more traditional immunomagnetic separation—IFA-microscopy, while virus (poliovirus) detection was based upon qPCR versus plaque assay. Filters were eluted using reagents consistent with the downstream detection technologies. Results showed higher mean recoveries using traditional detection methods over qPCR for Cryptosporidium (91% vs. 45%) and poliovirus (67% vs. 55%) whereas for Giardia the qPCR-based methods were characterized by higher mean recoveries (41% vs. 28%). Overall mean recoveries are considered high for all detection technologies. Results suggest that simultaneous filtration may be suitable for isolating different classes of pathogens from small marine water volumes. More research is needed to evaluate the suitability of this method for detecting pathogens at low ambient concentration levels. PMID:26114244
Quantification of Protozoa and Viruses from Small Water Volumes.
Bonilla, J Alfredo; Bonilla, Tonya D; Abdelzaher, Amir M; Scott, Troy M; Lukasik, Jerzy; Solo-Gabriele, Helena M; Palmer, Carol J
2015-06-24
Large sample volumes are traditionally required for the analysis of waterborne pathogens. The need for large volumes greatly limits the number of samples that can be processed. The aims of this study were to compare extraction and detection procedures for quantifying protozoan parasites and viruses from small volumes of marine water. The intent was to evaluate a logistically simpler method of sample collection and processing that would facilitate direct pathogen measures as part of routine monitoring programs. Samples were collected simultaneously using a bilayer device with protozoa capture by size (top filter) and viruses capture by charge (bottom filter). Protozoan detection technologies utilized for recovery of Cryptosporidium spp. and Giardia spp. were qPCR and the more traditional immunomagnetic separation-IFA-microscopy, while virus (poliovirus) detection was based upon qPCR versus plaque assay. Filters were eluted using reagents consistent with the downstream detection technologies. Results showed higher mean recoveries using traditional detection methods over qPCR for Cryptosporidium (91% vs. 45%) and poliovirus (67% vs. 55%) whereas for Giardia the qPCR-based methods were characterized by higher mean recoveries (41% vs. 28%). Overall mean recoveries are considered high for all detection technologies. Results suggest that simultaneous filtration may be suitable for isolating different classes of pathogens from small marine water volumes. More research is needed to evaluate the suitability of this method for detecting pathogens at low ambient concentration levels.
Development of source specific diatom lipids biomarkers as Antarctic Sea Ice proxies
NASA Astrophysics Data System (ADS)
Smik, Lukas; Belt, Simon T.; Brown, Thomas A.; Lieser, Jan L.; Armand, Leanne K.; Leventer, Amy; Allen, Claire S.
2016-04-01
C25 highly branched isoprenoid (HBI) are lipid biomarkers biosynthesised by a relatively small number of diatom genera, but are, nonetheless, common constituents of global marine sediments. The occurrence and variable abundance of certain C25 highly branched isoprenoid (HBI) biomarkers in Antarctic marine sediments has previously been proposed as a proxy measure of paleo sea-ice extent in the Southern Ocean and a small number of paleo sea-ice reconstructions based on the variable abundances of these HBIs have appeared in recent years. However, the development of HBIs as proxies for Antarctic sea ice is much less advanced than that for IP25 (another HBI) in the Arctic and has been based on relatively small number of analyses in sea ice, water column and sediment samples. To provide further insights into the use of these HBIs as proxies for Antarctic sea ice, we here describe an assessment of their distributions in surface water, surface sediment and sea ice samples collected from a number of Antarctic locations experiencing contrasting sea ice conditions in recent years. Our study shows that distributions of a di-unsaturated HBI (diene II) and tri-unsaturated HBI (triene III) in surface water samples were found to be extremely sensitive to the local sea-ice conditions, with diene II detected for sampling sites that experienced seasonal sea ice and highest concentrations found in coastal locations with longer-lasting ice cover and a recurrent polynya. In contrast, triene III was observed in all of the samples analysed, but with highest concentrations within the region of the retreating sea ice edge, an observation consistent with significant environmental control over the biosynthesis of diene II and triene III by sea ice diatoms and open water phytoplankton, respectively. However, additional local factors, such as those associated with polynya formation, may also exert some control over the distribution of triene III and the relative concentrations of diene II and triene III, in particular. This may have important implications for the use of these biomarkers for paleo sea ice reconstructions. Sedimentary distribution showed significant variation in abundances of diene II and triene III between different regions of Antarctica, but also on a more local scale, potentially reflecting a high degree of sensitivity towards individual sea ice dynamics that favour the individual species responsible for their biosynthesis. However, highest concentrations of diene II were generally observed in near coastal locations, consistent with the identification of elevated abundances of this HBI in first year or land fast ice in these settings. The identification of the sea ice diatom source of diene II will likely be significant in interpretations of the occurrence of this biomarker in paleo sea ice records.
Impact of spatial variability and sampling design on model performance
NASA Astrophysics Data System (ADS)
Schrape, Charlotte; Schneider, Anne-Kathrin; Schröder, Boris; van Schaik, Loes
2017-04-01
Many environmental physical and chemical parameters as well as species distributions display a spatial variability at different scales. In case measurements are very costly in labour time or money a choice has to be made between a high sampling resolution at small scales and a low spatial cover of the study area or a lower sampling resolution at the small scales resulting in local data uncertainties with a better spatial cover of the whole area. This dilemma is often faced in the design of field sampling campaigns for large scale studies. When the gathered field data are subsequently used for modelling purposes the choice of sampling design and resulting data quality influence the model performance criteria. We studied this influence with a virtual model study based on a large dataset of field information on spatial variation of earthworms at different scales. Therefore we built a virtual map of anecic earthworm distributions over the Weiherbach catchment (Baden-Württemberg in Germany). First of all the field scale abundance of earthworms was estimated using a catchment scale model based on 65 field measurements. Subsequently the high small scale variability was added using semi-variograms, based on five fields with a total of 430 measurements divided in a spatially nested sampling design over these fields, to estimate the nugget, range and standard deviation of measurements within the fields. With the produced maps, we performed virtual samplings of one up to 50 random points per field. We then used these data to rebuild the catchment scale models of anecic earthworm abundance with the same model parameters as in the work by Palm et al. (2013). The results of the models show clearly that a large part of the non-explained deviance of the models is due to the very high small scale variability in earthworm abundance: the models based on single virtual sampling points on average obtain an explained deviance of 0.20 and a correlation coefficient of 0.64. With increasing sampling points per field, we averaged the measured abundance of the sampling within each field to obtain a more representative value of the field average. Doubling the samplings per field strongly improved the model performance criteria (explained deviance 0.38 and correlation coefficient 0.73). With 50 sampling points per field the performance criteria were 0.91 and 0.97 respectively for explained deviance and correlation coefficient. The relationship between number of samplings and performance criteria can be described with a saturation curve. Beyond five samples per field the model improvement becomes rather small. With this contribution we wish to discuss the impact of data variability at sampling scale on model performance and the implications for sampling design and assessment of model results as well as ecological inferences.
The Effect of Small Sample Size on Two-Level Model Estimates: A Review and Illustration
ERIC Educational Resources Information Center
McNeish, Daniel M.; Stapleton, Laura M.
2016-01-01
Multilevel models are an increasingly popular method to analyze data that originate from a clustered or hierarchical structure. To effectively utilize multilevel models, one must have an adequately large number of clusters; otherwise, some model parameters will be estimated with bias. The goals for this paper are to (1) raise awareness of the…
Going to Scale: A Nonrandomized Nationwide Trial of the KiVa Antibullying Program for Grades 1-9
ERIC Educational Resources Information Center
Karna, Antti; Voeten, Marinus; Little, Todd D.; Poskiparta, Elisa; Alanen, Erkki; Salmivalli, Christina
2011-01-01
Objective: The effects of school-based antibullying programs have typically been examined on small samples, with number of schools ranging from 1 to 78 (Farrington & Ttofi, 2009). This study investigated the effectiveness of the KiVa antibullying program in the beginning of its nationwide implementation in Finland. Method: At each time point,…
ERIC Educational Resources Information Center
Patry, Marc W.; Magaletta, Philip R.; Diamond, Pamela M.; Weinman, Beth A.
2011-01-01
Although not originally designed for implementation in correctional settings, researchers and clinicians have begun to use the Personality Assessment Inventory (PAI) to assess offenders. A relatively small number of studies have made attempts to validate the alcohol and drug abuse scales of the PAI, and only a very few studies have validated those…
ERIC Educational Resources Information Center
Schmidt, Jennifer A.; Shumow, Lee; Kackar, Hayal Z.
2012-01-01
Youth who participate in service activities differ from those who do not on a number of key demographic characteristics like socio-economic status and other indicators of risk; and most studies demonstrating positive outcomes among service participants employ small non-representative samples. Thus, there is little evidence as to whether the…
Unbiased multi-fidelity estimate of failure probability of a free plane jet
NASA Astrophysics Data System (ADS)
Marques, Alexandre; Kramer, Boris; Willcox, Karen; Peherstorfer, Benjamin
2017-11-01
Estimating failure probability related to fluid flows is a challenge because it requires a large number of evaluations of expensive models. We address this challenge by leveraging multiple low fidelity models of the flow dynamics to create an optimal unbiased estimator. In particular, we investigate the effects of uncertain inlet conditions in the width of a free plane jet. We classify a condition as failure when the corresponding jet width is below a small threshold, such that failure is a rare event (failure probability is smaller than 0.001). We estimate failure probability by combining the frameworks of multi-fidelity importance sampling and optimal fusion of estimators. Multi-fidelity importance sampling uses a low fidelity model to explore the parameter space and create a biasing distribution. An unbiased estimate is then computed with a relatively small number of evaluations of the high fidelity model. In the presence of multiple low fidelity models, this framework offers multiple competing estimators. Optimal fusion combines all competing estimators into a single estimator with minimal variance. We show that this combined framework can significantly reduce the cost of estimating failure probabilities, and thus can have a large impact in fluid flow applications. This work was funded by DARPA.
Small Scale Biodiversity of an Alkaline Hot Spring in Yellowstone National Park
NASA Astrophysics Data System (ADS)
Walther, K.; Oiler, J.; Meyer-Dombard, D. R.
2012-12-01
To date, many phylogenetic diversity studies have been conducted in Yellowstone National Park (YNP) [1-7] focusing on the amplification of the 16S rRNA gene and "metagenomic" datasets. However, few reports focus on diversity at small scales. Here, we report on a small scale biodiversity study of sediment and biofilm communities within a confined area of a YNP hot spring, compare and contrast these communities to other sediment and biofilm communities from previous studies [1-7], and with other sediment and biofilm communities in the same system. Sediment and biofilm samples were collected, using a 30 x 50 cm sampling grid divided in 5 x 5 cm squares, which was placed in the outflow channel of "Bat Pool", an alkaline (pH 7.9) hot spring in YNP. Accompanying geochemical data included a full range of spectrophotometry measurements along with major ions, trace elements, and DIC/DOC. In addition, in situ temperature and conductivity arrays were placed within the grid location. The temperature array closest to the source varied between 83-88°C, while the temperature array 40 cm downstream varied between ~83.5-86.5°C. The two conductivity arrays yielded measurements of 5632 μS and 5710 μS showing little variation within the sampling area. Within the grid space, DO ranged from 0.5-1.33 mg/L, with relatively similar, but slightly lower values down the outflow channel. Sulfide values within the grid ranged from 1020-1671 μg/L, while sulfide values outside of the grid region fluctuated, but generally followed the trend of decreasing from source down the outflow. Despite the relative heterogeneity of chemical and physical parameters in the grid space, there was biological diversity in sediments and biofilms at the 5 cm scale. Small scale biodiversity was analyzed by selecting a representative number of samples from within the grid. DNA was extracted and variable regions V3 and V6 (Archaea and Bacteria, respectively) were sequenced with 454 pyrosequencing. The datasets from each of the samples were randomly subsampled and the same number of sequences was taken from each dataset so that the samples could be directly compared. Using the Ribosomal Database Project Pyrosequencing Pipeline (http://rdp.cme.msu.edu/), the sequences were aligned, complete linkage clustering was performed, Shannon and Chao1 indices were calculated, and rarefaction curves were made. The RDP Classifier tool afforded classification in a taxonomical hierarchy and the samples were compared on the order level to determine the variation of the microbial communities within the sampling grid. Additional alpha and beta diversity indices were also established. Through comparing the samples on the order level, it was determined that there is variation within a small sampling area despite similar geochemical and temperature conditions at the time of sampling. This variation is seen in both the sediment and biofilm communities, primarily seen among Bacteria. [1] Barns, S.M. et al. (1994) PNAS. 91: 1609-1613. [2] Barns, S.M. et al. (1996) PNAS. 93: 9188-9193. [3] Hall, J.R. et al. (2008) AEM. 74(15): 4910-4922. [4] Hugenholtz, P. et al. (1998) JofBac. 180(2): 366-376. [5] Meyer-Dombard, D. R. et al. (2005) Geobio. 3: 211-227. [6] Meyer-Dombard, D.R. et al. (2011) EM. 13(8): 2216-2231. [7] Reysenbach, A.L. et al. (1994) AEM. 60 (6): 2113-2119.
Profeta, Gerson S.; Pereira, Jessica A. S.; Costa, Samara G.; Azambuja, Patricia; Garcia, Eloi S.; Moraes, Caroline da Silva; Genta, Fernando A.
2017-01-01
Glycoside Hydrolases (GHs) are enzymes able to recognize and cleave glycosidic bonds. Insect GHs play decisive roles in digestion, in plant-herbivore, and host-pathogen interactions. GH activity is normally measured by the detection of a release from the substrate of products as sugars units, colored, or fluorescent groups. In most cases, the conditions for product release and detection differ, resulting in discontinuous assays. The current protocols result in using large amounts of reaction mixtures for the obtainment of time points in each experimental replica. These procedures restrain the analysis of biological materials with limited amounts of protein and, in the case of studies regarding small insects, implies in the pooling of samples from several individuals. In this respect, most studies do not assess the variability of GH activities across the population of individuals from the same species. The aim of this work is to approach this technical problem and have a deeper understanding of the variation of GH activities in insect populations, using as models the disease vectors Rhodnius prolixus (Hemiptera: Triatominae) and Lutzomyia longipalpis (Diptera: Phlebotominae). Here we standardized continuous assays using 4-methylumbelliferyl derived substrates for the detection of α-Glucosidase, β-Glucosidase, α-Mannosidase, N-acetyl-hexosaminidase, β-Galactosidase, and α-Fucosidase in the midgut of R. prolixus and L. longipalpis with results similar to the traditional discontinuous protocol. The continuous assays allowed us to measure GH activities using minimal sample amounts with a higher number of measurements, resulting in data that are more reliable and less time and reagent consumption. The continuous assay also allows the high-throughput screening of GH activities in small insect samples, which would be not applicable to the previous discontinuous protocol. We applied continuous GH measurements to 90 individual samples of R. prolixus anterior midgut homogenates using a high-throughput protocol. α-Glucosidase and α-Mannosidase activities showed the normal distribution in the population. β-Glucosidase, β-Galactosidase, N-acetyl-hexosaminidase, and α-Fucosidase activities showed non-normal distributions. These results indicate that GHs fluorescent-based high-throughput assays apply to insect samples and that the frequency distribution of digestive activities should be considered in data analysis, especially if a small number of samples is used. PMID:28553236
Lowe, Terrence (Peter); Tebbs, Kerry; Sparling, Donald W.
2016-01-01
Three types of macroinvertebrate collecting devices, Gerking box traps, D-shaped sweep nets, and activity traps, have commonly been used to sample macroinvertebrates when conducting rapid biological assessments of North American wetlands. We compared collections of macroinvertebrates identified to the family level made with these devices in 6 constructed and 2 natural wetlands on the Delmarva Peninsula of Maryland. We also assessed their potential efficacy in comparisons among wetlands using several proportional and richness attributes. Differences in median diversity among samples from the 3 devices were significant; the sweep-net samples had the greatest diversity and the activity-trap samples had the least diversity. Differences in median abundance were not significant between the Gerking box-trap samples and sweep-net samples, but median abundance among activity-trap samples was significantly lower than among samples of the other 2 devices. Within samples, the proportions of median diversity composed of major class and order groupings were similar among the 3 devices. However the proportions of median abundance composed of the major class and order groupings within activity-trap samples were not similar to those of the other 2 devices. There was a slight but significant increase in the total number of families captured when we combined activity-trap samples with Gerking box-trap samples or with sweep-net samples, and the per-sample median numbers of families of the combined activity-trap and sweep-net samples was significantly higher than that of the combined activity-trap and Gerking box-trap samples. We detected significant differences among wetlands for 4 macroinvertebrate attributes with the Gerking box-trap data, 6 attributes with sweep-net data, and 5 attributes with the activity-trap data. A small, but significant increase in the number of attributes showing differences among wetlands occurred when we combined activity-trap samples with those of the Gerking boxtrap or sweep net.
Salmonella Typhimurium DT193 and DT99 are present in great and blue tits in Flanders, Belgium
Verbrugghe, E.; Dekeukeleire, D.; De Beelde, R.; Rouffaer, L. O.; Haesendonck, R.; Strubbe, D.; Mattheus, W.; Bertrand, S.; Pasmans, F.; Bonte, D.; Verheyen, K.; Lens, L.; Martel, A.
2017-01-01
Endemic infections with the common avian pathogen Salmonella enterica subspecies enterica serovar Typhimurium (Salmonella Typhimurium) may incur a significant cost on the host population. In this study, we determined the potential of endemic Salmonella infections to reduce the reproductive success of blue (Cyanistes caeruleus) and great (Parus major) tits by correlating eggshell infection with reproductive parameters. The fifth egg of each clutch was collected from nest boxes in 19 deciduous forest fragments. Out of the 101 sampled eggs, 7 Salmonella Typhimurium isolates were recovered. The low bacterial prevalence was reflected by a similarly low serological prevalence in the fledglings. In this study with a relatively small sample size, presence of Salmonella did not affect reproductive parameters (egg volume, clutch size, number of nestlings and number of fledglings), nor the health status of the fledglings. However, in order to clarify the impact on health and reproduction a larger number of samples have to be analyzed. Phage typing showed that the isolates belonged to the definitive phage types (DT) 193 and 99, and multi-locus variable number tandem repeat analysis (MLVA) demonstrated a high similarity among the tit isolates, but distinction to human isolates. These findings suggest the presence of passerine-adapted Salmonella strains in free-ranging tit populations with host pathogen co-existence. PMID:29112955
Validating a biometric authentication system: sample size requirements.
Dass, Sarat C; Zhu, Yongfang; Jain, Anil K
2006-12-01
Authentication systems based on biometric features (e.g., fingerprint impressions, iris scans, human face images, etc.) are increasingly gaining widespread use and popularity. Often, vendors and owners of these commercial biometric systems claim impressive performance that is estimated based on some proprietary data. In such situations, there is a need to independently validate the claimed performance levels. System performance is typically evaluated by collecting biometric templates from n different subjects, and for convenience, acquiring multiple instances of the biometric for each of the n subjects. Very little work has been done in 1) constructing confidence regions based on the ROC curve for validating the claimed performance levels and 2) determining the required number of biometric samples needed to establish confidence regions of prespecified width for the ROC curve. To simplify the analysis that address these two problems, several previous studies have assumed that multiple acquisitions of the biometric entity are statistically independent. This assumption is too restrictive and is generally not valid. We have developed a validation technique based on multivariate copula models for correlated biometric acquisitions. Based on the same model, we also determine the minimum number of samples required to achieve confidence bands of desired width for the ROC curve. We illustrate the estimation of the confidence bands as well as the required number of biometric samples using a fingerprint matching system that is applied on samples collected from a small population.
Effects of Sample Selection Bias on the Accuracy of Population Structure and Ancestry Inference
Shringarpure, Suyash; Xing, Eric P.
2014-01-01
Population stratification is an important task in genetic analyses. It provides information about the ancestry of individuals and can be an important confounder in genome-wide association studies. Public genotyping projects have made a large number of datasets available for study. However, practical constraints dictate that of a geographical/ethnic population, only a small number of individuals are genotyped. The resulting data are a sample from the entire population. If the distribution of sample sizes is not representative of the populations being sampled, the accuracy of population stratification analyses of the data could be affected. We attempt to understand the effect of biased sampling on the accuracy of population structure analysis and individual ancestry recovery. We examined two commonly used methods for analyses of such datasets, ADMIXTURE and EIGENSOFT, and found that the accuracy of recovery of population structure is affected to a large extent by the sample used for analysis and how representative it is of the underlying populations. Using simulated data and real genotype data from cattle, we show that sample selection bias can affect the results of population structure analyses. We develop a mathematical framework for sample selection bias in models for population structure and also proposed a correction for sample selection bias using auxiliary information about the sample. We demonstrate that such a correction is effective in practice using simulated and real data. PMID:24637351
Chaibub Neto, Elias
2015-01-01
In this paper we propose a vectorized implementation of the non-parametric bootstrap for statistics based on sample moments. Basically, we adopt the multinomial sampling formulation of the non-parametric bootstrap, and compute bootstrap replications of sample moment statistics by simply weighting the observed data according to multinomial counts instead of evaluating the statistic on a resampled version of the observed data. Using this formulation we can generate a matrix of bootstrap weights and compute the entire vector of bootstrap replications with a few matrix multiplications. Vectorization is particularly important for matrix-oriented programming languages such as R, where matrix/vector calculations tend to be faster than scalar operations implemented in a loop. We illustrate the application of the vectorized implementation in real and simulated data sets, when bootstrapping Pearson’s sample correlation coefficient, and compared its performance against two state-of-the-art R implementations of the non-parametric bootstrap, as well as a straightforward one based on a for loop. Our investigations spanned varying sample sizes and number of bootstrap replications. The vectorized bootstrap compared favorably against the state-of-the-art implementations in all cases tested, and was remarkably/considerably faster for small/moderate sample sizes. The same results were observed in the comparison with the straightforward implementation, except for large sample sizes, where the vectorized bootstrap was slightly slower than the straightforward implementation due to increased time expenditures in the generation of weight matrices via multinomial sampling. PMID:26125965
Estimation of the bottleneck size in Florida panthers
Culver, M.; Hedrick, P.W.; Murphy, K.; O'Brien, S.; Hornocker, M.G.
2008-01-01
We have estimated the extent of genetic variation in museum (1890s) and contemporary (1980s) samples of Florida panthers Puma concolor coryi for both nuclear loci and mtDNA. The microsatellite heterozygosity in the contemporary sample was only 0.325 that in the museum samples although our sample size and number of loci are limited. Support for this estimate is provided by a sample of 84 microsatellite loci in contemporary Florida panthers and Idaho pumas Puma concolor hippolestes in which the contemporary Florida panther sample had only 0.442 the heterozygosity of Idaho pumas. The estimated diversities in mtDNA in the museum and contemporary samples were 0.600 and 0.000, respectively. Using a population genetics approach, we have estimated that to reduce either the microsatellite heterozygosity or the mtDNA diversity this much (in a period of c. 80years during the 20th century when the numbers were thought to be low) that a very small bottleneck size of c. 2 for several generations and a small effective population size in other generations is necessary. Using demographic data from Yellowstone pumas, we estimated the ratio of effective to census population size to be 0.315. Using this ratio, the census population size in the Florida panthers necessary to explain the loss of microsatellite variation was c .41 for the non-bottleneck generations and 6.2 for the two bottleneck generations. These low bottleneck population sizes and the concomitant reduced effectiveness of selection are probably responsible for the high frequency of several detrimental traits in Florida panthers, namely undescended testicles and poor sperm quality. The recent intensive monitoring both before and after the introduction of Texas pumas in 1995 will make the recovery and genetic restoration of Florida panthers a classic study of an endangered species. Our estimates of the bottleneck size responsible for the loss of genetic variation in the Florida panther completes an unknown aspect of this account. ?? 2008 The Authors. Journal compilation ?? 2008 The Zoological Society of London.
High-throughput biological small-angle X-ray scattering with a robotically loaded capillary cell
Nielsen, S. S.; Møller, M.; Gillilan, R. E.
2012-01-01
With the rise in popularity of biological small-angle X-ray scattering (BioSAXS) measurements, synchrotron beamlines are confronted with an ever-increasing number of samples from a wide range of solution conditions. To meet these demands, an increasing number of beamlines worldwide have begun to provide automated liquid-handling systems for sample loading. This article presents an automated sample-loading system for BioSAXS beamlines, which combines single-channel disposable-tip pipetting with a vacuum-enclosed temperature-controlled capillary flow cell. The design incorporates an easily changeable capillary to reduce the incidence of X-ray window fouling and cross contamination. Both the robot-control and the data-processing systems are written in Python. The data-processing code, RAW, has been enhanced with several new features to form a user-friendly BioSAXS pipeline for the robot. The flow cell also supports efficient manual loading and sample recovery. An effective rinse protocol for the sample cell is developed and tested. Fluid dynamics within the sample capillary reveals a vortex ring pattern of circulation that redistributes radiation-damaged material. Radiation damage is most severe in the boundary layer near the capillary surface. At typical flow speeds, capillaries below 2 mm in diameter are beginning to enter the Stokes (creeping flow) regime in which mixing due to oscillation is limited. Analysis within this regime shows that single-pass exposure and multiple-pass exposure of a sample plug are functionally the same with regard to exposed volume when plug motion reversal is slow. The robot was tested on three different beamlines at the Cornell High-Energy Synchrotron Source, with a variety of detectors and beam characteristics, and it has been used successfully in several published studies as well as in two introductory short courses on basic BioSAXS methods. PMID:22509071
Fuzzy support vector machine for microarray imbalanced data classification
NASA Astrophysics Data System (ADS)
Ladayya, Faroh; Purnami, Santi Wulan; Irhamah
2017-11-01
DNA microarrays are data containing gene expression with small sample sizes and high number of features. Furthermore, imbalanced classes is a common problem in microarray data. This occurs when a dataset is dominated by a class which have significantly more instances than the other minority classes. Therefore, it is needed a classification method that solve the problem of high dimensional and imbalanced data. Support Vector Machine (SVM) is one of the classification methods that is capable of handling large or small samples, nonlinear, high dimensional, over learning and local minimum issues. SVM has been widely applied to DNA microarray data classification and it has been shown that SVM provides the best performance among other machine learning methods. However, imbalanced data will be a problem because SVM treats all samples in the same importance thus the results is bias for minority class. To overcome the imbalanced data, Fuzzy SVM (FSVM) is proposed. This method apply a fuzzy membership to each input point and reformulate the SVM such that different input points provide different contributions to the classifier. The minority classes have large fuzzy membership so FSVM can pay more attention to the samples with larger fuzzy membership. Given DNA microarray data is a high dimensional data with a very large number of features, it is necessary to do feature selection first using Fast Correlation based Filter (FCBF). In this study will be analyzed by SVM, FSVM and both methods by applying FCBF and get the classification performance of them. Based on the overall results, FSVM on selected features has the best classification performance compared to SVM.
NASA Astrophysics Data System (ADS)
Ruf, B.; Erdnuess, B.; Weinmann, M.
2017-08-01
With the emergence of small consumer Unmanned Aerial Vehicles (UAVs), the importance and interest of image-based depth estimation and model generation from aerial images has greatly increased in the photogrammetric society. In our work, we focus on algorithms that allow an online image-based dense depth estimation from video sequences, which enables the direct and live structural analysis of the depicted scene. Therefore, we use a multi-view plane-sweep algorithm with a semi-global matching (SGM) optimization which is parallelized for general purpose computation on a GPU (GPGPU), reaching sufficient performance to keep up with the key-frames of input sequences. One important aspect to reach good performance is the way to sample the scene space, creating plane hypotheses. A small step size between consecutive planes, which is needed to reconstruct details in the near vicinity of the camera may lead to ambiguities in distant regions, due to the perspective projection of the camera. Furthermore, an equidistant sampling with a small step size produces a large number of plane hypotheses, leading to high computational effort. To overcome these problems, we present a novel methodology to directly determine the sampling points of plane-sweep algorithms in image space. The use of the perspective invariant cross-ratio allows us to derive the location of the sampling planes directly from the image data. With this, we efficiently sample the scene space, achieving higher sampling density in areas which are close to the camera and a lower density in distant regions. We evaluate our approach on a synthetic benchmark dataset for quantitative evaluation and on a real-image dataset consisting of aerial imagery. The experiments reveal that an inverse sampling achieves equal and better results than a linear sampling, with less sampling points and thus less runtime. Our algorithm allows an online computation of depth maps for subsequences of five frames, provided that the relative poses between all frames are given.
Quality control and quality assurance in genotypic data for genome-wide association studies
Laurie, Cathy C.; Doheny, Kimberly F.; Mirel, Daniel B.; Pugh, Elizabeth W.; Bierut, Laura J.; Bhangale, Tushar; Boehm, Frederick; Caporaso, Neil E.; Cornelis, Marilyn C.; Edenberg, Howard J.; Gabriel, Stacy B.; Harris, Emily L.; Hu, Frank B.; Jacobs, Kevin; Kraft, Peter; Landi, Maria Teresa; Lumley, Thomas; Manolio, Teri A.; McHugh, Caitlin; Painter, Ian; Paschall, Justin; Rice, John P.; Rice, Kenneth M.; Zheng, Xiuwen; Weir, Bruce S.
2011-01-01
Genome-wide scans of nucleotide variation in human subjects are providing an increasing number of replicated associations with complex disease traits. Most of the variants detected have small effects and, collectively, they account for a small fraction of the total genetic variance. Very large sample sizes are required to identify and validate findings. In this situation, even small sources of systematic or random error can cause spurious results or obscure real effects. The need for careful attention to data quality has been appreciated for some time in this field, and a number of strategies for quality control and quality assurance (QC/QA) have been developed. Here we extend these methods and describe a system of QC/QA for genotypic data in genome-wide association studies. This system includes some new approaches that (1) combine analysis of allelic probe intensities and called genotypes to distinguish gender misidentification from sex chromosome aberrations, (2) detect autosomal chromosome aberrations that may affect genotype calling accuracy, (3) infer DNA sample quality from relatedness and allelic intensities, (4) use duplicate concordance to infer SNP quality, (5) detect genotyping artifacts from dependence of Hardy-Weinberg equilibrium (HWE) test p-values on allelic frequency, and (6) demonstrate sensitivity of principal components analysis (PCA) to SNP selection. The methods are illustrated with examples from the ‘Gene Environment Association Studies’ (GENEVA) program. The results suggest several recommendations for QC/QA in the design and execution of genome-wide association studies. PMID:20718045
An adaptive multi-level simulation algorithm for stochastic biological systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lester, C., E-mail: lesterc@maths.ox.ac.uk; Giles, M. B.; Baker, R. E.
2015-01-14
Discrete-state, continuous-time Markov models are widely used in the modeling of biochemical reaction networks. Their complexity often precludes analytic solution, and we rely on stochastic simulation algorithms (SSA) to estimate system statistics. The Gillespie algorithm is exact, but computationally costly as it simulates every single reaction. As such, approximate stochastic simulation algorithms such as the tau-leap algorithm are often used. Potentially computationally more efficient, the system statistics generated suffer from significant bias unless tau is relatively small, in which case the computational time can be comparable to that of the Gillespie algorithm. The multi-level method [Anderson and Higham, “Multi-level Montemore » Carlo for continuous time Markov chains, with applications in biochemical kinetics,” SIAM Multiscale Model. Simul. 10(1), 146–179 (2012)] tackles this problem. A base estimator is computed using many (cheap) sample paths at low accuracy. The bias inherent in this estimator is then reduced using a number of corrections. Each correction term is estimated using a collection of paired sample paths where one path of each pair is generated at a higher accuracy compared to the other (and so more expensive). By sharing random variables between these paired paths, the variance of each correction estimator can be reduced. This renders the multi-level method very efficient as only a relatively small number of paired paths are required to calculate each correction term. In the original multi-level method, each sample path is simulated using the tau-leap algorithm with a fixed value of τ. This approach can result in poor performance when the reaction activity of a system changes substantially over the timescale of interest. By introducing a novel adaptive time-stepping approach where τ is chosen according to the stochastic behaviour of each sample path, we extend the applicability of the multi-level method to such cases. We demonstrate the efficiency of our method using a number of examples.« less
Min-su, Han
2013-08-01
This paper aims at identifying the provenance of Goryeo celadons by understanding its microstructural characteristics, such as particles, blisters, forms and amount of pores, and the presence of crystal formation, bodies, and glazes and its boundary, using an optical microscope and scanning electron microscopy (SEM). The analysis of the reproduced samples shows that the glazed layer of the sherd fired at higher temperatures has lower viscosity and therefore it encourages the blisters to be combined together and the layer to become more transparent. In addition, the result showed that the vitrification and melting process of clay minerals such as feldspars and quartzs on the bodies was accelerated for those samples. To factor such characteristics of the microstructure and apply it to the sherds, the samples could be divided into six categories based on status, such as small particles with many small pores or mainly large and small circular pores in the bodies, only a limited number of varied sized blisters in the glazes, and a few blisters and needle-shaped crystals on the boundary surface. In conclusion, the analysis of the microstructural characteristics using an optical microscope and SEM have proven to be useful as a categorizing reference factor in a provenance study on Goryeo celadons.
NASA Astrophysics Data System (ADS)
Fisher, W. P., Jr.; Petry, P.
2016-11-01
Many published research studies document item calibration invariance across samples using Rasch's probabilistic models for measurement. A new approach to outcomes evaluation for very small samples was employed for two workshop series focused on stress reduction and joyful living conducted for health system employees and caregivers since 2012. Rasch-calibrated self-report instruments measuring depression, anxiety and stress, and the joyful living effects of mindfulness behaviors were identified in peer-reviewed journal articles. Items from one instrument were modified for use with a US population, other items were simplified, and some new items were written. Participants provided ratings of their depression, anxiety and stress, and the effects of their mindfulness behaviors before and after each workshop series. The numbers of participants providing both pre- and post-workshop data were low (16 and 14). Analysis of these small data sets produce results showing that, with some exceptions, the item hierarchies defining the constructs retained the same invariant profiles they had exhibited in the published research (correlations (not disattenuated) range from 0.85 to 0.96). In addition, comparisons of the pre- and post-workshop measures for the three constructs showed substantively and statistically significant changes. Implications for program evaluation comparisons, quality improvement efforts, and the organization of communications concerning outcomes in clinical fields are explored.
Meteorite heat capacities: Results to date
NASA Astrophysics Data System (ADS)
Consolmagno, G.; Macke, R.; Britt, D.
2014-07-01
Heat capacity is an essential thermal property for modeling asteroid internal metamorphism or differentiation, and dynamical effects like YORP or Yarkovsky perturbations. We have developed a rapid, inexpensive, and non-destructive method for measuring the heat capacity of meteorites at low temperature [1]. A sample is introduced into a dewar of liquid nitrogen and an electronic scale measures the amount of nitrogen boiled away as the sample is cooled from the room temperature to the liquid nitrogen temperature; given the heat of vaporization of liquid nitrogen, one can then calculate the heat lost from the sample during the cooling process. Note that heat capacity in this temperature range is a strong function of temperature, but this functional relation is essentially the same for all materials; the values we determine are equivalent to the heat capacity of the sample at 175 K. To correct for systematic errors, samples of laboratory-grade quartz are measured along with the meteorite samples. To date, more than 70 samples of more than 50 different meteorites have been measured in this way, including ordinary chondrites [1], irons [2], basaltic achondrites [3], and a limited number of carbonaceous chondrites [1]. In general, one can draw a number of important conclusions from these results. First, the heat capacity of a meteorite is a function of its mineral composition, independent of shock, metamorphism, or other physical state. Second, given this relation, heat capacity can be strongly altered by terrestrial weathering. Third, the measurement of heat capacity in small (less than 1 g) samples as done typically by commercial systems runs a serious risk of giving misleading results for samples that are heterogeneous on scales of tens of grams or more. Finally, we demonstrate that heat capacity is a useful tool for determining and classifying a sample, especially if used in conjunction with other intrinsic variables such as grain density and magnetic susceptibility. We will present an updated list of our results, incorporating our latest corrections for a variety of small but measurable systematic errors, and new results for meteorites and meteorite types not previously measured or reported.
On the Structure of Cortical Microcircuits Inferred from Small Sample Sizes.
Vegué, Marina; Perin, Rodrigo; Roxin, Alex
2017-08-30
The structure in cortical microcircuits deviates from what would be expected in a purely random network, which has been seen as evidence of clustering. To address this issue, we sought to reproduce the nonrandom features of cortical circuits by considering several distinct classes of network topology, including clustered networks, networks with distance-dependent connectivity, and those with broad degree distributions. To our surprise, we found that all of these qualitatively distinct topologies could account equally well for all reported nonrandom features despite being easily distinguishable from one another at the network level. This apparent paradox was a consequence of estimating network properties given only small sample sizes. In other words, networks that differ markedly in their global structure can look quite similar locally. This makes inferring network structure from small sample sizes, a necessity given the technical difficulty inherent in simultaneous intracellular recordings, problematic. We found that a network statistic called the sample degree correlation (SDC) overcomes this difficulty. The SDC depends only on parameters that can be estimated reliably given small sample sizes and is an accurate fingerprint of every topological family. We applied the SDC criterion to data from rat visual and somatosensory cortex and discovered that the connectivity was not consistent with any of these main topological classes. However, we were able to fit the experimental data with a more general network class, of which all previous topologies were special cases. The resulting network topology could be interpreted as a combination of physical spatial dependence and nonspatial, hierarchical clustering. SIGNIFICANCE STATEMENT The connectivity of cortical microcircuits exhibits features that are inconsistent with a simple random network. Here, we show that several classes of network models can account for this nonrandom structure despite qualitative differences in their global properties. This apparent paradox is a consequence of the small numbers of simultaneously recorded neurons in experiment: when inferred via small sample sizes, many networks may be indistinguishable despite being globally distinct. We develop a connectivity measure that successfully classifies networks even when estimated locally with a few neurons at a time. We show that data from rat cortex is consistent with a network in which the likelihood of a connection between neurons depends on spatial distance and on nonspatial, asymmetric clustering. Copyright © 2017 the authors 0270-6474/17/378498-13$15.00/0.
Catch of channel catfish with tandem-set hoop nets and gill nets in lentic systems of Nebraska
Richters, Lindsey K.; Pope, Kevin L.
2011-01-01
Twenty-six Nebraska water bodies representing two ecosystem types (small standing waters and large standing waters) were surveyed during 2008 and 2009 with tandem-set hoop nets and experimental gill nets to determine if similar trends existed in catch rates and size structures of channel catfish Ictalurus punctatus captured with these gears. Gear efficiency was assessed as the number of sets (nets) that would be required to capture 100 channel catfish given observed catch per unit effort (CPUE). Efficiency of gill nets was not correlated with efficiency of hoop nets for capturing channel catfish. Small sample sizes prohibited estimation of proportional size distributions in most surveys; in the four surveys for which sample size was sufficient to quantify length-frequency distributions of captured channel catfish, distributions differed between gears. The CPUE of channel catfish did not differ between small and large water bodies for either gear. While catch rates of hoop nets were lower than rates recorded in previous studies, this gear was more efficient than gill nets at capturing channel catfish. However, comparisons of size structure between gears may be problematic.
Sekar, R; Deines, P; Machell, J; Osborn, A M; Biggs, C A; Boxall, J B
2012-06-01
To determine the spatial and temporal variability in the abundance, structure and composition of planktonic bacterial assemblages sampled from a small, looped water distribution system and to interpret results with respect to hydraulic conditions. Water samples were collected from five sampling points, twice a day at 06:00 h and 09:00 h on a Monday (following low weekend demand) and a Wednesday (higher midweek demand). All samples were fully compliant with current regulated parameter standards. This study did not show obvious changes in bacterial abundance (DAPI count) or community structure Denaturing gradient gel electrophoresis analysis with respect to sample site and hence to water age; however, the study did show temporal variability with respect to both sampling day and sample times. Data suggests that variations in the bacterial assemblages may be associated with the local system hydraulics: the bacterial composition and numbers, over short durations, are governed by the interaction of the bulk water and the biofilm influenced by the hydraulic conditions. This study demonstrates general stability in bacterial abundance, community structure and composition within the system studied. Trends and patterns supporting the transfer of idealized understanding to the real world were evident. Ultimately, such work will help to safeguard potable water quality, fundamental to public health. © 2012 The Authors. Journal of Applied Microbiology © 2012 The Society for Applied Microbiology.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Stevens, S.E. Jr.; Chung, K.T.
Anaerobic bacteria were isolated from deep subsurface sediment samples taken at study sites in Idaho (INEL) and Washington (HR) by culturing on dilute and concentrated medium. Morphologically distinct colonies were purified, and their responses to 21 selected physiological tests were determined. Although the number of isolates was small (18 INEL, 27 HR) some general patterns could be determined. Most strains could utilize all the carbon sources, however the glycerol and melizitose utilization was positive for 50% or less of the HR isolates. Catalase activity (27.78% at INEL, 74.07% at HR) and tryptophan metabolism (11.12% at INEL, 40.74% at HR) weremore » significantly different between the two study sites. MPN and viable counts indicate that sediments near the water table yield the greatest numbers of anaerobes. Deeper sediments also appear to be more selective with the greatest number of viable counts on low-nutrient mediums. Likewise, only strictly obligate anaerobes were found in the deepest sediment samples. Selective media indicated the presence of methanogens, acetogens, and sulfate reducers at only the HR site.« less
Dynamic Stability Testing of the Genesis Sample Return Capsule
NASA Technical Reports Server (NTRS)
Cheatwood, F. McNeil; Winchenbach, Gerald L.; Hathaway, Wayne; Chapman, Gary
2000-01-01
This paper documents a series of free flight tests of a scale model of the Genesis Sample Return Capsule. These tests were conducted in the Aeroballistic Research Facility (ARF), located at Eglin AFB, FL, during April 1999 and were sponsored by NASA Langley Research Center. Because these blunt atmospheric entry shapes tend to experience small angle of attack dynamic instabilities (frequently leading to limit cycle motions), the primary purpose of the present tests was to determine the dynamic stability characteristics of the Genesis configuration. The tests were conducted over a Mach number range of 1.0 to 4.5. The results for this configuration indicate that the models were dynamically unstable at low angles of attack for all Mach numbers tested. At Mach numbers below 2.5, the models were also unstable at the higher angles of attack (above 15 deg), and motion amplitudes of up to 40 deg were experienced. Above Mach 2.5, the models were dynamically stable at the higher angles of attack.
NASA Astrophysics Data System (ADS)
Yahyaei, Mohsen; Bashiri, Mahdi
2017-12-01
The hub location problem arises in a variety of domains such as transportation and telecommunication systems. In many real-world situations, hub facilities are subject to disruption. This paper deals with the multiple allocation hub location problem in the presence of facilities failure. To model the problem, a two-stage stochastic formulation is developed. In the proposed model, the number of scenarios grows exponentially with the number of facilities. To alleviate this issue, two approaches are applied simultaneously. The first approach is to apply sample average approximation to approximate the two stochastic problem via sampling. Then, by applying the multiple cuts Benders decomposition approach, computational performance is enhanced. Numerical studies show the effective performance of the SAA in terms of optimality gap for small problem instances with numerous scenarios. Moreover, performance of multi-cut Benders decomposition is assessed through comparison with the classic version and the computational results reveal the superiority of the multi-cut approach regarding the computational time and number of iterations.
Automatic liquid handling for life science: a critical review of the current state of the art.
Kong, Fanwei; Yuan, Liang; Zheng, Yuan F; Chen, Weidong
2012-06-01
Liquid handling plays a pivotal role in life science laboratories. In experiments such as gene sequencing, protein crystallization, antibody testing, and drug screening, liquid biosamples frequently must be transferred between containers of varying sizes and/or dispensed onto substrates of varying types. The sample volumes are usually small, at the micro- or nanoliter level, and the number of transferred samples can be huge when investigating large-scope combinatorial conditions. Under these conditions, liquid handling by hand is tedious, time-consuming, and impractical. Consequently, there is a strong demand for automated liquid-handling methods such as sensor-integrated robotic systems. In this article, we survey the current state of the art in automatic liquid handling, including technologies developed by both industry and research institutions. We focus on methods for dealing with small volumes at high throughput and point out challenges for future advancements.
A simple method for the construction of small format tissue arrays
Hidalgo, A; Piña, P; Guerrero, G; Lazos, M; Salcedo, M
2003-01-01
Tissue arrays can evaluate molecular targets in high numbers of samples in parallel. Array construction presents technical difficulties and tissue arrayers are expensive, particularly for small and medium sized laboratories. This report describes a method for the construction of 36 sample arrays using widely available materials. A blunted 16 gauge needle for bone marrow aspiration was used to extract paraffin wax cylinders and manually define a 6 × 6 matrix on a blank paraffin wax block. Tissue cores from 36 paraffin wax embedded premalignant lesions and invasive cervical carcinomas were injected into the matrix using a 14 gauge needle. This tissue array was sectioned using a standard microtome and used for the immunodetection of CD44 variant 9 and interleukin 18 with satisfactory results. This method can be applied in any laboratory, without the need of specialised equipment, offering a good alternative for the wider application of tissue arrays. PMID:12560397
Molecular kinetics. Ras activation by SOS: allosteric regulation by altered fluctuation dynamics.
Iversen, Lars; Tu, Hsiung-Lin; Lin, Wan-Chen; Christensen, Sune M; Abel, Steven M; Iwig, Jeff; Wu, Hung-Jen; Gureasko, Jodi; Rhodes, Christopher; Petit, Rebecca S; Hansen, Scott D; Thill, Peter; Yu, Cheng-Han; Stamou, Dimitrios; Chakraborty, Arup K; Kuriyan, John; Groves, Jay T
2014-07-04
Activation of the small guanosine triphosphatase H-Ras by the exchange factor Son of Sevenless (SOS) is an important hub for signal transduction. Multiple layers of regulation, through protein and membrane interactions, govern activity of SOS. We characterized the specific activity of individual SOS molecules catalyzing nucleotide exchange in H-Ras. Single-molecule kinetic traces revealed that SOS samples a broad distribution of turnover rates through stochastic fluctuations between distinct, long-lived (more than 100 seconds), functional states. The expected allosteric activation of SOS by Ras-guanosine triphosphate (GTP) was conspicuously absent in the mean rate. However, fluctuations into highly active states were modulated by Ras-GTP. This reveals a mechanism in which functional output may be determined by the dynamical spectrum of rates sampled by a small number of enzymes, rather than the ensemble average. Copyright © 2014, American Association for the Advancement of Science.
Portable x-ray fluorescence spectrometer for environmental monitoring of inorganic pollutants
NASA Technical Reports Server (NTRS)
Clark, III, Benton C. (Inventor); Thornton, Michael G. (Inventor)
1991-01-01
A portable x-ray fluorescence spectrometer has a portable sensor unit containing a battery, a high voltage power supply, an x-ray tube which produces a beam x-ray radiation directed toward a target sample, and a detector for fluorescent x-rays produced by the sample. If a silicon-lithium detector is used, the sensor unit also contains either a thermoelectric or thermochemical cooler, or a small dewar flask containing liquid nitrogen to cool the detector. A pulse height analyzer (PHA) generates a spectrum of data for each sample consisting of the number of fluorescent x-rays detected as a function of their energy level. The PHA can also store spectrum data for a number of samples in the field. A processing unit can be attached to the pulse height analyzer to upload and analyze the stored spectrum data for each sample. The processing unit provides a graphic display of the spectrum data for each sample, and provides qualitative and/or quantitative analysis of the elemental composition of the sample by comparing the peaks in the sample spectrum against known x-ray energies for various chemical elements. An optional filtration enclosure can be used to filter particles from a sample suspension, either in the form of a natural suspension or a chemically created precipitate. The sensor unit is then temporarily attached to the filtration unit to analyze the particles collected by the filter medium.
Braschel, Melissa C; Svec, Ivana; Darlington, Gerarda A; Donner, Allan
2016-04-01
Many investigators rely on previously published point estimates of the intraclass correlation coefficient rather than on their associated confidence intervals to determine the required size of a newly planned cluster randomized trial. Although confidence interval methods for the intraclass correlation coefficient that can be applied to community-based trials have been developed for a continuous outcome variable, fewer methods exist for a binary outcome variable. The aim of this study is to evaluate confidence interval methods for the intraclass correlation coefficient applied to binary outcomes in community intervention trials enrolling a small number of large clusters. Existing methods for confidence interval construction are examined and compared to a new ad hoc approach based on dividing clusters into a large number of smaller sub-clusters and subsequently applying existing methods to the resulting data. Monte Carlo simulation is used to assess the width and coverage of confidence intervals for the intraclass correlation coefficient based on Smith's large sample approximation of the standard error of the one-way analysis of variance estimator, an inverted modified Wald test for the Fleiss-Cuzick estimator, and intervals constructed using a bootstrap-t applied to a variance-stabilizing transformation of the intraclass correlation coefficient estimate. In addition, a new approach is applied in which clusters are randomly divided into a large number of smaller sub-clusters with the same methods applied to these data (with the exception of the bootstrap-t interval, which assumes large cluster sizes). These methods are also applied to a cluster randomized trial on adolescent tobacco use for illustration. When applied to a binary outcome variable in a small number of large clusters, existing confidence interval methods for the intraclass correlation coefficient provide poor coverage. However, confidence intervals constructed using the new approach combined with Smith's method provide nominal or close to nominal coverage when the intraclass correlation coefficient is small (<0.05), as is the case in most community intervention trials. This study concludes that when a binary outcome variable is measured in a small number of large clusters, confidence intervals for the intraclass correlation coefficient may be constructed by dividing existing clusters into sub-clusters (e.g. groups of 5) and using Smith's method. The resulting confidence intervals provide nominal or close to nominal coverage across a wide range of parameters when the intraclass correlation coefficient is small (<0.05). Application of this method should provide investigators with a better understanding of the uncertainty associated with a point estimator of the intraclass correlation coefficient used for determining the sample size needed for a newly designed community-based trial. © The Author(s) 2015.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cobb, G.P.; Wood, P.D.; O`Quinn, M.
1995-12-31
Investigation of contaminant burdens in threatened and endangered species is difficult due to the small number of samples that can be collected. Many samples can be collected if the sampling methods are non-lethal and more specifically non-invasive. Analysis of chorioallantoic membranes is demonstrated for American alligators and Loggerhead sea turtles. Significant differences were found in PCB, uptake by alligators from the Ashepoo-Combahee-Edisto Basin reference site and a contaminated site in Winyah Bay. Intrasite and intersite differences in uptake and distribution of PCB homologues were noted. These data will be discussed as they relate to egg viability and embryo development.
Junttila, Virpi; Kauranne, Tuomo; Finley, Andrew O.; Bradford, John B.
2015-01-01
Modern operational forest inventory often uses remotely sensed data that cover the whole inventory area to produce spatially explicit estimates of forest properties through statistical models. The data obtained by airborne light detection and ranging (LiDAR) correlate well with many forest inventory variables, such as the tree height, the timber volume, and the biomass. To construct an accurate model over thousands of hectares, LiDAR data must be supplemented with several hundred field sample measurements of forest inventory variables. This can be costly and time consuming. Different LiDAR-data-based and spatial-data-based sampling designs can reduce the number of field sample plots needed. However, problems arising from the features of the LiDAR data, such as a large number of predictors compared with the sample size (overfitting) or a strong correlation among predictors (multicollinearity), may decrease the accuracy and precision of the estimates and predictions. To overcome these problems, a Bayesian linear model with the singular value decomposition of predictors, combined with regularization, is proposed. The model performance in predicting different forest inventory variables is verified in ten inventory areas from two continents, where the number of field sample plots is reduced using different sampling designs. The results show that, with an appropriate field plot selection strategy and the proposed linear model, the total relative error of the predicted forest inventory variables is only 5%–15% larger using 50 field sample plots than the error of a linear model estimated with several hundred field sample plots when we sum up the error due to both the model noise variance and the model’s lack of fit.
Contaminants in landfill soils - Reliability of prefeasibility studies.
Hölzle, Ingo
2017-05-01
Recent landfill mining studies have researched the potential for resource recovery using samples from core drilling or grab cranes. However, most studies used small sample numbers, which may not represent the heterogeneous landfill composition. As a consequence, there exists a high risk of an incorrect economic and/or ecological evaluation. The main objective of this work is to investigate the possibilities and limitations of preliminary investigations concerning the crucial soil composition. The preliminary samples of landfill investigations were compared to the excavation samples from three completely excavated landfills in Germany. In addition, the research compared the reliability of prediction of the two investigation methods, core drilling and grab crane. Sampling using a grab crane led to better results, even for smaller investigations of 10 samples. Analyses of both methods showed sufficiently accurate results to make predictions (standard error 5%, level of confidence 95%) for most heavy metals, cyanide and PAH in the dry substance and for sulphate, barium, Benzo[a]pyrene, pH and the electrical conductivity in leachate analyses of soil type waste. While chrome and nickel showed less accurate results, the concentrations of hydrocarbons, TOC, DOC, PCB and fluorine (leachate) were not predictable even for sample numbers of up to 59. Overestimations of pollutant concentrations were more frequently apparent in drilling, and underestimations when using a grab crane. The dispersion of the element and elemental composition had no direct impact on the reliability of prediction. Thus, an individual consideration of the particular element or elemental composition for dry substance and leachate analyses is recommended to adapt the sample strategy and calculate an optimum sample number. Copyright © 2016 Elsevier Ltd. All rights reserved.
Delayed reward discounting and addictive behavior: a meta-analysis.
MacKillop, James; Amlung, Michael T; Few, Lauren R; Ray, Lara A; Sweet, Lawrence H; Munafò, Marcus R
2011-08-01
Delayed reward discounting (DRD) is a behavioral economic index of impulsivity and numerous studies have examined DRD in relation to addictive behavior. To synthesize the findings across the literature, the current review is a meta-analysis of studies comparing DRD between criterion groups exhibiting addictive behavior and control groups. The meta-analysis sought to characterize the overall patterns of findings, systematic variability by sample and study type, and possible small study (publication) bias. Literature reviews identified 310 candidate articles from which 46 studies reporting 64 comparisons were identified (total N=56,013). From the total comparisons identified, a small magnitude effect was evident (d= .15; p< .00001) with very high heterogeneity of effect size. Based on systematic observed differences, large studies assessing DRD with a small number of self-report items were removed and an analysis of 57 comparisons (n=3,329) using equivalent methods and exhibiting acceptable heterogeneity revealed a medium magnitude effect (d= .58; p< .00001). Further analyses revealed significantly larger effect sizes for studies using clinical samples (d= .61) compared with studies using nonclinical samples (d=.45). Indices of small study bias among the various comparisons suggested varying levels of influence by unpublished findings, ranging from minimal to moderate. These results provide strong evidence of greater DRD in individuals exhibiting addictive behavior in general and particularly in individuals who meet criteria for an addictive disorder. Implications for the assessment of DRD and research priorities are discussed.
Delayed reward discounting and addictive behavior: a meta-analysis
Amlung, Michael T.; Few, Lauren R.; Ray, Lara A.; Sweet, Lawrence H.; Munafò, Marcus R.
2011-01-01
Rationale Delayed reward discounting (DRD) is a behavioral economic index of impulsivity and numerous studies have examined DRD in relation to addictive behavior. To synthesize the findings across the literature, the current review is a meta-analysis of studies comparing DRD between criterion groups exhibiting addictive behavior and control groups. Objectives The meta-analysis sought to characterize the overall patterns of findings, systematic variability by sample and study type, and possible small study (publication) bias. Methods Literature reviews identified 310 candidate articles from which 46 studies reporting 64 comparisons were identified (total N=56,013). Results From the total comparisons identified, a small magnitude effect was evident (d=.15; p<.00001) with very high heterogeneity of effect size. Based on systematic observed differences, large studies assessing DRD with a small number of self-report items were removed and an analysis of 57 comparisons (n=3,329) using equivalent methods and exhibiting acceptable heterogeneity revealed a medium magnitude effect (d=.58; p<.00001). Further analyses revealed significantly larger effect sizes for studies using clinical samples (d=.61) compared with studies using nonclinical samples (d=.45). Indices of small study bias among the various comparisons suggested varying levels of influence by unpublished findings, ranging from minimal to moderate. Conclusions These results provide strong evidence of greater DRD in individuals exhibiting addictive behavior in general and particularly in individuals who meet criteria for an addictive disorder. Implications for the assessment of DRD and research priorities are discussed. PMID:21373791
ERIC Educational Resources Information Center
Byrd, Sharon R.
The study examined barriers to education perceived by a sample of 119 students aged 25 and over attending small private liberal arts colleges in Arkansas, Mississippi, and Tennessee. Among findings were that the number of children a respondent had impacted his or her perception of situational and dispositional barriers; the employment status of…
[Syagrus romanzoffiana (Arecaceae) seed utilization by ants in a secondary forest in South Brazil].
Silva, Fernanda R; Begnini, Romualdo M; Klier, Vinícius A; Scherer, Karla Z; Lopes, Benedito C; Castellani, Tânia T
2009-01-01
Ants can nest in a wide variety of substracts. This paper shows Syagrus romanzoffiana seed utilization by ants in an Atlantic secondary forest. We report 29 seeds occupied by small-bodied ants, with 27 of them showing at least two ant development stages. Although a large number of seeds were sampled, a low level of ant occupation was observed.
ERIC Educational Resources Information Center
Zacharakis, Jeff; Wang, Haiyan; Patterson, Margaret Becker; Andersen, Lori
2015-01-01
This research analyzed linked high-quality state data from K-12, adult education, and postsecondary state datasets in order to better understand the association between student demographics and successful completion of a postsecondary program. Due to the relatively small sample size compared to the large number of features, we analyzed the data…
[Scientific collaboration and article citations: practices in medical journals].
Bador, Pascal; Lafouge, Thierry
2012-01-01
In order to characterize scientific collaboration the best way is to study co-signature of articles. Two indicators are interesting: number of authors and international character. The objective is to study correlation between these two indicators and citation number. We selected two pharmacy and medicine journals in order to compare practices. We used a sample of about 800 articles published in 2002-2005 for which we collected citations up to 2010. We transformed numeric variables, authors number and citation number, into qualitative variables. "Authors" and "Citations" variables are not independent. Less cited articles are often published by one author or a very small team while international character of articles generally increases citation number. This micro-analysis also allowed us to better understand publication practices. © 2012 Société Française de Pharmacologie et de Thérapeutique.
Kim, Ki Hwan; Do, Won-Joon; Park, Sung-Hong
2018-05-04
The routine MRI scan protocol consists of multiple pulse sequences that acquire images of varying contrast. Since high frequency contents such as edges are not significantly affected by image contrast, down-sampled images in one contrast may be improved by high resolution (HR) images acquired in another contrast, reducing the total scan time. In this study, we propose a new deep learning framework that uses HR MR images in one contrast to generate HR MR images from highly down-sampled MR images in another contrast. The proposed convolutional neural network (CNN) framework consists of two CNNs: (a) a reconstruction CNN for generating HR images from the down-sampled images using HR images acquired with a different MRI sequence and (b) a discriminator CNN for improving the perceptual quality of the generated HR images. The proposed method was evaluated using a public brain tumor database and in vivo datasets. The performance of the proposed method was assessed in tumor and no-tumor cases separately, with perceptual image quality being judged by a radiologist. To overcome the challenge of training the network with a small number of available in vivo datasets, the network was pretrained using the public database and then fine-tuned using the small number of in vivo datasets. The performance of the proposed method was also compared to that of several compressed sensing (CS) algorithms. Incorporating HR images of another contrast improved the quantitative assessments of the generated HR image in reference to ground truth. Also, incorporating a discriminator CNN yielded perceptually higher image quality. These results were verified in regions of normal tissue as well as tumors for various MRI sequences from pseudo k-space data generated from the public database. The combination of pretraining with the public database and fine-tuning with the small number of real k-space datasets enhanced the performance of CNNs in in vivo application compared to training CNNs from scratch. The proposed method outperformed the compressed sensing methods. The proposed method can be a good strategy for accelerating routine MRI scanning. © 2018 American Association of Physicists in Medicine.
Small mammal abundance in Mediterranean post-fire habitats: a role for predators?
NASA Astrophysics Data System (ADS)
Torre, I.; Díaz, M.
2004-05-01
We studied patterns of small mammal abundance and species richness in post-fire habitats by sampling 33 plots (225 m 2 each) representing different stages of vegetation recovery after fire. Small mammal abundance was estimated by live trapping during early spring 1999 and vegetation structure was sampled by visual estimation at the same plots. Recently-burnt areas were characterised by shrubby and herbaceous vegetation with low structural variability, and unburnt areas were characterised by well developed forest cover with high structural complexity. Small mammal abundance and species richness decreased with time elapsed since the last fire (from 5 to at least 50 years), and these differences were associated to the decreasing cover of short shrubs as the post-fire succession of plant communities advanced. However, relationships between vegetation structure and small mammals differed among areas burned in different times, with weak or negative relationship in recently burnt areas and positive and stronger relationship in unburnt areas. Furthermore, the abundance of small mammals was larger than expected from vegetation structure in plots burned recently whereas the contrary pattern was found in unburned areas. We hypothesised that the pattern observed could be related to the responses of small mammal predators to changes in vegetation and landscape structure promoted by fire. Fire-related fragmentation could have promoted the isolation of forest predators (owls and carnivores) in unburned forest patches, a fact that could have produced a higher predation pressure for small mammals. Conversely, small mammal populations would have been enhanced in early post-fire stages by lower predator numbers combined with better predator protection in areas covered by resprouting woody vegetation.
Zhang, Fang; Wagner, Anita K; Ross-Degnan, Dennis
2011-11-01
Interrupted time series is a strong quasi-experimental research design to evaluate the impacts of health policy interventions. Using simulation methods, we estimated the power requirements for interrupted time series studies under various scenarios. Simulations were conducted to estimate the power of segmented autoregressive (AR) error models when autocorrelation ranged from -0.9 to 0.9 and effect size was 0.5, 1.0, and 2.0, investigating balanced and unbalanced numbers of time periods before and after an intervention. Simple scenarios of autoregressive conditional heteroskedasticity (ARCH) models were also explored. For AR models, power increased when sample size or effect size increased, and tended to decrease when autocorrelation increased. Compared with a balanced number of study periods before and after an intervention, designs with unbalanced numbers of periods had less power, although that was not the case for ARCH models. The power to detect effect size 1.0 appeared to be reasonable for many practical applications with a moderate or large number of time points in the study equally divided around the intervention. Investigators should be cautious when the expected effect size is small or the number of time points is small. We recommend conducting various simulations before investigation. Copyright © 2011 Elsevier Inc. All rights reserved.
The evolution of ovule number and flower size in wind-pollinated plants.
Friedman, Jannice; Barrett, Spencer C H
2011-02-01
In angiosperms, ovules are "packaged" within individual flowers, and an optimal strategy should occur depending on pollination and resource conditions. In animal-pollinated species, wide variation in ovule number per flower occurs, and this contrasts with wind-pollinated plants, where most species possess uniovulate flowers. This pattern is usually explained as an adaptive response to low pollen receipt in wind-pollinated species. Here, we develop a phenotypic model for the evolution of ovule number per flower that incorporates the aerodynamics of pollen capture and a fixed resource pool for provisioning of flowers, ovules, and seeds. Our results challenge the prevailing explanation for the association between uniovulate flowers and wind pollination. We demonstrate that when flowers are small and inexpensive, as they are in wind-pollinated species, ovule number should be minimized and lower than the average number of pollen tubes per style, even under stochastic pollination and fertilization regimes. The model predicts that plants benefit from producing many small inexpensive flowers, even though some flowers capture too few pollen grains to fertilize their ovules. Wind-pollinated plants with numerous flowers distributed throughout the inflorescence, each with a single ovule or a few ovules, sample more of the airstream, and this should maximize pollen capture and seed production.
Current developments in forensic interpretation of mixed DNA samples (Review).
Hu, Na; Cong, Bin; Li, Shujin; Ma, Chunling; Fu, Lihong; Zhang, Xiaojing
2014-05-01
A number of recent improvements have provided contemporary forensic investigations with a variety of tools to improve the analysis of mixed DNA samples in criminal investigations, producing notable improvements in the analysis of complex trace samples in cases of sexual assult and homicide. Mixed DNA contains DNA from two or more contributors, compounding DNA analysis by combining DNA from one or more major contributors with small amounts of DNA from potentially numerous minor contributors. These samples are characterized by a high probability of drop-out or drop-in combined with elevated stutter, significantly increasing analysis complexity. At some loci, minor contributor alleles may be completely obscured due to amplification bias or over-amplification, creating the illusion of additional contributors. Thus, estimating the number of contributors and separating contributor genotypes at a given locus is significantly more difficult in mixed DNA samples, requiring the application of specialized protocols that have only recently been widely commercialized and standardized. Over the last decade, the accuracy and repeatability of mixed DNA analyses available to conventional forensic laboratories has greatly advanced in terms of laboratory technology, mathematical models and biostatistical software, generating more accurate, rapid and readily available data for legal proceedings and criminal cases.
Current developments in forensic interpretation of mixed DNA samples (Review)
HU, NA; CONG, BIN; LI, SHUJIN; MA, CHUNLING; FU, LIHONG; ZHANG, XIAOJING
2014-01-01
A number of recent improvements have provided contemporary forensic investigations with a variety of tools to improve the analysis of mixed DNA samples in criminal investigations, producing notable improvements in the analysis of complex trace samples in cases of sexual assult and homicide. Mixed DNA contains DNA from two or more contributors, compounding DNA analysis by combining DNA from one or more major contributors with small amounts of DNA from potentially numerous minor contributors. These samples are characterized by a high probability of drop-out or drop-in combined with elevated stutter, significantly increasing analysis complexity. At some loci, minor contributor alleles may be completely obscured due to amplification bias or over-amplification, creating the illusion of additional contributors. Thus, estimating the number of contributors and separating contributor genotypes at a given locus is significantly more difficult in mixed DNA samples, requiring the application of specialized protocols that have only recently been widely commercialized and standardized. Over the last decade, the accuracy and repeatability of mixed DNA analyses available to conventional forensic laboratories has greatly advanced in terms of laboratory technology, mathematical models and biostatistical software, generating more accurate, rapid and readily available data for legal proceedings and criminal cases. PMID:24748965
NASA Astrophysics Data System (ADS)
Beck, Joakim; Dia, Ben Mansour; Espath, Luis F. R.; Long, Quan; Tempone, Raúl
2018-06-01
In calculating expected information gain in optimal Bayesian experimental design, the computation of the inner loop in the classical double-loop Monte Carlo requires a large number of samples and suffers from underflow if the number of samples is small. These drawbacks can be avoided by using an importance sampling approach. We present a computationally efficient method for optimal Bayesian experimental design that introduces importance sampling based on the Laplace method to the inner loop. We derive the optimal values for the method parameters in which the average computational cost is minimized according to the desired error tolerance. We use three numerical examples to demonstrate the computational efficiency of our method compared with the classical double-loop Monte Carlo, and a more recent single-loop Monte Carlo method that uses the Laplace method as an approximation of the return value of the inner loop. The first example is a scalar problem that is linear in the uncertain parameter. The second example is a nonlinear scalar problem. The third example deals with the optimal sensor placement for an electrical impedance tomography experiment to recover the fiber orientation in laminate composites.
Effect of finite sample size on feature selection and classification: a simulation study.
Way, Ted W; Sahiner, Berkman; Hadjiiski, Lubomir M; Chan, Heang-Ping
2010-02-01
The small number of samples available for training and testing is often the limiting factor in finding the most effective features and designing an optimal computer-aided diagnosis (CAD) system. Training on a limited set of samples introduces bias and variance in the performance of a CAD system relative to that trained with an infinite sample size. In this work, the authors conducted a simulation study to evaluate the performances of various combinations of classifiers and feature selection techniques and their dependence on the class distribution, dimensionality, and the training sample size. The understanding of these relationships will facilitate development of effective CAD systems under the constraint of limited available samples. Three feature selection techniques, the stepwise feature selection (SFS), sequential floating forward search (SFFS), and principal component analysis (PCA), and two commonly used classifiers, Fisher's linear discriminant analysis (LDA) and support vector machine (SVM), were investigated. Samples were drawn from multidimensional feature spaces of multivariate Gaussian distributions with equal or unequal covariance matrices and unequal means, and with equal covariance matrices and unequal means estimated from a clinical data set. Classifier performance was quantified by the area under the receiver operating characteristic curve Az. The mean Az values obtained by resubstitution and hold-out methods were evaluated for training sample sizes ranging from 15 to 100 per class. The number of simulated features available for selection was chosen to be 50, 100, and 200. It was found that the relative performance of the different combinations of classifier and feature selection method depends on the feature space distributions, the dimensionality, and the available training sample sizes. The LDA and SVM with radial kernel performed similarly for most of the conditions evaluated in this study, although the SVM classifier showed a slightly higher hold-out performance than LDA for some conditions and vice versa for other conditions. PCA was comparable to or better than SFS and SFFS for LDA at small samples sizes, but inferior for SVM with polynomial kernel. For the class distributions simulated from clinical data, PCA did not show advantages over the other two feature selection methods. Under this condition, the SVM with radial kernel performed better than the LDA when few training samples were available, while LDA performed better when a large number of training samples were available. None of the investigated feature selection-classifier combinations provided consistently superior performance under the studied conditions for different sample sizes and feature space distributions. In general, the SFFS method was comparable to the SFS method while PCA may have an advantage for Gaussian feature spaces with unequal covariance matrices. The performance of the SVM with radial kernel was better than, or comparable to, that of the SVM with polynomial kernel under most conditions studied.
Adaptively biased molecular dynamics: An umbrella sampling method with a time-dependent potential
NASA Astrophysics Data System (ADS)
Babin, Volodymyr; Karpusenka, Vadzim; Moradi, Mahmoud; Roland, Christopher; Sagui, Celeste
We discuss an adaptively biased molecular dynamics (ABMD) method for the computation of a free energy surface for a set of reaction coordinates. The ABMD method belongs to the general category of umbrella sampling methods with an evolving biasing potential. It is characterized by a small number of control parameters and an O(t) numerical cost with simulation time t. The method naturally allows for extensions based on multiple walkers and replica exchange mechanism. The workings of the method are illustrated with a number of examples, including sugar puckering, and free energy landscapes for polymethionine and polyproline peptides, and for a short β-turn peptide. ABMD has been implemented into the latest version (Case et al., AMBER 10; University of California: San Francisco, 2008) of the AMBER software package and is freely available to the simulation community.
Image analysis for quantification of bacterial rock weathering.
Puente, M Esther; Rodriguez-Jaramillo, M Carmen; Li, Ching Y; Bashan, Yoav
2006-02-01
A fast, quantitative image analysis technique was developed to assess potential rock weathering by bacteria. The technique is based on reduction in the surface area of rock particles and counting the relative increase in the number of small particles in ground rock slurries. This was done by recording changes in ground rock samples with an electronic image analyzing process. The slurries were previously amended with three carbon sources, ground to a uniform particle size and incubated with rock weathering bacteria for 28 days. The technique was developed and tested, using two rock-weathering bacteria Pseudomonas putida R-20 and Azospirillum brasilense Cd on marble, granite, apatite, quartz, limestone, and volcanic rock as substrates. The image analyzer processed large number of particles (10(7)-10(8) per sample), so that the weathering capacity of bacteria can be detected.
NASA Technical Reports Server (NTRS)
Nebenfuhr, A.; Lomax, T. L.
1998-01-01
We have developed an improved method for determination of gene expression levels with RT-PCR. The procedure is rapid and does not require extensive optimization or densitometric analysis. Since the detection of individual transcripts is PCR-based, small amounts of tissue samples are sufficient for the analysis of expression patterns in large gene families. Using this method, we were able to rapidly screen nine members of the Aux/IAA family of auxin-responsive genes and identify those genes which vary in message abundance in a tissue- and light-specific manner. While not offering the accuracy of conventional semi-quantitative or competitive RT-PCR, our method allows quick screening of large numbers of genes in a wide range of RNA samples with just a thermal cycler and standard gel analysis equipment.
The large bright quasar survey. 6: Quasar catalog and survey parameters
NASA Astrophysics Data System (ADS)
Hewett, Paul C.; Foltz, Craig B.; Chaffee, Frederic H.
1995-04-01
Positions, redshifts, and magnitudes for the 1055 quasars in the Large Bright Quasar Survey (LBQS) are presented in a single catalog. Celestial positions have been derived using the PPM catalog to provide an improved reference frame. J2000.0 coordinates are given together with improved b1950.0 positions. Redshifts calculated via cross correlation with a high signal-to-noise ratio composite quasar spectrum are included and the small number of typographic and redshift misidentifications in the discovery papers are corrected. Spectra of the 12 quasars added to the sample since the publication of the discovery papers are included. Discriptions of the plate material, magnitude calibration, quasar candidate selection procedures, and the identification spectroscopy are given. Calculation of the effective area of the survey for the 1055 quasars comprising the well-defined LBQS sample specified in detail. Number-redshift and number-magnitude relations for the quasars are derived and the strengths and limitastions of the LBSQ sample summarized. Comparison with existing surveys is made and a qualitative assessment of the effectiveness of the LBQS undertaken. Positions, magnitudes, and optical spectra of the eight objects (less than 1%) in the survey that remain unidentified are also presented.
Sarkar, F H; Valdivieso, M; Borders, J; Yao, K L; Raval, M M; Madan, S K; Sreepathi, P; Shimoyama, R; Steiger, Z; Visscher, D W
1995-12-01
The p53 tumor suppressor gene has been found to be altered in almost all human solid tumors, whereas K-ras gene mutations have been observed in a limited number of human cancers (adenocarcinoma of colon, pancreas, and lung). Studies of mutational inactivation for both genes in the same patient's sample on non-small-cell lung cancer have been limited. In an effort to perform such an analysis, we developed and compared methods (for the mutational detection of p53 and K-ras gene) that represent a modified and universal protocol, in terms of DNA extraction, polymerase chain reaction (PCR) amplification, and nonradioisotopic PCR-single-strand conformation polymorphism (PCR-SSCP) analysis, which is readily applicable to either formalin-fixed, paraffin-embedded tissues or frozen tumor specimens. We applied this method to the evaluation of p53 (exons 5-8) and K-ras (codon 12 and 13) gene mutations in 55 cases of non-small-cell lung cancer. The mutational status in the p53 gene was evaluated by radioisotopic PCR-SSCP and compared with PCR-SSCP utilizing our standardized nonradioisotopic detection system using a single 6-microns tissue section. The mutational patterns observed by PCR-SSCP were subsequently confirmed by PCR-DNA sequencing. The mutational status in the K-ras gene was similarly evaluated by PCR-SSCP, and the specific mutation was confirmed by Southern slot-blot hybridization using 32P-labeled sequence-specific oligonucleotide probes for codons 12 and 13. Mutational changes in K-ras (codon 12) were found in 10 of 55 (18%) of non-small-cell lung cancers. Whereas adenocarcinoma showed K-ras mutation in 33% of the cases at codon 12, only one mutation was found at codon 13. As expected, squamous cell carcinoma samples (25 cases) did not show K-ras mutations. Mutations at exons 5-8 of the p53 gene were documented in 19 of 55 (34.5%) cases. Ten of the 19 mutations were single nucleotide point mutations, leading to amino acid substitution. Six showed insertional mutation, and three showed deletion mutations. Only three samples showed mutations of both K-ras and p53 genes. We conclude that although K-ras and p53 gene mutations are frequent in non-small-cell lung cancer, mutations of both genes in the same patient's samples are not common. We also conclude that this universal nonradioisotopic method is superior to other similar methods and is readily applicable to the rapid screening of large numbers of formalin-fixed, paraffin-embedded or frozen samples for the mutational analysis of multiple genes.
Practice does make perfect. A longitudinal look at repeated taste exposure.
Williams, Keith E; Paul, Candace; Pizzo, Bianca; Riegel, Katherine
2008-11-01
Previous research has found that 10-15 exposures to a novel food found can increase liking and consumption. This research has been, however, largely limited cross-sectional studies in which participants are offered only one or a few novel foods. The goal of the current study uses a small clinical sample to demonstrate the number of exposures required for consumption of novel foods decreases as a greater number of foods are added to the diet. Evidence that fewer exposures are needed over time may make interventions based upon repeated exposure more acceptable to parents and clinicians.
Effects of sample size on estimates of population growth rates calculated with matrix models.
Fiske, Ian J; Bruna, Emilio M; Bolker, Benjamin M
2008-08-28
Matrix models are widely used to study the dynamics and demography of populations. An important but overlooked issue is how the number of individuals sampled influences estimates of the population growth rate (lambda) calculated with matrix models. Even unbiased estimates of vital rates do not ensure unbiased estimates of lambda-Jensen's Inequality implies that even when the estimates of the vital rates are accurate, small sample sizes lead to biased estimates of lambda due to increased sampling variance. We investigated if sampling variability and the distribution of sampling effort among size classes lead to biases in estimates of lambda. Using data from a long-term field study of plant demography, we simulated the effects of sampling variance by drawing vital rates and calculating lambda for increasingly larger populations drawn from a total population of 3842 plants. We then compared these estimates of lambda with those based on the entire population and calculated the resulting bias. Finally, we conducted a review of the literature to determine the sample sizes typically used when parameterizing matrix models used to study plant demography. We found significant bias at small sample sizes when survival was low (survival = 0.5), and that sampling with a more-realistic inverse J-shaped population structure exacerbated this bias. However our simulations also demonstrate that these biases rapidly become negligible with increasing sample sizes or as survival increases. For many of the sample sizes used in demographic studies, matrix models are probably robust to the biases resulting from sampling variance of vital rates. However, this conclusion may depend on the structure of populations or the distribution of sampling effort in ways that are unexplored. We suggest more intensive sampling of populations when individual survival is low and greater sampling of stages with high elasticities.
A prototype splitter apparatus for dividing large catches of small fish
Stapanian, Martin A.; Edwards, William H.
2012-01-01
Due to financial and time constraints, it is often necessary in fisheries studies to divide large samples of fish and estimate total catch from the subsample. The subsampling procedure may involve potential human biases or may be difficult to perform in rough conditions. We present a prototype gravity-fed splitter apparatus for dividing large samples of small fish (30–100 mm TL). The apparatus features a tapered hopper with a sliding and removable shutter. The apparatus provides a comparatively stable platform for objectively obtaining subsamples, and it can be modified to accommodate different sizes of fish and different sample volumes. The apparatus is easy to build, inexpensive, and convenient to use in the field. To illustrate the performance of the apparatus, we divided three samples (total N = 2,000 fish) composed of four fish species. Our results indicated no significant bias in estimating either the number or proportion of each species from the subsample. Use of this apparatus or a similar apparatus can help to standardize subsampling procedures in large surveys of fish. The apparatus could be used for other applications that require dividing a large amount of material into one or more smaller subsamples.
Accounting for randomness in measurement and sampling in studying cancer cell population dynamics.
Ghavami, Siavash; Wolkenhauer, Olaf; Lahouti, Farshad; Ullah, Mukhtar; Linnebacher, Michael
2014-10-01
Knowing the expected temporal evolution of the proportion of different cell types in sample tissues gives an indication about the progression of the disease and its possible response to drugs. Such systems have been modelled using Markov processes. We here consider an experimentally realistic scenario in which transition probabilities are estimated from noisy cell population size measurements. Using aggregated data of FACS measurements, we develop MMSE and ML estimators and formulate two problems to find the minimum number of required samples and measurements to guarantee the accuracy of predicted population sizes. Our numerical results show that the convergence mechanism of transition probabilities and steady states differ widely from the real values if one uses the standard deterministic approach for noisy measurements. This provides support for our argument that for the analysis of FACS data one should consider the observed state as a random variable. The second problem we address is about the consequences of estimating the probability of a cell being in a particular state from measurements of small population of cells. We show how the uncertainty arising from small sample sizes can be captured by a distribution for the state probability.
Sampling effects on the identification of roadkill hotspots: Implications for survey design.
Santos, Sara M; Marques, J Tiago; Lourenço, André; Medinas, Denis; Barbosa, A Márcia; Beja, Pedro; Mira, António
2015-10-01
Although locating wildlife roadkill hotspots is essential to mitigate road impacts, the influence of study design on hotspot identification remains uncertain. We evaluated how sampling frequency affects the accuracy of hotspot identification, using a dataset of vertebrate roadkills (n = 4427) recorded over a year of daily surveys along 37 km of roads. "True" hotspots were identified using this baseline dataset, as the 500-m segments where the number of road-killed vertebrates exceeded the upper 95% confidence limit of the mean, assuming a Poisson distribution of road-kills per segment. "Estimated" hotspots were identified likewise, using datasets representing progressively lower sampling frequencies, which were produced by extracting data from the baseline dataset at appropriate time intervals (1-30 days). Overall, 24.3% of segments were "true" hotspots, concentrating 40.4% of roadkills. For different groups, "true" hotspots accounted from 6.8% (bats) to 29.7% (small birds) of road segments, concentrating from <40% (frogs and toads, snakes) to >60% (lizards, lagomorphs, carnivores) of roadkills. Spatial congruence between "true" and "estimated" hotspots declined rapidly with increasing time interval between surveys, due primarily to increasing false negatives (i.e., missing "true" hotspots). There were also false positives (i.e., wrong "estimated" hotspots), particularly at low sampling frequencies. Spatial accuracy decay with increasing time interval between surveys was higher for smaller-bodied (amphibians, reptiles, small birds, small mammals) than for larger-bodied species (birds of prey, hedgehogs, lagomorphs, carnivores). Results suggest that widely used surveys at weekly or longer intervals may produce poor estimates of roadkill hotspots, particularly for small-bodied species. Surveying daily or at two-day intervals may be required to achieve high accuracy in hotspot identification for multiple species. Copyright © 2015 Elsevier Ltd. All rights reserved.
Voineskos, Sophocles H; Coroneos, Christopher J; Ziolkowski, Natalia I; Kaur, Manraj N; Banfield, Laura; Meade, Maureen O; Chung, Kevin C; Thoma, Achilleas; Bhandari, Mohit
2016-02-01
The authors examined industry support, conflict of interest, and sample size in plastic surgery randomized controlled trials that compared surgical interventions. They hypothesized that industry-funded trials demonstrate statistically significant outcomes more often, and randomized controlled trials with small sample sizes report statistically significant results more frequently. An electronic search identified randomized controlled trials published between 2000 and 2013. Independent reviewers assessed manuscripts and performed data extraction. Funding source, conflict of interest, primary outcome direction, and sample size were examined. Chi-squared and independent-samples t tests were used in the analysis. The search identified 173 randomized controlled trials, of which 100 (58 percent) did not acknowledge funding status. A relationship between funding source and trial outcome direction was not observed. Both funding status and conflict of interest reporting improved over time. Only 24 percent (six of 25) of industry-funded randomized controlled trials reported authors to have independent control of data and manuscript contents. The mean number of patients randomized was 73 per trial (median, 43, minimum, 3, maximum, 936). Small trials were not found to be positive more often than large trials (p = 0.87). Randomized controlled trials with small sample size were common; however, this provides great opportunity for the field to engage in further collaboration and produce larger, more definitive trials. Reporting of trial funding and conflict of interest is historically poor, but it greatly improved over the study period. Underreporting at author and journal levels remains a limitation when assessing the relationship between funding source and trial outcomes. Improved reporting and manuscript control should be goals that both authors and journals can actively achieve.
DNA extraction for streamlined metagenomics of diverse environmental samples.
Marotz, Clarisse; Amir, Amnon; Humphrey, Greg; Gaffney, James; Gogul, Grant; Knight, Rob
2017-06-01
A major bottleneck for metagenomic sequencing is rapid and efficient DNA extraction. Here, we compare the extraction efficiencies of three magnetic bead-based platforms (KingFisher, epMotion, and Tecan) to a standardized column-based extraction platform across a variety of sample types, including feces, oral, skin, soil, and water. Replicate sample plates were extracted and prepared for 16S rRNA gene amplicon sequencing in parallel to assess extraction bias and DNA quality. The data demonstrate that any effect of extraction method on sequencing results was small compared with the variability across samples; however, the KingFisher platform produced the largest number of high-quality reads in the shortest amount of time. Based on these results, we have identified an extraction pipeline that dramatically reduces sample processing time without sacrificing bacterial taxonomic or abundance information.
Sampling design for the 1980 commercial and multifamily residential building survey
NASA Astrophysics Data System (ADS)
Bowen, W. M.; Olsen, A. R.; Nieves, A. L.
1981-06-01
The extent to which new building design practices comply with the proposed 1980 energy budget levels for commercial and multifamily residential building designs (DEB-80) can be assessed by: (1) identifying small number of building types which account for the majority of commercial buildings constructed in the U.S.A.; (2) conducting a separate survey for each building type; and (3) including only buildings designed during 1980. For each building, the design energy consumption (DEC-80) will be determined by the DOE2.1 computer program. The quantity X = (DEC-80 - DEB-80). These X quantities can then be used to compute sample statistics. Inferences about nationwide compliance with DEB-80 may then be made for each building type. Details of the population, sampling frame, stratification, sample size, and implementation of the sampling plan are provided.
Ko, Ryo; Kenmotsu, Hirotsugu; Serizawa, Masakuni; Koh, Yasuhiro; Wakuda, Kazushige; Ono, Akira; Taira, Tetsuhiko; Naito, Tateaki; Murakami, Haruyasu; Isaka, Mitsuhiro; Endo, Masahiro; Nakajima, Takashi; Ohde, Yasuhisa; Yamamoto, Nobuyuki; Takahashi, Kazuhisa; Takahashi, Toshiaki
2016-11-08
The majority of non-small cell lung cancer (NSCLC) patients with epidermal growth factor receptor (EGFR) mutation eventually develop resistance to EGFR tyrosine kinase inhibitors (TKIs). Minimal information exists regarding genetic alterations in rebiopsy samples from Asian NSCLC patients who develop acquired resistance to EGFR-TKIs. We retrospectively reviewed the medical records of patients with NSCLC harboring EGFR mutations who had undergone rebiopsies after developing acquired resistance to EGFR-TKIs. We analyzed 27 practicable samples using a tumor genotyping panel to assess 23 hot-spot sites of genetic alterations in nine genes (EGFR, KRAS, BRAF, PIK3CA, NRAS, MEK1, AKT1, PTEN, and HER2), gene copy number of EGFR, MET, PIK3CA, FGFR1, and FGFR2, and ALK, ROS1, and RET fusions. Additionally, 34 samples were analyzed by commercially available EGFR mutation tests. Sixty-one patients underwent rebiopsy. Twenty-seven samples were analyzed using our tumor genotyping panel, and 34 samples were analyzed for EGFR mutations only by commercial clinical laboratories. Twenty-one patients (34 %) had EGFR T790M mutation. Using our tumor genotyping panel, MET gene copy number gain was observed in two of 27 (7 %) samples. Twenty patients received continuous treatment with EGFR-TKIs even after disease progression, and 11 of these patients had T790M mutation in rebiopsy samples. In contrast, only 10 of 41 patients who finished EGFR-TKI treatment at disease progression had T790M mutation. The frequency of T790M mutation in patients who received continuous treatment with EGFR-TKIs after disease progression was significantly higher than that in patients who finished EGFR-TKI treatment at disease progression (55 % versus 24 %, p = 0.018). The frequency of T790M mutation in this study was lower than that in previous reports examining western patients. These results suggest that continuous treatment with EGFR-TKI after disease progression may enhance the frequency of EGFR T790M mutation in rebiopsy samples.
NASA Astrophysics Data System (ADS)
Tycko, Robert
2015-04-01
Twenty years ago, applications of solid state nuclear magnetic resonance (NMR) methods to real problems involving biological systems or biological materials were few and far between. Starting in the 1980s, a small number of research groups had begun to explore the possibility of obtaining structural and dynamical information about peptides, proteins, and other biopolymers from solid state NMR spectra. Progress was initially slow due to the relatively primitive state of solid state NMR probes, spectrometers, sample preparation methods, and pulse sequence techniques, coupled with the small number of people contributing to this research area. By the early 1990s, with the advent of new ideas about pulse sequence techniques such as dipolar recoupling, improvements in techniques for orienting membrane proteins and in technology for magic-angle spinning (MAS), improvements in the capabilities of commercial NMR spectrometers, and general developments in multidimensional spectroscopy, it began to appear that biomolecular solid state NMR might have a viable future. It was not until 1993 that the annual number of publications in this area crept above twenty.
Schumacher, Robin F; Malone, Amelia S
2017-09-01
The goal of the present study was to describe fraction-calculation errors among 4 th -grade students and determine whether error patterns differed as a function of problem type (addition vs. subtraction; like vs. unlike denominators), orientation (horizontal vs. vertical), or mathematics-achievement status (low- vs. average- vs. high-achieving). We specifically addressed whether mathematics-achievement status was related to students' tendency to operate with whole number bias. We extended this focus by comparing low-performing students' errors in two instructional settings that focused on two different types of fraction understandings: core instruction that focused on part-whole understanding vs. small-group tutoring that focused on magnitude understanding. Results showed students across the sample were more likely to operate with whole number bias on problems with unlike denominators. Students with low or average achievement (who only participated in core instruction) were more likely to operate with whole number bias than students with low achievement who participated in small-group tutoring. We suggest instruction should emphasize magnitude understanding to sufficiently increase fraction understanding for all students in the upper elementary grades.
A Statistical Treatment of Bioassay Pour Fractions
NASA Technical Reports Server (NTRS)
Barengoltz, Jack; Hughes, David W.
2014-01-01
The binomial probability distribution is used to treat the statistics of a microbiological sample that is split into two parts, with only one part evaluated for spore count. One wishes to estimate the total number of spores in the sample based on the counts obtained from the part that is evaluated (pour fraction). Formally, the binomial distribution is recharacterized as a function of the observed counts (successes), with the total number (trials) an unknown. The pour fraction is the probability of success per spore (trial). This distribution must be renormalized in terms of the total number. Finally, the new renormalized distribution is integrated and mathematically inverted to yield the maximum estimate of the total number as a function of a desired level of confidence ( P(
2011-01-01
Background Cochrane systematic reviews collate and summarise studies of the effects of healthcare interventions. The characteristics of these reviews and the meta-analyses and individual studies they contain provide insights into the nature of healthcare research and important context for the development of relevant statistical and other methods. Methods We classified every meta-analysis with at least two studies in every review in the January 2008 issue of the Cochrane Database of Systematic Reviews (CDSR) according to the medical specialty, the types of interventions being compared and the type of outcome. We provide descriptive statistics for numbers of meta-analyses, numbers of component studies and sample sizes of component studies, broken down by these categories. Results We included 2321 reviews containing 22,453 meta-analyses, which themselves consist of data from 112,600 individual studies (which may appear in more than one meta-analysis). Meta-analyses in the areas of gynaecology, pregnancy and childbirth (21%), mental health (13%) and respiratory diseases (13%) are well represented in the CDSR. Most meta-analyses address drugs, either with a control or placebo group (37%) or in a comparison with another drug (25%). The median number of meta-analyses per review is six (inter-quartile range 3 to 12). The median number of studies included in the meta-analyses with at least two studies is three (inter-quartile range 2 to 6). Sample sizes of individual studies range from 2 to 1,242,071, with a median of 91 participants. Discussion It is clear that the numbers of studies eligible for meta-analyses are typically very small for all medical areas, outcomes and interventions covered by Cochrane reviews. This highlights the particular importance of suitable methods for the meta-analysis of small data sets. There was little variation in number of studies per meta-analysis across medical areas, across outcome data types or across types of interventions being compared. PMID:22114982
Taguchi, Y-H
2016-05-10
MicroRNA(miRNA)-mRNA interactions are important for understanding many biological processes, including development, differentiation and disease progression, but their identification is highly context-dependent. When computationally derived from sequence information alone, the identification should be verified by integrated analyses of mRNA and miRNA expression. The drawback of this strategy is the vast number of identified interactions, which prevents an experimental or detailed investigation of each pair. In this paper, we overcome this difficulty by the recently proposed principal component analysis (PCA)-based unsupervised feature extraction (FE), which reduces the number of identified miRNA-mRNA interactions that properly discriminate between patients and healthy controls without losing biological feasibility. The approach is applied to six cancers: hepatocellular carcinoma, non-small cell lung cancer, esophageal squamous cell carcinoma, prostate cancer, colorectal/colon cancer and breast cancer. In PCA-based unsupervised FE, the significance does not depend on the number of samples (as in the standard case) but on the number of features, which approximates the number of miRNAs/mRNAs. To our knowledge, we have newly identified miRNA-mRNA interactions in multiple cancers based on a single common (universal) criterion. Moreover, the number of identified interactions was sufficiently small to be sequentially curated by literature searches.
Long-Term Follow-up of HPV Infection Using Urine and Cervical Quantitative HPV DNA Testing
Vorsters, Alex; Van Keer, Severien; Biesmans, Samantha; Hens, Annick; De Coster, Ilse; Goossens, Herman; Ieven, Margareta; Van Damme, Pierre
2016-01-01
The link between infection with high-risk human papillomavirus (hrHPV) and cervical cancer has been clearly demonstrated. Virological end-points showing the absence of persistent HPV infection are now accepted as a way of monitoring the impact of prophylactic vaccination programs and therapeutic vaccine trials. This study investigated the use of urine samples, which can be collected by self-sampling at home, instead of cervical samples for follow-up of an HPV intervention trial. Eighteen initially HPV DNA-positive women participating in an HPV therapeutic vaccine trial were monitored during a three-year follow-up period. A total of 172 urine samples and 85 cervical samples were collected. We obtained a paired urine sample for each of the 85 cervical samples by recovering urine samples from six monthly gynaecological examinations. We performed a small pilot study in which the participating women used a urine collection device at home and returned their urine sample to the laboratory by mail. All samples were analyzed using quantitative real-time HPV DNA PCR. A good association (κ value of 0.65) was found between the presence of HPV DNA in urine and a subsequent cervical sample. Comparisons of the number of HPV DNA copies in urine and paired cervical samples revealed a significant Spearman rho of 0.676. This correlation was superior in women with severe lesions. The HPV DNA results of the small pilot study based on self-collected urine samples at home are consistent with previous and subsequent urine and/or cervical results. We demonstrated that urine sampling may be a valid alternative to cervical samples for the follow-up of HPV intervention trials or programs. The potential clinical value of urine viral load monitoring should be further investigated. PMID:27196899
Long-Term Follow-up of HPV Infection Using Urine and Cervical Quantitative HPV DNA Testing.
Vorsters, Alex; Van Keer, Severien; Biesmans, Samantha; Hens, Annick; De Coster, Ilse; Goossens, Herman; Ieven, Margareta; Van Damme, Pierre
2016-05-17
The link between infection with high-risk human papillomavirus (hrHPV) and cervical cancer has been clearly demonstrated. Virological end-points showing the absence of persistent HPV infection are now accepted as a way of monitoring the impact of prophylactic vaccination programs and therapeutic vaccine trials. This study investigated the use of urine samples, which can be collected by self-sampling at home, instead of cervical samples for follow-up of an HPV intervention trial. Eighteen initially HPV DNA-positive women participating in an HPV therapeutic vaccine trial were monitored during a three-year follow-up period. A total of 172 urine samples and 85 cervical samples were collected. We obtained a paired urine sample for each of the 85 cervical samples by recovering urine samples from six monthly gynaecological examinations. We performed a small pilot study in which the participating women used a urine collection device at home and returned their urine sample to the laboratory by mail. All samples were analyzed using quantitative real-time HPV DNA PCR. A good association (κ value of 0.65) was found between the presence of HPV DNA in urine and a subsequent cervical sample. Comparisons of the number of HPV DNA copies in urine and paired cervical samples revealed a significant Spearman rho of 0.676. This correlation was superior in women with severe lesions. The HPV DNA results of the small pilot study based on self-collected urine samples at home are consistent with previous and subsequent urine and/or cervical results. We demonstrated that urine sampling may be a valid alternative to cervical samples for the follow-up of HPV intervention trials or programs. The potential clinical value of urine viral load monitoring should be further investigated.
Belitz, Kenneth; Jurgens, Bryant C.; Landon, Matthew K.; Fram, Miranda S.; Johnson, Tyler D.
2010-01-01
The proportion of an aquifer with constituent concentrations above a specified threshold (high concentrations) is taken as a nondimensional measure of regional scale water quality. If computed on the basis of area, it can be referred to as the aquifer scale proportion. A spatially unbiased estimate of aquifer scale proportion and a confidence interval for that estimate are obtained through the use of equal area grids and the binomial distribution. Traditionally, the confidence interval for a binomial proportion is computed using either the standard interval or the exact interval. Research from the statistics literature has shown that the standard interval should not be used and that the exact interval is overly conservative. On the basis of coverage probability and interval width, the Jeffreys interval is preferred. If more than one sample per cell is available, cell declustering is used to estimate the aquifer scale proportion, and Kish's design effect may be useful for estimating an effective number of samples. The binomial distribution is also used to quantify the adequacy of a grid with a given number of cells for identifying a small target, defined as a constituent that is present at high concentrations in a small proportion of the aquifer. Case studies illustrate a consistency between approaches that use one well per grid cell and many wells per cell. The methods presented in this paper provide a quantitative basis for designing a sampling program and for utilizing existing data.
Gautam, Aarti; Kumar, Raina; Dimitrov, George; Hoke, Allison; Hammamieh, Rasha; Jett, Marti
2016-10-01
miRNAs act as important regulators of gene expression by promoting mRNA degradation or by attenuating protein translation. Since miRNAs are stably expressed in bodily fluids, there is growing interest in profiling these miRNAs, as it is minimally invasive and cost-effective as a diagnostic matrix. A technical hurdle in studying miRNA dynamics is the ability to reliably extract miRNA as small sample volumes and low RNA abundance create challenges for extraction and downstream applications. The purpose of this study was to develop a pipeline for the recovery of miRNA using small volumes of archived serum samples. The RNA was extracted employing several widely utilized RNA isolation kits/methods with and without addition of a carrier. The small RNA library preparation was carried out using Illumina TruSeq small RNA kit and sequencing was carried out using Illumina platform. A fraction of five microliters of total RNA was used for library preparation as quantification is below the detection limit. We were able to profile miRNA levels in serum from all the methods tested. We found out that addition of nucleic acid based carrier molecules had higher numbers of processed reads but it did not enhance the mapping of any miRBase annotated sequences. However, some of the extraction procedures offer certain advantages: RNA extracted by TRIzol seemed to align to the miRBase best; extractions using TRIzol with carrier yielded higher miRNA-to-small RNA ratios. Nuclease free glycogen can be carrier of choice for miRNA sequencing. Our findings illustrate that miRNA extraction and quantification is influenced by the choice of methodologies. Addition of nucleic acid- based carrier molecules during extraction procedure is not a good choice when assaying miRNA using sequencing. The careful selection of an extraction method permits the archived serum samples to become valuable resources for high-throughput applications.
NASA Astrophysics Data System (ADS)
Theveneau, P.; Baker, R.; Barrett, R.; Beteva, A.; Bowler, M. W.; Carpentier, P.; Caserotto, H.; de Sanctis, D.; Dobias, F.; Flot, D.; Guijarro, M.; Giraud, T.; Lentini, M.; Leonard, G. A.; Mattenet, M.; McCarthy, A. A.; McSweeney, S. M.; Morawe, C.; Nanao, M.; Nurizzo, D.; Ohlsson, S.; Pernot, P.; Popov, A. N.; Round, A.; Royant, A.; Schmid, W.; Snigirev, A.; Surr, J.; Mueller-Dieckmann, C.
2013-03-01
Automation and advances in technology are the key elements in addressing the steadily increasing complexity of Macromolecular Crystallography (MX) experiments. Much of this complexity is due to the inter-and intra-crystal heterogeneity in diffraction quality often observed for crystals of multi-component macromolecular assemblies or membrane proteins. Such heterogeneity makes high-throughput sample evaluation an important and necessary tool for increasing the chances of a successful structure determination. The introduction at the ESRF of automatic sample changers in 2005 dramatically increased the number of samples that were tested for diffraction quality. This "first generation" of automation, coupled with advances in software aimed at optimising data collection strategies in MX, resulted in a three-fold increase in the number of crystal structures elucidated per year using data collected at the ESRF. In addition, sample evaluation can be further complemented using small angle scattering experiments on the newly constructed bioSAXS facility on BM29 and the micro-spectroscopy facility (ID29S). The construction of a second generation of automated facilities on the MASSIF (Massively Automated Sample Screening Integrated Facility) beam lines will build on these advances and should provide a paradigm shift in how MX experiments are carried out which will benefit the entire Structural Biology community.
NASA Astrophysics Data System (ADS)
Huang, Jian; Yuen, Pong C.; Chen, Wen-Sheng; Lai, J. H.
2005-05-01
Many face recognition algorithms/systems have been developed in the last decade and excellent performances have also been reported when there is a sufficient number of representative training samples. In many real-life applications such as passport identification, only one well-controlled frontal sample image is available for training. Under this situation, the performance of existing algorithms will degrade dramatically or may not even be implemented. We propose a component-based linear discriminant analysis (LDA) method to solve the one training sample problem. The basic idea of the proposed method is to construct local facial feature component bunches by moving each local feature region in four directions. In this way, we not only generate more samples with lower dimension than the original image, but also consider the face detection localization error while training. After that, we propose a subspace LDA method, which is tailor-made for a small number of training samples, for the local feature projection to maximize the discrimination power. Theoretical analysis and experiment results show that our proposed subspace LDA is efficient and overcomes the limitations in existing LDA methods. Finally, we combine the contributions of each local component bunch with a weighted combination scheme to draw the recognition decision. A FERET database is used for evaluating the proposed method and results are encouraging.
Gao, Jing; Zhong, Shaoyun; Zhou, Yanting; He, Han; Peng, Shuying; Zhu, Zhenyun; Liu, Xing; Zheng, Jing; Xu, Bin; Zhou, Hu
2017-06-06
Detergents and salts are widely used in lysis buffers to enhance protein extraction from biological samples, facilitating in-depth proteomic analysis. However, these detergents and salt additives must be efficiently removed from the digested samples prior to LC-MS/MS analysis to obtain high-quality mass spectra. Although filter-aided sample preparation (FASP), acetone precipitation (AP), followed by in-solution digestion, and strong cation exchange-based centrifugal proteomic reactors (CPRs) are commonly used for proteomic sample processing, little is known about their efficiencies at removing detergents and salt additives. In this study, we (i) developed an integrative workflow for the quantification of small molecular additives in proteomic samples, developing a multiple reaction monitoring (MRM)-based LC-MS approach for the quantification of six additives (i.e., Tris, urea, CHAPS, SDS, SDC, and Triton X-100) and (ii) systematically evaluated the relationships between the level of additive remaining in samples following sample processing and the number of peptides/proteins identified by mass spectrometry. Although FASP outperformed the other two methods, the results were complementary in terms of peptide/protein identification, as well as the GRAVY index and amino acid distributions. This is the first systematic and quantitative study of the effect of detergents and salt additives on protein identification. This MRM-based approach can be used for an unbiased evaluation of the performance of new sample preparation methods. Data are available via ProteomeXchange under identifier PXD005405.
Wright, Mark H.; Tung, Chih-Wei; Zhao, Keyan; Reynolds, Andy; McCouch, Susan R.; Bustamante, Carlos D.
2010-01-01
Motivation: The development of new high-throughput genotyping products requires a significant investment in testing and training samples to evaluate and optimize the product before it can be used reliably on new samples. One reason for this is current methods for automated calling of genotypes are based on clustering approaches which require a large number of samples to be analyzed simultaneously, or an extensive training dataset to seed clusters. In systems where inbred samples are of primary interest, current clustering approaches perform poorly due to the inability to clearly identify a heterozygote cluster. Results: As part of the development of two custom single nucleotide polymorphism genotyping products for Oryza sativa (domestic rice), we have developed a new genotype calling algorithm called ‘ALCHEMY’ based on statistical modeling of the raw intensity data rather than modelless clustering. A novel feature of the model is the ability to estimate and incorporate inbreeding information on a per sample basis allowing accurate genotyping of both inbred and heterozygous samples even when analyzed simultaneously. Since clustering is not used explicitly, ALCHEMY performs well on small sample sizes with accuracy exceeding 99% with as few as 18 samples. Availability: ALCHEMY is available for both commercial and academic use free of charge and distributed under the GNU General Public License at http://alchemy.sourceforge.net/ Contact: mhw6@cornell.edu Supplementary information: Supplementary data are available at Bioinformatics online. PMID:20926420
Yan, Junhao; Fan, Lingling; Zhao, Yueran; You, Li; Wang, Laicheng; Zhao, Han; Li, Yuan; Chen, Zi-Jiang
2011-12-01
To find the association between recurrent spontaneous abortion (RSA)/early embryo growth arrest and Y chromosome polymorphism. Peripheral blood samples of the male patients of big Y chromosome, small Y chromosome and other male patients whose partners suffered from unexplained RSA/early embryo growth arrest were collected. PCR and real-time fluorescent quantitative PCR were used to test the deletion and the copy number variation of DYZ1 region in Y chromosome of the patients. A total of 79 big Y chromosome patients (48 of whose partners suffered from RSA or early embryo growth arrest), 7 small Y chromosome patients, 106 other male patients whose partners had suffered from unexplained RSA or early embryo growth arrest, and 100 normal male controls were enrolled. There was no fraction deletion of DYZ1 detected both in big Y patients and in normal men. Of RSA patients, 1 case showed deletion of 266bp from the gene locus 25-290bp, and 2 cases showed deletion of 773bp from 1347 to 2119bp. Of only 7 small Y chromosome patients, 2 cases showed deletion of 266bp from 25 to 290bp, and 4 cases showed deletion of 773bp from 1347 to 2119bp and 275bp from 3128 to 3420bp. The mean of DYZ1 copies was 3900 in normal control men; the mean in big Y patients was 5571, in RSA patients was 2655, and in small Y patients was 1059. All of the others were significantly different (P<0.01) compared with normal control men, which meant that DYZ1 copy number in normal control men was less than that of big Y chromosome patients, and was more than that of unexplained early RSA patients and small Y patients. The integrity and copy number variation of DYZ1 are closely related to the Y chromosome length under microscope. The cause of RSA/early embryo growth arrest in some couples may be the increase (big Y patients) or decrease of DYZ1 copy number in the husbands' Y chromosome. Copyright © 2011 Elsevier Ireland Ltd. All rights reserved.
Detection of the Emerging Picornavirus Senecavirus A in Pigs, Mice, and Houseflies.
Joshi, Lok R; Mohr, Kristin A; Clement, Travis; Hain, Kyle S; Myers, Bryan; Yaros, Joseph; Nelson, Eric A; Christopher-Hennings, Jane; Gava, Danielle; Schaefer, Rejane; Caron, Luizinho; Dee, Scott; Diel, Diego G
2016-06-01
Senecavirus A (SVA) is an emerging picornavirus that has been recently associated with an increased number of outbreaks of vesicular disease and neonatal mortality in swine. Many aspects of SVA infection biology and epidemiology remain unknown. Here, we present a diagnostic investigation conducted in swine herds affected by vesicular disease and increased neonatal mortality. Clinical and environmental samples were collected from affected and unaffected herds and were screened for the presence of SVA by real-time reverse transcriptase PCR and virus isolation. Notably, SVA was detected and isolated from vesicular lesions and tissues of affected pigs, environmental samples, mouse feces, and mouse small intestine. SVA nucleic acid was also detected in houseflies collected from affected farms and from a farm with no history of vesicular disease. Detection of SVA in mice and housefly samples and recovery of viable virus from mouse feces and small intestine suggest that these pests may play a role on the epidemiology of SVA. These results provide important information that may allow the development of improved prevention and control strategies for SVA. Copyright © 2016, American Society for Microbiology. All Rights Reserved.
Detection of the Emerging Picornavirus Senecavirus A in Pigs, Mice, and Houseflies
Joshi, Lok R.; Mohr, Kristin A.; Clement, Travis; Hain, Kyle S.; Myers, Bryan; Yaros, Joseph; Nelson, Eric A.; Christopher-Hennings, Jane; Gava, Danielle; Schaefer, Rejane; Caron, Luizinho; Dee, Scott
2016-01-01
Senecavirus A (SVA) is an emerging picornavirus that has been recently associated with an increased number of outbreaks of vesicular disease and neonatal mortality in swine. Many aspects of SVA infection biology and epidemiology remain unknown. Here, we present a diagnostic investigation conducted in swine herds affected by vesicular disease and increased neonatal mortality. Clinical and environmental samples were collected from affected and unaffected herds and were screened for the presence of SVA by real-time reverse transcriptase PCR and virus isolation. Notably, SVA was detected and isolated from vesicular lesions and tissues of affected pigs, environmental samples, mouse feces, and mouse small intestine. SVA nucleic acid was also detected in houseflies collected from affected farms and from a farm with no history of vesicular disease. Detection of SVA in mice and housefly samples and recovery of viable virus from mouse feces and small intestine suggest that these pests may play a role on the epidemiology of SVA. These results provide important information that may allow the development of improved prevention and control strategies for SVA. PMID:27030489
Bentzen, Amalie Kai; Marquard, Andrea Marion; Lyngaa, Rikke; Saini, Sunil Kumar; Ramskov, Sofie; Donia, Marco; Such, Lina; Furness, Andrew J S; McGranahan, Nicholas; Rosenthal, Rachel; Straten, Per Thor; Szallasi, Zoltan; Svane, Inge Marie; Swanton, Charles; Quezada, Sergio A; Jakobsen, Søren Nyboe; Eklund, Aron Charles; Hadrup, Sine Reker
2016-10-01
Identification of the peptides recognized by individual T cells is important for understanding and treating immune-related diseases. Current cytometry-based approaches are limited to the simultaneous screening of 10-100 distinct T-cell specificities in one sample. Here we use peptide-major histocompatibility complex (MHC) multimers labeled with individual DNA barcodes to screen >1,000 peptide specificities in a single sample, and detect low-frequency CD8 T cells specific for virus- or cancer-restricted antigens. When analyzing T-cell recognition of shared melanoma antigens before and after adoptive cell therapy in melanoma patients, we observe a greater number of melanoma-specific T-cell populations compared with cytometry-based approaches. Furthermore, we detect neoepitope-specific T cells in tumor-infiltrating lymphocytes and peripheral blood from patients with non-small cell lung cancer. Barcode-labeled pMHC multimers enable the combination of functional T-cell analysis with large-scale epitope recognition profiling for the characterization of T-cell recognition in various diseases, including in small clinical samples.
NASA Astrophysics Data System (ADS)
Li, Hui; Yu, Jun-Ling; Yu, Le-An; Sun, Jie
2014-05-01
Case-based reasoning (CBR) is one of the main forecasting methods in business forecasting, which performs well in prediction and holds the ability of giving explanations for the results. In business failure prediction (BFP), the number of failed enterprises is relatively small, compared with the number of non-failed ones. However, the loss is huge when an enterprise fails. Therefore, it is necessary to develop methods (trained on imbalanced samples) which forecast well for this small proportion of failed enterprises and performs accurately on total accuracy meanwhile. Commonly used methods constructed on the assumption of balanced samples do not perform well in predicting minority samples on imbalanced samples consisting of the minority/failed enterprises and the majority/non-failed ones. This article develops a new method called clustering-based CBR (CBCBR), which integrates clustering analysis, an unsupervised process, with CBR, a supervised process, to enhance the efficiency of retrieving information from both minority and majority in CBR. In CBCBR, various case classes are firstly generated through hierarchical clustering inside stored experienced cases, and class centres are calculated out by integrating cases information in the same clustered class. When predicting the label of a target case, its nearest clustered case class is firstly retrieved by ranking similarities between the target case and each clustered case class centre. Then, nearest neighbours of the target case in the determined clustered case class are retrieved. Finally, labels of the nearest experienced cases are used in prediction. In the empirical experiment with two imbalanced samples from China, the performance of CBCBR was compared with the classical CBR, a support vector machine, a logistic regression and a multi-variant discriminate analysis. The results show that compared with the other four methods, CBCBR performed significantly better in terms of sensitivity for identifying the minority samples and generated high total accuracy meanwhile. The proposed approach makes CBR useful in imbalanced forecasting.
NASA Astrophysics Data System (ADS)
Kalb, Wolfgang L.; Haas, Simon; Krellner, Cornelius; Mathis, Thomas; Batlogg, Bertram
2010-04-01
We show that it is possible to reach one of the ultimate goals of organic electronics: producing organic field-effect transistors with trap densities as low as in the bulk of single crystals. We studied the spectral density of localized states in the band gap [trap density of states (trap DOS)] of small-molecule organic semiconductors as derived from electrical characteristics of organic field-effect transistors or from space-charge-limited current measurements. This was done by comparing data from a large number of samples including thin-film transistors (TFT’s), single crystal field-effect transistors (SC-FET’s) and bulk samples. The compilation of all data strongly suggests that structural defects associated with grain boundaries are the main cause of “fast” hole traps in TFT’s made with vacuum-evaporated pentacene. For high-performance transistors made with small-molecule semiconductors such as rubrene it is essential to reduce the dipolar disorder caused by water adsorbed on the gate dielectric surface. In samples with very low trap densities, we sometimes observe a steep increase in the trap DOS very close (<0.15eV) to the mobility edge with a characteristic slope of 10-20 meV. It is discussed to what degree band broadening due to the thermal fluctuation of the intermolecular transfer integral is reflected in this steep increase in the trap DOS. Moreover, we show that the trap DOS in TFT’s with small-molecule semiconductors is very similar to the trap DOS in hydrogenated amorphous silicon even though polycrystalline films of small-molecules with van der Waals-type interaction on the one hand are compared with covalently bound amorphous silicon on the other hand.
Small Business Procurement Event
2014-08-13
Small Business Procurement Event 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) 5d. PROJECT NUMBER 5e. TASK...NUMBER 5f. WORK UNIT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) Department of the Navy,Office of Small Business Programs,720 Kennon...distribution unlimited 13. SUPPLEMENTARY NOTES NDIA 27th Navy Gold Coast Small Business Procurement Event, 12-13 Aug 2014, San Diego, CA. 14. ABSTRACT
2011-01-01
To obtain approval for the use vertebrate animals in research, an investigator must assure an ethics committee that the proposed number of animals is the minimum necessary to achieve a scientific goal. How does an investigator make that assurance? A power analysis is most accurate when the outcome is known before the study, which it rarely is. A ‘pilot study’ is appropriate only when the number of animals used is a tiny fraction of the numbers that will be invested in the main study because the data for the pilot animals cannot legitimately be used again in the main study without increasing the rate of type I errors (false discovery). Traditional significance testing requires the investigator to determine the final sample size before any data are collected and then to delay analysis of any of the data until all of the data are final. An investigator often learns at that point either that the sample size was larger than necessary or too small to achieve significance. Subjects cannot be added at this point in the study without increasing type I errors. In addition, journal reviewers may require more replications in quantitative studies than are truly necessary. Sequential stopping rules used with traditional significance tests allow incremental accumulation of data on a biomedical research problem so that significance, replicability, and use of a minimal number of animals can be assured without increasing type I errors. PMID:21838970
Micro-CT scouting for transmission electron microscopy of human tissue specimens
DOE Office of Scientific and Technical Information (OSTI.GOV)
Morales, A. G.; Stempinski, E. S.; XIAO, X.
Transmission electron microscopy (TEM) provides sub-nanometre-scale details in volumetric samples. Samples such as pathology tissue specimens are often stained with a metal element to enhance contrast, which makes them opaque to optical microscopes. As a result, it can be a lengthy procedure to find the region of interest inside a sample through sectioning. Here, we describe micro-CT scouting for TEM that allows noninvasive identification of regions of interest within a block sample to guide the sectioning step. In a tissue pathology study, a bench-top micro-CT scanner with 10 m resolution was used to determine the location of patches of themore » mucous membrane in osmium-stained human nasal scraping samples. Furthermore, once the regions of interest were located, the sample block was sectioned to expose that location, followed by ultra-thin sectioning and TEM to inspect the internal structure of the cilia of the membrane epithelial cells with nanometre resolution. This method substantially reduced the time and labour of the search process from typically 20 sections for light microscopy to three sections with no added sample preparation. Lay description Electron microscopy provides very high levels of detail in a small area, and thus the question of where to look in an opaque sample, such as a stained tissue specimen, needs to be answered by sectioning the sample in small steps and examining the sections under a light microscope, until the region of interest is found. The search process can be lengthy and labor intensive, especially for a study involving a large number of samples. Small areas of interest can be missed in the process if not enough regions are examined. We also describe a method to directly locate the region of interest within a whole sample using micro-CT imaging, bypassing the need of blindly sectioning. Micro-CT enables locating the region within 3D space; this information provides a guide for sectioning the sample to expose that precise location for high resolution electron microscopy imaging. In a human tissue specimen study, this method considerably reduced the time and labor of the search process.« less
Micro-CT scouting for transmission electron microscopy of human tissue specimens
Morales, A. G.; Stempinski, E. S.; XIAO, X.; ...
2016-02-08
Transmission electron microscopy (TEM) provides sub-nanometre-scale details in volumetric samples. Samples such as pathology tissue specimens are often stained with a metal element to enhance contrast, which makes them opaque to optical microscopes. As a result, it can be a lengthy procedure to find the region of interest inside a sample through sectioning. Here, we describe micro-CT scouting for TEM that allows noninvasive identification of regions of interest within a block sample to guide the sectioning step. In a tissue pathology study, a bench-top micro-CT scanner with 10 m resolution was used to determine the location of patches of themore » mucous membrane in osmium-stained human nasal scraping samples. Furthermore, once the regions of interest were located, the sample block was sectioned to expose that location, followed by ultra-thin sectioning and TEM to inspect the internal structure of the cilia of the membrane epithelial cells with nanometre resolution. This method substantially reduced the time and labour of the search process from typically 20 sections for light microscopy to three sections with no added sample preparation. Lay description Electron microscopy provides very high levels of detail in a small area, and thus the question of where to look in an opaque sample, such as a stained tissue specimen, needs to be answered by sectioning the sample in small steps and examining the sections under a light microscope, until the region of interest is found. The search process can be lengthy and labor intensive, especially for a study involving a large number of samples. Small areas of interest can be missed in the process if not enough regions are examined. We also describe a method to directly locate the region of interest within a whole sample using micro-CT imaging, bypassing the need of blindly sectioning. Micro-CT enables locating the region within 3D space; this information provides a guide for sectioning the sample to expose that precise location for high resolution electron microscopy imaging. In a human tissue specimen study, this method considerably reduced the time and labor of the search process.« less
Fearon, Elizabeth; Chabata, Sungai T; Thompson, Jennifer A; Cowan, Frances M; Hargreaves, James R
2017-09-14
While guidance exists for obtaining population size estimates using multiplier methods with respondent-driven sampling surveys, we lack specific guidance for making sample size decisions. To guide the design of multiplier method population size estimation studies using respondent-driven sampling surveys to reduce the random error around the estimate obtained. The population size estimate is obtained by dividing the number of individuals receiving a service or the number of unique objects distributed (M) by the proportion of individuals in a representative survey who report receipt of the service or object (P). We have developed an approach to sample size calculation, interpreting methods to estimate the variance around estimates obtained using multiplier methods in conjunction with research into design effects and respondent-driven sampling. We describe an application to estimate the number of female sex workers in Harare, Zimbabwe. There is high variance in estimates. Random error around the size estimate reflects uncertainty from M and P, particularly when the estimate of P in the respondent-driven sampling survey is low. As expected, sample size requirements are higher when the design effect of the survey is assumed to be greater. We suggest a method for investigating the effects of sample size on the precision of a population size estimate obtained using multipler methods and respondent-driven sampling. Uncertainty in the size estimate is high, particularly when P is small, so balancing against other potential sources of bias, we advise researchers to consider longer service attendance reference periods and to distribute more unique objects, which is likely to result in a higher estimate of P in the respondent-driven sampling survey. ©Elizabeth Fearon, Sungai T Chabata, Jennifer A Thompson, Frances M Cowan, James R Hargreaves. Originally published in JMIR Public Health and Surveillance (http://publichealth.jmir.org), 14.09.2017.
Diffraction Pattern Analysis as an Optical Inspection Technique
1991-08-01
BACKGROUND Diameters of fiber samples have commonly been measured manually with an optical microscope. Marcuse and Presby developed an automatic...by analyzing the back-scattered light when a beam of laser light impinged upon the fiber [2]. Presby and Marcuse extended this back-scattering tech...be im- proved further in order to become a feasible method for detecting a small number of blocked openings in CRT screens. 20 REFERENCES 1. Marcuse
Cipora, Krzysztof; Nuerk, Hans-Christoph
2013-01-01
The SNARC (spatial-numerical association of response codes) described that larger numbers are responded faster with the right hand and smaller numbers with the left hand. It is held in the literature that arithmetically skilled and nonskilled adults differ in the SNARC. However, the respective data are descriptive, and the decisive tests are nonsignificant. Possible reasons for this nonsignificance could be that in previous studies (a) very small samples were used, (b) there were too few repetitions producing too little power and, consequently, reliabilities that were too small to reach conventional significance levels for the descriptive skill differences in the SNARC, and (c) general mathematical ability was assessed by the field of study of students, while individual arithmetic skills were not examined. Therefore we used a much bigger sample, a lot more repetitions, and direct assessment of arithmetic skills to explore relations between the SNARC effect and arithmetic skills. Nevertheless, a difference in SNARC effect between arithmetically skilled and nonskilled participants was not obtained. Bayesian analysis showed positive evidence of a true null effect, not just a power problem. Hence we conclude that the idea that arithmetically skilled and nonskilled participants generally differ in the SNARC effect is not warranted by our data.
Perry, G.; Rodda, G.H.; Fritts, T.H.; Sharp, T.R.
1998-01-01
We sampled the lizard fauna of twenty-two small islets fringing the Pacific island of Guam and used these data to shed light on the processes responsible for present-day diversity. Habitat diversity, measured by islet area and vegetation complexity, was significantly correlated with the number of species found on an islet. However, islet distance and elevation were not significant predictors of diversity. Distribution patterns were slightly different for the two major families in our sample, Scincidae and Gekkonidae: skinks needed larger islets to maintain a population than did geckos. Presence/absence patterns were highly and significantly nested, and population density was correlated with the number of islets on which a species was found. An area cladogram was poorly supported and showed no faunal similarity between nearby islands. These patterns indicate that extinctions on most islets were due mostly to non-catastrophic, long-acting biological causes. The presence on the islets of species extirpated on Guam and the lack of significant nestedness on islands with greater maximum elevation highlight the impact that predators (primarily brown treesnakes) can have. Our findings also show that small reserves will not suffice to protect endangered lizard faunas, and that the islets may serve as a short-term repository of such species until snake-free areas can be established on Guam.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Robison, W.L.; Conrado, C.L.
1996-03-01
The data presented in the following tables is the total available for each northern island; they include both the data from the 1978 Northern Marshall Island Radiological Survey (NMIRS) and trips to Rongelap Atoll from 1986 through 1989. In one table we present the number of vegetation samples collected in the 1978 NMIRS and from 1986 through 1989. Again the majority of the {sup 137}Cs data is from the 1986-1989 trips. We have not made additional analyses of {sup 239+240}Pu, {sup 241}Am and {sup 90}Sr because the concentrations are very low and these radionuclides contribute less than 5% of anmore » already very small dose. In another table we show the number of soil samples collected at each island in 1978 and the number collected since 1986. Most of the data are from 1986 through 1989. The major exception is {sup 90}Sr where all of the data are from the 1978 NMIRS. We have done some additional Pu analyses of surface soils (0-5 cm depth) in the northern islands. A significant amount of new data for {sup 137}Cs and {sup 241}Am have been generated from the samples collected from 1986 through 1989. The data are presented in the form of summary tables, graphics, detailed appendices and aerial photographs of the islands with the sample locations marked. The identified sample locations from the 1978 NMIRS will be added later.« less
Thompson, W.L.
2003-01-01
Hankin and Reeves' (1988) approach to estimating fish abundance in small streams has been applied in stream fish studies across North America. However, their population estimator relies on two key assumptions: (1) removal estimates are equal to the true numbers of fish, and (2) removal estimates are highly correlated with snorkel counts within a subset of sampled stream units. Violations of these assumptions may produce suspect results. To determine possible sources of the assumption violations, I used data on the abundance of steelhead Oncorhynchus mykiss from Hankin and Reeves' (1988) in a simulation composed of 50,000 repeated, stratified systematic random samples from a spatially clustered distribution. The simulation was used to investigate effects of a range of removal estimates, from 75% to 100% of true fish abundance, on overall stream fish population estimates. The effects of various categories of removal-estimates-to-snorkel-count correlation levels (r = 0.75-1.0) on fish population estimates were also explored. Simulation results indicated that Hankin and Reeves' approach may produce poor results unless removal estimates exceed at least 85% of the true number of fish within sampled units and unless correlations between removal estimates and snorkel counts are at least 0.90. A potential modification to Hankin and Reeves' approach is the inclusion of environmental covariates that affect detection rates of fish into the removal model or other mark-recapture model. A potential alternative approach is to use snorkeling combined with line transect sampling to estimate fish densities within stream units. As with any method of population estimation, a pilot study should be conducted to evaluate its usefulness, which requires a known (or nearly so) population of fish to serve as a benchmark for evaluating bias and precision of estimators.
Magnotti, John F; Basu Mallick, Debshila; Feng, Guo; Zhou, Bin; Zhou, Wen; Beauchamp, Michael S
2015-09-01
Humans combine visual information from mouth movements with auditory information from the voice to recognize speech. A common method for assessing multisensory speech perception is the McGurk effect: When presented with particular pairings of incongruent auditory and visual speech syllables (e.g., the auditory speech sounds for "ba" dubbed onto the visual mouth movements for "ga"), individuals perceive a third syllable, distinct from the auditory and visual components. Chinese and American cultures differ in the prevalence of direct facial gaze and in the auditory structure of their languages, raising the possibility of cultural- and language-related group differences in the McGurk effect. There is no consensus in the literature about the existence of these group differences, with some studies reporting less McGurk effect in native Mandarin Chinese speakers than in English speakers and others reporting no difference. However, these studies sampled small numbers of participants tested with a small number of stimuli. Therefore, we collected data on the McGurk effect from large samples of Mandarin-speaking individuals from China and English-speaking individuals from the USA (total n = 307) viewing nine different stimuli. Averaged across participants and stimuli, we found similar frequencies of the McGurk effect between Chinese and American participants (48 vs. 44 %). In both groups, we observed a large range of frequencies both across participants (range from 0 to 100 %) and stimuli (15 to 83 %) with the main effect of culture and language accounting for only 0.3 % of the variance in the data. High individual variability in perception of the McGurk effect necessitates the use of large sample sizes to accurately estimate group differences.
Average of delta: a new quality control tool for clinical laboratories.
Jones, Graham R D
2016-01-01
Average of normals is a tool used to control assay performance using the average of a series of results from patients' samples. Delta checking is a process of identifying errors in individual patient results by reviewing the difference from previous results of the same patient. This paper introduces a novel alternate approach, average of delta, which combines these concepts to use the average of a number of sequential delta values to identify changes in assay performance. Models for average of delta and average of normals were developed in a spreadsheet application. The model assessed the expected scatter of average of delta and average of normals functions and the effect of assay bias for different values of analytical imprecision and within- and between-subject biological variation and the number of samples included in the calculations. The final assessment was the number of patients' samples required to identify an added bias with 90% certainty. The model demonstrated that with larger numbers of delta values, the average of delta function was tighter (lower coefficient of variation). The optimal number of samples for bias detection with average of delta was likely to be between 5 and 20 for most settings and that average of delta outperformed average of normals when the within-subject biological variation was small relative to the between-subject variation. Average of delta provides a possible additional assay quality control tool which theoretical modelling predicts may be more valuable than average of normals for analytes where the group biological variation is wide compared with within-subject variation and where there is a high rate of repeat testing in the laboratory patient population. © The Author(s) 2015.
Radiological conditions at the Southern Islands of Rongelap Atoll
DOE Office of Scientific and Technical Information (OSTI.GOV)
Robison, W.L.; Conrado, C.L.
1996-03-01
The data presented in the following tables is the total available for each southern island; they include both the data from the 1978 Northern Marshall Island Radiological Survey (NMIRS) and trips to Rongelap Atoll from 1986 through 1991. There are additional samples that were taken at Rongelap Island in 1990 and 1991, and the data are unavailable for this report. In one table we present the number of vegetation samples collected in the 1978 NMIRS and from 1986 through 1991. Again, the majority of the {sup 137}Cs is from the 1986-1991 trips. We have not made additional analyses of {supmore » 239+240}Pu, {sup 241}Am and {sup 90}Sr because the concentrations are very low and these radionuclides contribute less than 5% of an already very small dose. In another table we show the number of soil samples collected at each island in 1978 and the number collected since 1986. Most of the data are from 1986 through 1991. The major exception is {sup 90}Sr where all of the data are from the 1978 NMIRS. We have done some additional Pu analyses of soils from Rongelap Eniaetok, and Borukka Island but none of the other southern islands. A significant amount of new data for {sup 137}Cs and {sup 241}Am have been generated from the samples collected from 1986 through 1991. The data are presented in the form of summary tables, graphics, detailed appendices and aerial photographs of the islands with the sample locations marked. The identified sample locations from the 1978 NMIRS will be added later.« less
NASA Astrophysics Data System (ADS)
Janniche, G. S.; Mouvet, C.; Albrechtsen, H.-J.
2011-04-01
Vertical variation in sorption and mineralization potential of mecoprop (MCPP), isoproturon and acetochlor were investigated at low concentrations (μg-range) at the cm-scale in unsaturated sub-surface limestone samples and saturated sandy aquifer samples from an agricultural catchment in Brévilles, France. From two intact core drills, four heterogenic limestone sections were collected from 4.50 to 26.40 m below surface (mbs) and divided into 12 sub-samples of 8-25 cm length, and one sandy aquifer section from 19.20 to 19.53 m depth divided into 7 sub-samples of 4-5 cm length. In the sandy aquifer section acetochlor and isoproturon sorption increased substantially with depth; in average 78% (acetochlor) and 61% (isoproturon) per 5 cm. Also the number of acetochlor and isoproturon degraders (most-probable-number) was higher in the bottom half of the aquifer section (93-> 16 000/g) than in the upper half (4-71/g). One 50 cm long limestone section with a distinct shift in color showed a clear shift in mineralization, number of degraders and sorption: In the two brown, uppermost samples, up to 31% mecoprop and up to 9% isoproturon was mineralized during 231 days, the numbers of mecoprop and isoproturon degraders were 1300 to > 16 000/g, and the sorption of both isoproturon and acetochlor was more than three times higher, compared to the two deeper, grayish samples just below where mineralization (≤ 4%) and numbers of degraders (1-520/g) were low for all three herbicides. In both unsaturated limestone and sandy aquifer, variations and even distinct shifts in both mineralization, number of specific degraders and sorption were seen within just 4-15 cm of vertical distance. A simple conceptual model of herbicides leaching to groundwater through a 10 m unsaturated limestone was established, and calculations showed that a 30 cm active layer with the measured sorption and mineralization values hardly impacted the fate of the investigated herbicides, whereas a total thickness of layers of 1 m would substantially increase natural attenuation.
Borges, Chad R
2007-07-01
A chemometrics-based data analysis concept has been developed as a substitute for manual inspection of extracted ion chromatograms (XICs), which facilitates rapid, analyst-mediated interpretation of GC- and LC/MS(n) data sets from samples undergoing qualitative batchwise screening for prespecified sets of analytes. Automatic preparation of data into two-dimensional row space-derived scatter plots (row space plots) eliminates the need to manually interpret hundreds to thousands of XICs per batch of samples while keeping all interpretation of raw data directly in the hands of the analyst-saving great quantities of human time without loss of integrity in the data analysis process. For a given analyte, two analyte-specific variables are automatically collected by a computer algorithm and placed into a data matrix (i.e., placed into row space): the first variable is the ion abundance corresponding to scan number x and analyte-specific m/z value y, and the second variable is the ion abundance corresponding to scan number x and analyte-specific m/z value z (a second ion). These two variables serve as the two axes of the aforementioned row space plots. In order to collect appropriate scan number (retention time) information, it is necessary to analyze, as part of every batch, a sample containing a mixture of all analytes to be tested. When pure standard materials of tested analytes are unavailable, but representative ion m/z values are known and retention time can be approximated, data are evaluated based on two-dimensional scores plots from principal component analysis of small time range(s) of mass spectral data. The time-saving efficiency of this concept is directly proportional to the percentage of negative samples and to the total number of samples processed simultaneously.
Wu, Dongrui; Lance, Brent J; Parsons, Thomas D
2013-01-01
Brain-computer interaction (BCI) and physiological computing are terms that refer to using processed neural or physiological signals to influence human interaction with computers, environment, and each other. A major challenge in developing these systems arises from the large individual differences typically seen in the neural/physiological responses. As a result, many researchers use individually-trained recognition algorithms to process this data. In order to minimize time, cost, and barriers to use, there is a need to minimize the amount of individual training data required, or equivalently, to increase the recognition accuracy without increasing the number of user-specific training samples. One promising method for achieving this is collaborative filtering, which combines training data from the individual subject with additional training data from other, similar subjects. This paper describes a successful application of a collaborative filtering approach intended for a BCI system. This approach is based on transfer learning (TL), active class selection (ACS), and a mean squared difference user-similarity heuristic. The resulting BCI system uses neural and physiological signals for automatic task difficulty recognition. TL improves the learning performance by combining a small number of user-specific training samples with a large number of auxiliary training samples from other similar subjects. ACS optimally selects the classes to generate user-specific training samples. Experimental results on 18 subjects, using both k nearest neighbors and support vector machine classifiers, demonstrate that the proposed approach can significantly reduce the number of user-specific training data samples. This collaborative filtering approach will also be generalizable to handling individual differences in many other applications that involve human neural or physiological data, such as affective computing.
Wu, Dongrui; Lance, Brent J.; Parsons, Thomas D.
2013-01-01
Brain-computer interaction (BCI) and physiological computing are terms that refer to using processed neural or physiological signals to influence human interaction with computers, environment, and each other. A major challenge in developing these systems arises from the large individual differences typically seen in the neural/physiological responses. As a result, many researchers use individually-trained recognition algorithms to process this data. In order to minimize time, cost, and barriers to use, there is a need to minimize the amount of individual training data required, or equivalently, to increase the recognition accuracy without increasing the number of user-specific training samples. One promising method for achieving this is collaborative filtering, which combines training data from the individual subject with additional training data from other, similar subjects. This paper describes a successful application of a collaborative filtering approach intended for a BCI system. This approach is based on transfer learning (TL), active class selection (ACS), and a mean squared difference user-similarity heuristic. The resulting BCI system uses neural and physiological signals for automatic task difficulty recognition. TL improves the learning performance by combining a small number of user-specific training samples with a large number of auxiliary training samples from other similar subjects. ACS optimally selects the classes to generate user-specific training samples. Experimental results on 18 subjects, using both nearest neighbors and support vector machine classifiers, demonstrate that the proposed approach can significantly reduce the number of user-specific training data samples. This collaborative filtering approach will also be generalizable to handling individual differences in many other applications that involve human neural or physiological data, such as affective computing. PMID:23437188
Green, Charlotte J; Charlton, Catriona A; Wang, Lai-Mun; Silva, Michael; Morten, Karl J; Hodson, Leanne
2017-12-01
Two-step perfusion is considered the gold standard method for isolating hepatocytes from human liver tissue. As perfusion may require a large tissue specimen, which is encapsulated and has accessible vessels for cannulation, only a limited number of tissue samples may be suitable. Therefore, the aim of this work was to develop an alternative method to isolate hepatocytes from non-encapsulated and small samples of human liver tissue. Healthy tissue from 44 human liver resections were graded for steatosis and tissue weights between 7.8 and 600 g were used for hepatocyte isolations. Tissue was diced and underwent a two-step digestion (EDTA and collagenase). Red cell lysis buffer was used to prevent red blood cell contamination and toxicity. Isolated hepatocyte viability was determined by trypan blue exclusion. Western blot and biochemical analyses were undertaken to ascertain cellular phenotype and function. Liver tissue that weighed ≥50 g yielded significantly higher (P < 0.01) cell viability than tissue <50 g. Viable cells secreted urea and displayed the phenotypic hepatocyte markers albumin and cytochrome P450. Presence of steatosis in liver tissue or intra-hepatocellular triglyceride content had no effect on cell viability. This methodology allows for the isolation of viable primary human hepatocytes from small amounts of "healthy" resected liver tissue which are not suitable for perfusion. This work provides the opportunity to increase the utilisation of resection surplus tissue, and may ultimately lead to an increased number of in vitro cellular studies being undertaken using the gold-standard model of human primary hepatocytes.
Cannistraci, Carlo Vittorio; Ravasi, Timothy; Montevecchi, Franco Maria; Ideker, Trey; Alessio, Massimo
2010-09-15
Nonlinear small datasets, which are characterized by low numbers of samples and very high numbers of measures, occur frequently in computational biology, and pose problems in their investigation. Unsupervised hybrid-two-phase (H2P) procedures-specifically dimension reduction (DR), coupled with clustering-provide valuable assistance, not only for unsupervised data classification, but also for visualization of the patterns hidden in high-dimensional feature space. 'Minimum Curvilinearity' (MC) is a principle that-for small datasets-suggests the approximation of curvilinear sample distances in the feature space by pair-wise distances over their minimum spanning tree (MST), and thus avoids the introduction of any tuning parameter. MC is used to design two novel forms of nonlinear machine learning (NML): Minimum Curvilinear embedding (MCE) for DR, and Minimum Curvilinear affinity propagation (MCAP) for clustering. Compared with several other unsupervised and supervised algorithms, MCE and MCAP, whether individually or combined in H2P, overcome the limits of classical approaches. High performance was attained in the visualization and classification of: (i) pain patients (proteomic measurements) in peripheral neuropathy; (ii) human organ tissues (genomic transcription factor measurements) on the basis of their embryological origin. MC provides a valuable framework to estimate nonlinear distances in small datasets. Its extension to large datasets is prefigured for novel NMLs. Classification of neuropathic pain by proteomic profiles offers new insights for future molecular and systems biology characterization of pain. Improvements in tissue embryological classification refine results obtained in an earlier study, and suggest a possible reinterpretation of skin attribution as mesodermal. https://sites.google.com/site/carlovittoriocannistraci/home.
Trichinella nelsoni in carnivores from the Serengeti ecosystem, Tanzania.
Pozio, E; De Meneghi, D; Roelke-Parker, M E; La Rosa, G
1997-12-01
A survey of trichinellosis among sylvatic carnivore mammals from the Serengeti ecosystem (Tanzania) demonstrated the presence of Trichinella nelsoni in 5 of 9 species examined. Muscle samples were collected from carcasses of 56 carnivores from 1993 to 1995 and frozen before transport and examination. Following artificial digestion of the samples, collected larvae were analyzed by the random amplified polymorphic DNA technique. Trichinella nelsoni was identified in 1 bat-eared fox (Otocyon megalotis), 1 cheetah (Acinonyx jubatus), 1 leopard (Panthera pardus), 3 lions (Panthera leo), and 3 spotted hyenas (Crocuta crocuta). The numbers of bat-eared foxes (6), cheetahs (5), and leopards (3) examined were too small to reveal the roles of these carnivore species in the ecology of T. nelsoni. The numbers of lions and spotted hyenas examined, with a prevalence of 12% and 23%, respectively, suggest that these species may be reservoirs of T. nelsoni in the area under study.
NASA Technical Reports Server (NTRS)
Wasson, John T.; Ouyang, Xinwei; Wang, Jianmin; Jerde, Eric
1989-01-01
Concentrations of 14 elements in the metal of 38 iron meteorites and a pallasite are reported. Three samples are paired with previously classified irons, raising the number of well-classified, independent iron meteorites to 598. Several of the new irons are from Antarctica. Of 24 independent irons from Antarctica, eight are ungrouped, a much higher fraction than that among all classified irons. The difference is probably related to the fact that the median mass of Antarctic irons is about two orders of magnitude smaller than that of non-Antarctic irons. Smaller meteoroids may tend to sample a larger number of asteroidal source regions, perhaps because small meteoroids tend to have higher ejection velocities or because they have random-walked a greater increment of orbital semimajor axis away from that of the parent body.
Karyological features of wild and cultivated forms of myrtle (Myrtus communis, Myrtaceae).
Serçe, S; Ekbiç, E; Suda, J; Gündüz, K; Kiyga, Y
2010-03-09
Myrtle is an evergreen shrub or small tree widespread throughout the Mediterranean region. In Turkey, both cultivated and wild forms, differing in plant and fruit size and fruit composition, can be found. These differences may have resulted from the domestication of the cultivated form over a long period of time. We investigated whether wild and cultivated forms of myrtle differ in karyological features (i.e., number of somatic chromosomes and relative genome size). We sampled two wild forms and six cultivated types of myrtle. All the samples had the same chromosome number (2n = 2x = 22). The results were confirmed by 4',6-diamidino-2-phenylindole (DAPI) flow cytometry. Only negligible variation (approximately 3%) in relative fluorescence intensity was observed among the different myrtle accessions, with wild genotypes having the smallest values. We concluded that despite considerable morphological differentiation, cultivated and wild myrtle genotypes in Turkey have similar karyological features.
A Dynamic Time Warping Approach to Real-Time Activity Recognition for Food Preparation
NASA Astrophysics Data System (ADS)
Pham, Cuong; Plötz, Thomas; Olivier, Patrick
We present a dynamic time warping based activity recognition system for the analysis of low-level food preparation activities. Accelerometers embedded into kitchen utensils provide continuous sensor data streams while people are using them for cooking. The recognition framework analyzes frames of contiguous sensor readings in real-time with low latency. It thereby adapts to the idiosyncrasies of utensil use by automatically maintaining a template database. We demonstrate the effectiveness of the classification approach by a number of real-world practical experiments on a publically available dataset. The adaptive system shows superior performance compared to a static recognizer. Furthermore, we demonstrate the generalization capabilities of the system by gradually reducing the amount of training samples. The system achieves excellent classification results even if only a small number of training samples is available, which is especially relevant for real-world scenarios.
Importance-sampling computation of statistical properties of coupled oscillators
NASA Astrophysics Data System (ADS)
Gupta, Shamik; Leitão, Jorge C.; Altmann, Eduardo G.
2017-07-01
We introduce and implement an importance-sampling Monte Carlo algorithm to study systems of globally coupled oscillators. Our computational method efficiently obtains estimates of the tails of the distribution of various measures of dynamical trajectories corresponding to states occurring with (exponentially) small probabilities. We demonstrate the general validity of our results by applying the method to two contrasting cases: the driven-dissipative Kuramoto model, a paradigm in the study of spontaneous synchronization; and the conservative Hamiltonian mean-field model, a prototypical system of long-range interactions. We present results for the distribution of the finite-time Lyapunov exponent and a time-averaged order parameter. Among other features, our results show most notably that the distributions exhibit a vanishing standard deviation but a skewness that is increasing in magnitude with the number of oscillators, implying that nontrivial asymmetries and states yielding rare or atypical values of the observables persist even for a large number of oscillators.
Neeser, Rudolph; Ackermann, Rebecca Rogers; Gain, James
2009-09-01
Various methodological approaches have been used for reconstructing fossil hominin remains in order to increase sample sizes and to better understand morphological variation. Among these, morphometric quantitative techniques for reconstruction are increasingly common. Here we compare the accuracy of three approaches--mean substitution, thin plate splines, and multiple linear regression--for estimating missing landmarks of damaged fossil specimens. Comparisons are made varying the number of missing landmarks, sample sizes, and the reference species of the population used to perform the estimation. The testing is performed on landmark data from individuals of Homo sapiens, Pan troglodytes and Gorilla gorilla, and nine hominin fossil specimens. Results suggest that when a small, same-species fossil reference sample is available to guide reconstructions, thin plate spline approaches perform best. However, if no such sample is available (or if the species of the damaged individual is uncertain), estimates of missing morphology based on a single individual (or even a small sample) of close taxonomic affinity are less accurate than those based on a large sample of individuals drawn from more distantly related extant populations using a technique (such as a regression method) able to leverage the information (e.g., variation/covariation patterning) contained in this large sample. Thin plate splines also show an unexpectedly large amount of error in estimating landmarks, especially over large areas. Recommendations are made for estimating missing landmarks under various scenarios. Copyright 2009 Wiley-Liss, Inc.
Integrating diffusion maps with umbrella sampling: Application to alanine dipeptide
NASA Astrophysics Data System (ADS)
Ferguson, Andrew L.; Panagiotopoulos, Athanassios Z.; Debenedetti, Pablo G.; Kevrekidis, Ioannis G.
2011-04-01
Nonlinear dimensionality reduction techniques can be applied to molecular simulation trajectories to systematically extract a small number of variables with which to parametrize the important dynamical motions of the system. For molecular systems exhibiting free energy barriers exceeding a few kBT, inadequate sampling of the barrier regions between stable or metastable basins can lead to a poor global characterization of the free energy landscape. We present an adaptation of a nonlinear dimensionality reduction technique known as the diffusion map that extends its applicability to biased umbrella sampling simulation trajectories in which restraining potentials are employed to drive the system into high free energy regions and improve sampling of phase space. We then propose a bootstrapped approach to iteratively discover good low-dimensional parametrizations by interleaving successive rounds of umbrella sampling and diffusion mapping, and we illustrate the technique through a study of alanine dipeptide in explicit solvent.
Matching-to-sample by an echolocating dolphin (Tursiops truncatus).
Roitblat, H L; Penner, R H; Nachtigall, P E
1990-01-01
An adult male dolphin was trained to perform a three-alternative delayed matching-to-sample task while wearing eyecups to occlude its vision. Sample and comparison stimuli consisted of a small and a large PVC plastic tube, a water-filled stainless steel sphere, and a solid aluminum cone. Stimuli were presented under water and the dolphin was allowed to identify the stimuli through echolocation. The echolocation clicks emitted by the dolphin to each sample and each comparison stimulus were recorded and analyzed. Over 48 sessions of testing, choice accuracy averaged 94.5% correct. This high level of accuracy was apparently achieved by varying the number of echolocation clicks emitted to various stimuli. Performance appeared to reflect a preexperimental stereotyped search pattern that dictated the order in which comparison items were examined and a complex sequential-sampling decision process. A model for the dolphin's decision-making processes is described.
Lambertini, Elisabetta; Spencer, Susan K.; Bertz, Phillip D.; Loge, Frank J.; Kieke, Burney A.; Borchardt, Mark A.
2008-01-01
Available filtration methods to concentrate waterborne viruses are either too costly for studies requiring large numbers of samples, limited to small sample volumes, or not very portable for routine field applications. Sodocalcic glass wool filtration is a cost-effective and easy-to-use method to retain viruses, but its efficiency and reliability are not adequately understood. This study evaluated glass wool filter performance to concentrate the four viruses on the U.S. Environmental Protection Agency contaminant candidate list, i.e., coxsackievirus, echovirus, norovirus, and adenovirus, as well as poliovirus. Total virus numbers recovered were measured by quantitative reverse transcription-PCR (qRT-PCR); infectious polioviruses were quantified by integrated cell culture (ICC)-qRT-PCR. Recovery efficiencies averaged 70% for poliovirus, 14% for coxsackievirus B5, 19% for echovirus 18, 21% for adenovirus 41, and 29% for norovirus. Virus strain and water matrix affected recovery, with significant interaction between the two variables. Optimal recovery was obtained at pH 6.5. No evidence was found that water volume, filtration rate, and number of viruses seeded influenced recovery. The method was successful in detecting indigenous viruses in municipal wells in Wisconsin. Long-term continuous filtration retained viruses sufficiently for their detection for up to 16 days after seeding for qRT-PCR and up to 30 days for ICC-qRT-PCR. Glass wool filtration is suitable for large-volume samples (1,000 liters) collected at high filtration rates (4 liters min−1), and its low cost makes it advantageous for studies requiring large numbers of samples. PMID:18359827
Size distribution and growth rate of crystal nuclei near critical undercooling in small volumes
NASA Astrophysics Data System (ADS)
Kožíšek, Z.; Demo, P.
2017-11-01
Kinetic equations are numerically solved within standard nucleation model to determine the size distribution of nuclei in small volumes near critical undercooling. Critical undercooling, when first nuclei are detected within the system, depends on the droplet volume. The size distribution of nuclei reaches the stationary value after some time delay and decreases with nucleus size. Only a certain maximum size of nuclei is reached in small volumes near critical undercooling. As a model system, we selected recently studied nucleation in Ni droplet [J. Bokeloh et al., Phys. Rev. Let. 107 (2011) 145701] due to available experimental and simulation data. However, using these data for sample masses from 23 μg up to 63 mg (corresponding to experiments) leads to the size distribution of nuclei, when no critical nuclei in Ni droplet are formed (the number of critical nuclei < 1). If one takes into account the size dependence of the interfacial energy, the size distribution of nuclei increases to reasonable values. In lower volumes (V ≤ 10-9 m3) nucleus size reaches some maximum extreme size, which quickly increases with undercooling. Supercritical clusters continue their growth only if the number of critical nuclei is sufficiently high.
Schillaci, Michael A; Schillaci, Mario E
2009-02-01
The use of small sample sizes in human and primate evolutionary research is commonplace. Estimating how well small samples represent the underlying population, however, is not commonplace. Because the accuracy of determinations of taxonomy, phylogeny, and evolutionary process are dependant upon how well the study sample represents the population of interest, characterizing the uncertainty, or potential error, associated with analyses of small sample sizes is essential. We present a method for estimating the probability that the sample mean is within a desired fraction of the standard deviation of the true mean using small (n<10) or very small (n < or = 5) sample sizes. This method can be used by researchers to determine post hoc the probability that their sample is a meaningful approximation of the population parameter. We tested the method using a large craniometric data set commonly used by researchers in the field. Given our results, we suggest that sample estimates of the population mean can be reasonable and meaningful even when based on small, and perhaps even very small, sample sizes.
Observational studies of patients in the emergency department: a comparison of 4 sampling methods.
Valley, Morgan A; Heard, Kennon J; Ginde, Adit A; Lezotte, Dennis C; Lowenstein, Steven R
2012-08-01
We evaluate the ability of 4 sampling methods to generate representative samples of the emergency department (ED) population. We analyzed the electronic records of 21,662 consecutive patient visits at an urban, academic ED. From this population, we simulated different models of study recruitment in the ED by using 2 sample sizes (n=200 and n=400) and 4 sampling methods: true random, random 4-hour time blocks by exact sample size, random 4-hour time blocks by a predetermined number of blocks, and convenience or "business hours." For each method and sample size, we obtained 1,000 samples from the population. Using χ(2) tests, we measured the number of statistically significant differences between the sample and the population for 8 variables (age, sex, race/ethnicity, language, triage acuity, arrival mode, disposition, and payer source). Then, for each variable, method, and sample size, we compared the proportion of the 1,000 samples that differed from the overall ED population to the expected proportion (5%). Only the true random samples represented the population with respect to sex, race/ethnicity, triage acuity, mode of arrival, language, and payer source in at least 95% of the samples. Patient samples obtained using random 4-hour time blocks and business hours sampling systematically differed from the overall ED patient population for several important demographic and clinical variables. However, the magnitude of these differences was not large. Common sampling strategies selected for ED-based studies may affect parameter estimates for several representative population variables. However, the potential for bias for these variables appears small. Copyright © 2012. Published by Mosby, Inc.
Collaborative Indoor Access Point Localization Using Autonomous Mobile Robot Swarm.
Awad, Fahed; Naserllah, Muhammad; Omar, Ammar; Abu-Hantash, Alaa; Al-Taj, Abrar
2018-01-31
Localization of access points has become an important research problem due to the wide range of applications it addresses such as dismantling critical security threats caused by rogue access points or optimizing wireless coverage of access points within a service area. Existing proposed solutions have mostly relied on theoretical hypotheses or computer simulation to demonstrate the efficiency of their methods. The techniques that rely on estimating the distance using samples of the received signal strength usually assume prior knowledge of the signal propagation characteristics of the indoor environment in hand and tend to take a relatively large number of uniformly distributed random samples. This paper presents an efficient and practical collaborative approach to detect the location of an access point in an indoor environment without any prior knowledge of the environment. The proposed approach comprises a swarm of wirelessly connected mobile robots that collaboratively and autonomously collect a relatively small number of non-uniformly distributed random samples of the access point's received signal strength. These samples are used to efficiently and accurately estimate the location of the access point. The experimental testing verified that the proposed approach can identify the location of the access point in an accurate and efficient manner.
Collaborative Indoor Access Point Localization Using Autonomous Mobile Robot Swarm
Awad, Fahed; Naserllah, Muhammad; Omar, Ammar; Abu-Hantash, Alaa; Al-Taj, Abrar
2018-01-01
Localization of access points has become an important research problem due to the wide range of applications it addresses such as dismantling critical security threats caused by rogue access points or optimizing wireless coverage of access points within a service area. Existing proposed solutions have mostly relied on theoretical hypotheses or computer simulation to demonstrate the efficiency of their methods. The techniques that rely on estimating the distance using samples of the received signal strength usually assume prior knowledge of the signal propagation characteristics of the indoor environment in hand and tend to take a relatively large number of uniformly distributed random samples. This paper presents an efficient and practical collaborative approach to detect the location of an access point in an indoor environment without any prior knowledge of the environment. The proposed approach comprises a swarm of wirelessly connected mobile robots that collaboratively and autonomously collect a relatively small number of non-uniformly distributed random samples of the access point’s received signal strength. These samples are used to efficiently and accurately estimate the location of the access point. The experimental testing verified that the proposed approach can identify the location of the access point in an accurate and efficient manner. PMID:29385042
Caron, Alexis; Lelong, Christine; Pascual, Marie-Hélène; Benning, Véronique
2015-03-01
Minimizing the number of animals in regulatory toxicity studies while achieving study objectives to support the development of future medicines contributes to good scientific and ethical practices. Recent advances in technology have enabled the development of miniaturized blood sampling methods (including microsampling and dried blood spots) applicable to toxicokinetic determinations of small-molecule drugs. Implementation of miniaturized blood sampling methods in the context of biotherapeutic drugs is desirable because a limitation to this type of medicine remains the total blood volume needed from a single animal to support toxicokinetic determinations of several analytes (parent drug, metabolites[s], antidrug antibodies, and so forth). We describe here the technical details, applicability, and relevance of new miniaturized blood sampling procedures in mice and nonhuman primates in the context of the toxicologic evaluation of biotherapeutic drugs consisting of antibody-drug conjugates developed for oncology indications. These examples illustrate how these techniques can benefit the reduction of animal usage in mouse toxicity studies by decreasing the number of animals dedicated to toxicokinetic determinations and the refinement of practices in nonhuman primate toxicity studies by decreasing the blood volume repeatedly drawn for toxicokinetic determinations.
Caron, Alexis; Lelong, Christine; Pascual, Marie-Hélène; Benning, Véronique
2015-01-01
Minimizing the number of animals in regulatory toxicity studies while achieving study objectives to support the development of future medicines contributes to good scientific and ethical practices. Recent advances in technology have enabled the development of miniaturized blood sampling methods (including microsampling and dried blood spots) applicable to toxicokinetic determinations of small-molecule drugs. Implementation of miniaturized blood sampling methods in the context of biotherapeutic drugs is desirable because a limitation to this type of medicine remains the total blood volume needed from a single animal to support toxicokinetic determinations of several analytes (parent drug, metabolites[s], antidrug antibodies, and so forth). We describe here the technical details, applicability, and relevance of new miniaturized blood sampling procedures in mice and nonhuman primates in the context of the toxicologic evaluation of biotherapeutic drugs consisting of antibody–drug conjugates developed for oncology indications. These examples illustrate how these techniques can benefit the reduction of animal usage in mouse toxicity studies by decreasing the number of animals dedicated to toxicokinetic determinations and the refinement of practices in nonhuman primate toxicity studies by decreasing the blood volume repeatedly drawn for toxicokinetic determinations. PMID:25836960
Occurrence of Cymbasoma longispinosum Bourne, 1890 in the Curuçá River estuary.
Leite, Natália R; Pereira, Luci C C; Abrunhosa, Fernando; Pires, Marcus A B; Costa, Rauquírio M da
2010-09-01
The present work was carried out to verify the occurrence and distribution of Cymbasoma longispinosum Bourne, 1890 in a tropical Amazon estuary from North Brazil. Samplings were performed bimonthly from July/2003 to July/2004 at two different transects (Muriá and Curuçá rivers) situated along the Curuçá estuary (Pará, North Brazil). Samples were collected during neap tides via gentle (1 to 1.5 knots) 200 μm-mesh net tows from a small boat. Additional subsurface water samples were collected for the determination of environmental parameters. Males and females of Cymbasoma longispinosum were only observed during September and November/2003. The highest number of organisms was found in September/2003 at the Muriá River transect. The presence of C. longispinosum in samples obtained during September and November/2003 could probably be related to the reproductive period of this species in the studied estuary, which is directly related to the dry period in the region. The highest salinity values and the highest number of individuals observed in September/2003 corroborate with the previous assumption, since no C. longispinosum was found during the months comprising the rainy period (January to June).
Reduction in training time of a deep learning model in detection of lesions in CT
NASA Astrophysics Data System (ADS)
Makkinejad, Nazanin; Tajbakhsh, Nima; Zarshenas, Amin; Khokhar, Ashfaq; Suzuki, Kenji
2018-02-01
Deep learning (DL) emerged as a powerful tool for object detection and classification in medical images. Building a well-performing DL model, however, requires a huge number of images for training, and it takes days to train a DL model even on a cutting edge high-performance computing platform. This study is aimed at developing a method for selecting a "small" number of representative samples from a large collection of training samples to train a DL model for the could be used to detect polyps in CT colonography (CTC), without compromising the classification performance. Our proposed method for representative sample selection (RSS) consists of a K-means clustering algorithm. For the performance evaluation, we applied the proposed method to select samples for the training of a massive training artificial neural network based DL model, to be used for the classification of polyps and non-polyps in CTC. Our results show that the proposed method reduce the training time by a factor of 15, while maintaining the classification performance equivalent to the model trained using the full training set. We compare the performance using area under the receiveroperating- characteristic curve (AUC).
Research on computer virus database management system
NASA Astrophysics Data System (ADS)
Qi, Guoquan
2011-12-01
The growing proliferation of computer viruses becomes the lethal threat and research focus of the security of network information. While new virus is emerging, the number of viruses is growing, virus classification increasing complex. Virus naming because of agencies' capture time differences can not be unified. Although each agency has its own virus database, the communication between each other lacks, or virus information is incomplete, or a small number of sample information. This paper introduces the current construction status of the virus database at home and abroad, analyzes how to standardize and complete description of virus characteristics, and then gives the information integrity, storage security and manageable computer virus database design scheme.
DOA Finding with Support Vector Regression Based Forward-Backward Linear Prediction.
Pan, Jingjing; Wang, Yide; Le Bastard, Cédric; Wang, Tianzhen
2017-05-27
Direction-of-arrival (DOA) estimation has drawn considerable attention in array signal processing, particularly with coherent signals and a limited number of snapshots. Forward-backward linear prediction (FBLP) is able to directly deal with coherent signals. Support vector regression (SVR) is robust with small samples. This paper proposes the combination of the advantages of FBLP and SVR in the estimation of DOAs of coherent incoming signals with low snapshots. The performance of the proposed method is validated with numerical simulations in coherent scenarios, in terms of different angle separations, numbers of snapshots, and signal-to-noise ratios (SNRs). Simulation results show the effectiveness of the proposed method.
Khan, Sardar; Cao, Qing; Lin, Ai-Jun; Zhu, Yong-Guan
2008-06-01
Polycyclic aromatic hydrocarbons (PAHs) are widespread environmental contaminants and contribute to the pollution of soil environment. Soil ingestion is of increasing concern for assessing health risk from PAH-contaminated soils because soil ingestion is one of the potentially important pathways of exposure to environmental pollutants, particularly relevant for children playing at contaminated sites due to their hand-to-mouth activities. In vitro gastro-intestinal tests imitate the human digestive tract, based on the physiology of humans, generally more simple, less time-consuming, and especially more reproducible than animal tests. This study was conducted to investigate the level of PAH contamination and oral bioaccessibility in surface soils, using physiologically based in vitro gastro-intestinal tests regarding both gastric and small intestinal conditions. Wastewater-irrigated soils were sampled from the metropolitan areas of Beijing and Tianjin, China, which were highly contaminated with PAHs. Reference soil samples were also collected for comparisons. At each site, four soils were sampled in the upper horizon at the depth of 0-20 cm randomly and were bulked together to form one composite sample. PAH concentrations and origin were investigated and a physiologically based in vitro test was conducted using all analytical grade reagents. Linear regression model was used to assess the relationship between total PAH concentrations in soils and soil organic carbon (SOC). A wide range of total PAH concentrations ranging from 1,304 to 3,369 mug kg(-1) in soils collected from different wastewater-irrigated sites in Tianjin, while ranging from 2,687 to 4,916 mug kg(-1) in soils collected from different wastewater-irrigated sites in Beijing, was detected. In general, total PAH concentrations in soils from Beijing sites were significantly higher than those from Tianjin sites, indicating a dominant contribution from both pyrogenic and petrogenic sources. Results indicated that the oral bioaccessibility of PAHs in small intestinal was significantly higher (from P < 0.05 to P < 0.001) than gastric condition. Similarly, the oral bioaccessibility of PAHs in contaminated sites was significantly higher (from P < or = 0.05 to P < 0.001) than in reference sites. Individual PAH ratios (three to six rings), a more accurate and reliable estimation about the emission sources, were used to distinguish the natural and anthropogenic PAH inputs in the soils. Results indicated that PAHs were both pyrogenic and petrogenic in nature. The identification of PAH sources and importance of in vitro test for PAH bioaccessibility were emphasized in this study. The oral bioaccessibility of individual PAHs in soils generally decreased with increasing ring numbers of PAHs in both the gastric and small intestinal conditions. However, the ratio of bioaccessibility of individual PAHs in gastric conditions to that in the small intestinal condition generally increased with increasing ring numbers, indicating the relatively pronounced effect of bile extract on improving the bioaccessibility of PAHs with relatively high ring numbers characterized by their high K ( ow ) values. Similarly, total PAH concentrations in soils were strongly correlated with SOC, indicating that SOC was the key factor determining the retention of PAHs in soils. Soils were contaminated with PAHs due to long-term wastewater irrigation. PAHs with two to six rings showed high concentrations with a significant increase over reference soils. Based on the molecular indices, it was suggested that PAHs in soils had both pyrogenic and petrogenic sources. It was also concluded that the oral bioaccessibility of total PAHs in the small intestinal condition was significantly higher than that in the gastric condition. Furthermore, the bioaccessibility of individual PAHs in soils generally decreased with the increasing ring numbers in both the gastric and small intestinal conditions. It is suggested that more care should be given while establishing reliable soil criteria for PAHs, especially concerning the health of children who may ingest a considerable amount of PAH-contaminated soil via outdoor hand-to-mouth activities.
Ng, Zhi Xiang; Chua, Kek Heng; Kuppusamy, Umah Rani
2014-04-01
This study aimed to investigate the changes in the proteome of bitter gourd prior to and after subjecting to boiling and microwaving. A comparative analysis of the proteome profiles of raw and thermally treated bitter gourds was performed using 2D-DIGE. The protein content and number of protein spots in raw sample was higher when compared to the cooked samples. Qualitative analysis revealed that 103 (boiled sample) and 110 (microwaved sample) protein spots were up regulated whereas 120 (boiled sample) and 107 (microwaved sample) protein spots were down regulated. Ten protein spots with the highest significant fold change in the cooked samples were involved in carbohydrate/energy metabolisms and stress responses. Small heat shock proteins, superoxide dismutase, quinone oxidoreductase, UDP-glucose pyrophosphorylase and phosphoglycerate kinase play a role in heat-stress-mediated protection of bitter gourd. This study suggests that appropriate heat treatment (cooking methods) can lead to induction of selected proteins in bitter gourd. Copyright © 2013 Elsevier Ltd. All rights reserved.
Recognition Using Hybrid Classifiers.
Osadchy, Margarita; Keren, Daniel; Raviv, Dolev
2016-04-01
A canonical problem in computer vision is category recognition (e.g., find all instances of human faces, cars etc., in an image). Typically, the input for training a binary classifier is a relatively small sample of positive examples, and a huge sample of negative examples, which can be very diverse, consisting of images from a large number of categories. The difficulty of the problem sharply increases with the dimension and size of the negative example set. We propose to alleviate this problem by applying a "hybrid" classifier, which replaces the negative samples by a prior, and then finds a hyperplane which separates the positive samples from this prior. The method is extended to kernel space and to an ensemble-based approach. The resulting binary classifiers achieve an identical or better classification rate than SVM, while requiring far smaller memory and lower computational complexity to train and apply.
NASA Astrophysics Data System (ADS)
Yang, Chun; Quarles, C. A.
2007-10-01
We have used positron Doppler Broadening Spectroscopy (DBS) to investigate the uniformity of rubber-carbon black composite samples. The amount of carbon black added to a rubber sample is characterized by phr, the number of grams of carbon black per hundred grams of rubber. Typical concentrations in rubber tires are 50 phr. It has been shown that the S parameter measured by DBS depends on the phr of the sample, so the variation in carbon black concentration can be easily measured to 0.5 phr. In doing the experiments we observed a dependence of the S parameter on small variation in the counting rate or deadtime. By carefully calibrating this deadtime correction we can significantly reduce the experimental run time and thus make faster determination of the uniformity of extended samples.
Heat transfer to small horizontal cylinders immersed in a fluidized bed
DOE Office of Scientific and Technical Information (OSTI.GOV)
Friedman, J.; Koundakjian, P.; Naylor, D.
2006-10-15
Heat transfer to horizontal cylinders immersed in fluidized beds has been extensively studied, but mainly in the context of heat transfer to boiler tubes in coal-fired beds. As a result, most correlations in the literature have been derived for cylinders of 25-50 mm diameter in vigorously fluidizing beds. In recent years, fluidized bed heat treating furnaces fired by natural gas have become increasingly popular, particularly in the steel wire manufacturing industry. These fluidized beds typically operate at relatively low fluidizing rates and with small diameter wires (1-6 mm). Nusselt number correlations developed based on boiler tube studies do not extrapolatemore » down to these small size ranges and low fluidizing rates. In order to obtain reliable Nusselt number data for these size ranges, an experimental investigation has been undertaken using two heat treating fluidized beds; one a pilot-scale industrial unit and the other a lab-scale (300 mm diameter) unit. Heat transfer measurements were obtained using resistively heated cylindrical samples ranging from 1.3 to 9.5 mm in diameter at fluidizing rates ranging from approximately 0.5 x G{sub mf} (packed bed condition) to over 10 x G{sub mf} using aluminum oxide sand particles ranging from d{sub p}=145-330 {mu}m (50-90 grit). It has been found that for all cylinder sizes tested, the Nusselt number reaches a maximum near 2 x G{sub mf}, then remains relatively steady ({+-}5-10%) to the maximum fluidizing rate tested, typically 8-12xG{sub mf}. A correlation for maximum Nusselt number is developed.« less
Vaeth, Michael; Skovlund, Eva
2004-06-15
For a given regression problem it is possible to identify a suitably defined equivalent two-sample problem such that the power or sample size obtained for the two-sample problem also applies to the regression problem. For a standard linear regression model the equivalent two-sample problem is easily identified, but for generalized linear models and for Cox regression models the situation is more complicated. An approximately equivalent two-sample problem may, however, also be identified here. In particular, we show that for logistic regression and Cox regression models the equivalent two-sample problem is obtained by selecting two equally sized samples for which the parameters differ by a value equal to the slope times twice the standard deviation of the independent variable and further requiring that the overall expected number of events is unchanged. In a simulation study we examine the validity of this approach to power calculations in logistic regression and Cox regression models. Several different covariate distributions are considered for selected values of the overall response probability and a range of alternatives. For the Cox regression model we consider both constant and non-constant hazard rates. The results show that in general the approach is remarkably accurate even in relatively small samples. Some discrepancies are, however, found in small samples with few events and a highly skewed covariate distribution. Comparison with results based on alternative methods for logistic regression models with a single continuous covariate indicates that the proposed method is at least as good as its competitors. The method is easy to implement and therefore provides a simple way to extend the range of problems that can be covered by the usual formulas for power and sample size determination. Copyright 2004 John Wiley & Sons, Ltd.
Surface degassing and modifications to vesicle size distributions in active basalt flows
Cashman, K.V.; Mangan, M.T.; Newman, S.
1994-01-01
The character of the vesicle population in lava flows includes several measurable parameters that may provide important constraints on lava flow dynamics and rheology. Interpretation of vesicle size distributions (VSDs), however, requires an understanding of vesiculation processes in feeder conduits, and of post-eruption modifications to VSDs during transport and emplacement. To this end we collected samples from active basalt flows at Kilauea Volcano: (1) near the effusive Kupaianaha vent; (2) through skylights in the approximately isothermal Wahaula and Kamoamoa tube systems transporting lava to the coast; (3) from surface breakouts at different locations along the lava tubes; and (4) from different locations in a single breakout from a lava tube 1 km from the 51 vent at Pu'u 'O'o. Near-vent samples are characterized by VSDs that show exponentially decreasing numbers of vesicles with increasing vesicle size. These size distributions suggest that nucleation and growth of bubbles were continuous during ascent in the conduit, with minor associated bubble coalescence resulting from differential bubble rise. The entire vesicle population can be attributed to shallow exsolution of H2O-dominated gases at rates consistent with those predicted by simple diffusion models. Measurements of H2O, CO2 and S in the matrix glass show that the melt equilibrated rapidly at atmospheric pressure. Down-tube samples maintain similar VSD forms but show a progressive decrease in both overall vesicularity and mean vesicle size. We attribute this change to open system, "passive" rise and escape of larger bubbles to the surface. Such gas loss from the tube system results in the output of 1.2 ?? 106 g/day SO2, an output representing an addition of approximately 1% to overall volatile budget calculations. A steady increase in bubble number density with downstream distance is best explained by continued bubble nucleation at rates of 7-8/cm3s. Rates are ???25% of those estimated from the vent samples, and thus represent volatile supersaturations considerably less than those of the conduit. We note also that the small total volume represented by this new bubble population does not: (1) measurably deplete the melt in volatiles; or (2) make up for the overall vesicularity decrease resulting from the loss of larger bubbles. Surface breakout samples have distinctive VSDs characterized by an extreme depletion in the small vesicle population. This results in samples with much lower number densities and larger mean vesicle sizes than corresponding tube samples. Similar VSD patterns have been observed in solidified lava flows and are interpreted to result from either static (wall rupture) or dynamic (bubble rise and capture) coalescence. Through comparison with vent and tube vesicle populations, we suggest that, in addition to coalescence, the observed vesicle populations in the breakout samples have experienced a rapid loss of small vesicles consistent with 'ripening' of the VSD resulting from interbubble diffusion of volatiles. Confinement of ripening features to surface flows suggests that the thin skin that forms on surface breakouts may play a role in the observed VSD modification. ?? 1994.
An ancient depleted mantle source for Archean crust in Rajasthan, India
NASA Technical Reports Server (NTRS)
Macdougall, J. D.; Gopalan, K.; Lugmair, G. W.; Roy, A. B.
1983-01-01
Data from an initial set of Banded Gneiss Complex (BGC) east of the city of Udaipur are given. In this region the BGC comprises typical grey gneiss with variably abundant granitic and mafic components. Efforts to date were concentrated on the mafic components which, based on chemical data, appear to be metavolcanic. All samples examined were recrystallized under amphibolite or upper amphibolite facies conditions. Pertinent chemical data for a small number of amphibolites analyzed so far are: SiO2: 49-53%; MgO: 5.7-7.3%; K2O: 0.24-0.50%; Ni: 106-140 ppm; Zr: 37-159 ppm. From Sm/Nd data, all amphibolites show small to moderate LREE enrichments.
Small rural hospitals: an example of market segmentation analysis.
Mainous, A G; Shelby, R L
1991-01-01
In recent years, market segmentation analysis has shown increased popularity among health care marketers, although marketers tend to focus upon hospitals as sellers. The present analysis suggests that there is merit to viewing hospitals as a market of consumers. Employing a random sample of 741 small rural hospitals, the present investigation sought to determine, through the use of segmentation analysis, the variables associated with hospital success (occupancy). The results of a discriminant analysis yielded a model which classifies hospitals with a high degree of predictive accuracy. Successful hospitals have more beds and employees, and are generally larger and have more resources. However, there was no significant relationship between organizational success and number of services offered by the institution.
Hudson, W. R.; Roberts, T. A.; Whelehan, O. P.
1987-01-01
The bacteriological status of beef carcasses was monitored at a commercial abattoir before and after two stages of modernization to the beef slaughterline which included changing from cradle dressing to dressing on an overhead rail, and the introduction of hot water spray cleaning of carcasses. Although small significant (P less than 0.05) differences in bacterial count occurred among carcass sites within modernization stages, significant visit within stage variation and stage X site interactions prevented any significant change in overall count being observed among stages and carcass sites. Principal components analysis revealed small changes in the distribution of bacterial numbers on the sites sampled. PMID:3556439
Sahara 99555 and D'Orbigny: Possible Pristine Parent Magma of Quenched Angrites
NASA Technical Reports Server (NTRS)
Mikouchi, T.; McKay, G. A.; Jones, J. H.
2004-01-01
Angrites constitute a small, but important group of basaltic achondrites showing unusual mineralogy and old crystallization ages. The currently known angrites are divided into two subgroups. Angra dos Reis (ADOR) and LEW86010 show slow cooling histories ("slowly-cooled" angrites) and differ from the later found angrites (LEW87051, Asuka 881371, Sahara 99555, D Orbigny, NWA1670, NWA1298). This second group has textures that suggest rapid cooling histories ("quenched" angrites). The petrogenesis of angrites has been controversial, partly due to the small number of available samples. In this abstract, we suggest a possible parent melt composition for the quenched angrites and its relationship to the partial melts of carbonaceous chondrites.
Radiomics-based Prognosis Analysis for Non-Small Cell Lung Cancer
NASA Astrophysics Data System (ADS)
Zhang, Yucheng; Oikonomou, Anastasia; Wong, Alexander; Haider, Masoom A.; Khalvati, Farzad
2017-04-01
Radiomics characterizes tumor phenotypes by extracting large numbers of quantitative features from radiological images. Radiomic features have been shown to provide prognostic value in predicting clinical outcomes in several studies. However, several challenges including feature redundancy, unbalanced data, and small sample sizes have led to relatively low predictive accuracy. In this study, we explore different strategies for overcoming these challenges and improving predictive performance of radiomics-based prognosis for non-small cell lung cancer (NSCLC). CT images of 112 patients (mean age 75 years) with NSCLC who underwent stereotactic body radiotherapy were used to predict recurrence, death, and recurrence-free survival using a comprehensive radiomics analysis. Different feature selection and predictive modeling techniques were used to determine the optimal configuration of prognosis analysis. To address feature redundancy, comprehensive analysis indicated that Random Forest models and Principal Component Analysis were optimum predictive modeling and feature selection methods, respectively, for achieving high prognosis performance. To address unbalanced data, Synthetic Minority Over-sampling technique was found to significantly increase predictive accuracy. A full analysis of variance showed that data endpoints, feature selection techniques, and classifiers were significant factors in affecting predictive accuracy, suggesting that these factors must be investigated when building radiomics-based predictive models for cancer prognosis.
Probing the Small-scale Structure in Strongly Lensed Systems via Transdimensional Inference
NASA Astrophysics Data System (ADS)
Daylan, Tansu; Cyr-Racine, Francis-Yan; Diaz Rivero, Ana; Dvorkin, Cora; Finkbeiner, Douglas P.
2018-02-01
Strong lensing is a sensitive probe of the small-scale density fluctuations in the Universe. We implement a pipeline to model strongly lensed systems using probabilistic cataloging, which is a transdimensional, hierarchical, and Bayesian framework to sample from a metamodel (union of models with different dimensionality) consistent with observed photon count maps. Probabilistic cataloging allows one to robustly characterize modeling covariances within and across lens models with different numbers of subhalos. Unlike traditional cataloging of subhalos, it does not require model subhalos to improve the goodness of fit above the detection threshold. Instead, it allows the exploitation of all information contained in the photon count maps—for instance, when constraining the subhalo mass function. We further show that, by not including these small subhalos in the lens model, fixed-dimensional inference methods can significantly mismodel the data. Using a simulated Hubble Space Telescope data set, we show that the subhalo mass function can be probed even when many subhalos in the sample catalogs are individually below the detection threshold and would be absent in a traditional catalog. The implemented software, Probabilistic Cataloger (PCAT) is made publicly available at https://github.com/tdaylan/pcat.
Evaluation of genetic variability in a small, insular population of spruce grouse
O'Connell, A.F.; Rhymer, Judith; Keppie, D.M.; Svenson, K.L.; Paigan, B.J.
2002-01-01
Using microsatellite markers we determined genetic variability for two populations of spruce grouse in eastern North America, one on a coastal Maine island where breeding habitat is limited and highly fragmented, the other in central New Brunswick (NB), where suitable breeding habitat is generally contiguous across the region. We examined six markers for both populations and all were polymorphic. Although the number of alleles per locus and the proportion of unique alleles were lower in the island population, and probably a result of small sample.size, heterozygosity and a breeding coefficient (Fis) indicated slightly more variability in the island population. Deviation from Hardy-Weinberg equilibrium also was more evident in loci for the mainland population. Several traits previously documented in the island population: relatively long natal dispersal distances, reproductive success, territoriality, adult survival, and longevity support the maintenance of hetrerzygosity, at least in the short-term. Sample collection from two small (500 ha), separate areas in NB, and the predicted importance of immigration density to supplement this population demonstrate the need for behavioral and ecological information when interpreting genetic variation. We discuss the relevance of these issues with respect to genetic variability and viability.
Potential role of Alternaria and Cladosporium species in canine lymphoplasmacytic rhinitis.
Mercier, E; Peters, I R; Billen, F; Battaille, G; Clercx, C; Day, M J; Peeters, D
2013-04-01
To evaluate the possible role of Alternaria and Cladosporium species in the pathogenesis of canine lymphoplasmacytic rhinitis by comparing the amount of specific fungal DNA in nasal mucosal biopsies between dogs without nasal neoplasia and those with lymphoplasmacytic rhinitis or nasal neoplasia. Quantitative real-time polymerase chain reaction (qPCR) assays detecting DNA from Alternaria and Cladosporium fungi were applied to nasal mucosal biopsies collected from dogs with lymphoplasmacytic rhinitis (n = 8), dogs with nasal neoplasia (n = 10) and control animals (n = 10). A copy number for each sample was calculated using a standard curve of known copy number and differences amongst groups were assessed using Kruskal-Wallis tests. No significant difference was found between the groups. Low levels of Alternaria DNA (10-100 copies/PCR) were detected in one sample; very low levels of DNA (<10 copies/qPCR) were detected in 6 samples, and 21 samples were negative. Low levels of Cladosporium DNA were detected in 2 samples; very low levels of DNA in 18; and 8 were negative. Results of this study reveal that Alternaria and Cladosporium species are part of the canine nasal flora, and that these fungi are probably not involved in the pathogenesis of lymphoplasmacytic rhinitis. © 2013 British Small Animal Veterinary Association.
Translations on Environmental Quality, Number 150
1977-10-17
again and again.... I recently did some calculations and it turned out that in 10 years the supplementary pay for those sick, the costs of treatment ...They required hospital treatment for a month and a half and 3 weeks respectively. At the Pal Heim Hospital "We have treated three small children...instruments showed an accumula- tion of lead in urine samples. After treatment , we released them in good condition. We received another three children during
Noise Levels and Data Analyses for Small Prop-Driven Aircraft
1983-08-01
assumption is that the acoustical emission characteristics of the test aircraft remain constant over the 3000 feet between sites. 7.1 Intensity metric...assumed that acoustical emission characteristics of the aircraft are nominally the same as the aircraft passes over the two measurement locations. As...associated with the emission of AIM. Table 12-2 lists the aircraft tested, number of samples, and the mean and standard deviation of the acoustical angle. The
Aitken, C G
1999-07-01
It is thought that, in a consignment of discrete units, a certain proportion of the units contain illegal material. A sample of the consignment is to be inspected. Various methods for the determination of the sample size are compared. The consignment will be considered as a random sample from some super-population of units, a certain proportion of which contain drugs. For large consignments, a probability distribution, known as the beta distribution, for the proportion of the consignment which contains illegal material is obtained. This distribution is based on prior beliefs about the proportion. Under certain specific conditions the beta distribution gives the same numerical results as an approach based on the binomial distribution. The binomial distribution provides a probability for the number of units in a sample which contain illegal material, conditional on knowing the proportion of the consignment which contains illegal material. This is in contrast to the beta distribution which provides probabilities for the proportion of a consignment which contains illegal material, conditional on knowing the number of units in the sample which contain illegal material. The interpretation when the beta distribution is used is much more intuitively satisfactory. It is also much more flexible in its ability to cater for prior beliefs which may vary given the different circumstances of different crimes. For small consignments, a distribution, known as the beta-binomial distribution, for the number of units in the consignment which are found to contain illegal material, is obtained, based on prior beliefs about the number of units in the consignment which are thought to contain illegal material. As with the beta and binomial distributions for large samples, it is shown that, in certain specific conditions, the beta-binomial and hypergeometric distributions give the same numerical results. However, the beta-binomial distribution, as with the beta distribution, has a more intuitively satisfactory interpretation and greater flexibility. The beta and the beta-binomial distributions provide methods for the determination of the minimum sample size to be taken from a consignment in order to satisfy a certain criterion. The criterion requires the specification of a proportion and a probability.
A Monte Carlo study of Weibull reliability analysis for space shuttle main engine components
NASA Technical Reports Server (NTRS)
Abernethy, K.
1986-01-01
The incorporation of a number of additional capabilities into an existing Weibull analysis computer program and the results of Monte Carlo computer simulation study to evaluate the usefulness of the Weibull methods using samples with a very small number of failures and extensive censoring are discussed. Since the censoring mechanism inherent in the Space Shuttle Main Engine (SSME) data is hard to analyze, it was decided to use a random censoring model, generating censoring times from a uniform probability distribution. Some of the statistical techniques and computer programs that are used in the SSME Weibull analysis are described. The methods documented in were supplemented by adding computer calculations of approximate (using iteractive methods) confidence intervals for several parameters of interest. These calculations are based on a likelihood ratio statistic which is asymptotically a chisquared statistic with one degree of freedom. The assumptions built into the computer simulations are described. The simulation program and the techniques used in it are described there also. Simulation results are tabulated for various combinations of Weibull shape parameters and the numbers of failures in the samples.
The reliability and stability of visual working memory capacity.
Xu, Z; Adam, K C S; Fang, X; Vogel, E K
2018-04-01
Because of the central role of working memory capacity in cognition, many studies have used short measures of working memory capacity to examine its relationship to other domains. Here, we measured the reliability and stability of visual working memory capacity, measured using a single-probe change detection task. In Experiment 1, the participants (N = 135) completed a large number of trials of a change detection task (540 in total, 180 each of set sizes 4, 6, and 8). With large numbers of both trials and participants, reliability estimates were high (α > .9). We then used an iterative down-sampling procedure to create a look-up table for expected reliability in experiments with small sample sizes. In Experiment 2, the participants (N = 79) completed 31 sessions of single-probe change detection. The first 30 sessions took place over 30 consecutive days, and the last session took place 30 days later. This unprecedented number of sessions allowed us to examine the effects of practice on stability and internal reliability. Even after much practice, individual differences were stable over time (average between-session r = .76).
Comparison of water-quality samples collected by siphon samplers and automatic samplers in Wisconsin
Graczyk, David J.; Robertson, Dale M.; Rose, William J.; Steur, Jeffrey J.
2000-01-01
In small streams, flow and water-quality concentrations often change quickly in response to meteorological events. Hydrologists, field technicians, or locally hired stream ob- servers involved in water-data collection are often unable to reach streams quickly enough to observe or measure these rapid changes. Therefore, in hydrologic studies designed to describe changes in water quality, a combination of manual and automated sampling methods have commonly been used manual methods when flow is relatively stable and automated methods when flow is rapidly changing. Auto- mated sampling, which makes use of equipment programmed to collect samples in response to changes in stage and flow of a stream, has been shown to be an effective method of sampling to describe the rapid changes in water quality (Graczyk and others, 1993). Because of the high cost of automated sampling, however, especially for studies examining a large number of sites, alternative methods have been considered for collecting samples during rapidly changing stream conditions. One such method employs the siphon sampler (fig. 1). also referred to as the "single-stage sampler." Siphon samplers are inexpensive to build (about $25- $50 per sampler), operate, and maintain, so they are cost effective to use at a large number of sites. Their ability to collect samples representing the average quality of water passing though the entire cross section of a stream, however, has not been fully demonstrated for many types of stream sites.
Hobbs, Brian P.; Sargent, Daniel J.; Carlin, Bradley P.
2014-01-01
Assessing between-study variability in the context of conventional random-effects meta-analysis is notoriously difficult when incorporating data from only a small number of historical studies. In order to borrow strength, historical and current data are often assumed to be fully homogeneous, but this can have drastic consequences for power and Type I error if the historical information is biased. In this paper, we propose empirical and fully Bayesian modifications of the commensurate prior model (Hobbs et al., 2011) extending Pocock (1976), and evaluate their frequentist and Bayesian properties for incorporating patient-level historical data using general and generalized linear mixed regression models. Our proposed commensurate prior models lead to preposterior admissible estimators that facilitate alternative bias-variance trade-offs than those offered by pre-existing methodologies for incorporating historical data from a small number of historical studies. We also provide a sample analysis of a colon cancer trial comparing time-to-disease progression using a Weibull regression model. PMID:24795786
Astrelin, A V; Sokolov, M V; Behnisch, T; Reymann, K G; Voronin, L L
1997-04-25
A statistical approach to analysis of amplitude fluctuations of postsynaptic responses is described. This includes (1) using a L1-metric in the space of distribution functions for minimisation with application of linear programming methods to decompose amplitude distributions into a convolution of Gaussian and discrete distributions; (2) deconvolution of the resulting discrete distribution with determination of the release probabilities and the quantal amplitude for cases with a small number (< 5) of discrete components. The methods were tested against simulated data over a range of sample sizes and signal-to-noise ratios which mimicked those observed in physiological experiments. In computer simulation experiments, comparisons were made with other methods of 'unconstrained' (generalized) and constrained reconstruction of discrete components from convolutions. The simulation results provided additional criteria for improving the solutions to overcome 'over-fitting phenomena' and to constrain the number of components with small probabilities. Application of the programme to recordings from hippocampal neurones demonstrated its usefulness for the analysis of amplitude distributions of postsynaptic responses.
NASA Astrophysics Data System (ADS)
Takayama, T.; Iwasaki, A.
2016-06-01
Above-ground biomass prediction of tropical rain forest using remote sensing data is of paramount importance to continuous large-area forest monitoring. Hyperspectral data can provide rich spectral information for the biomass prediction; however, the prediction accuracy is affected by a small-sample-size problem, which widely exists as overfitting in using high dimensional data where the number of training samples is smaller than the dimensionality of the samples due to limitation of require time, cost, and human resources for field surveys. A common approach to addressing this problem is reducing the dimensionality of dataset. Also, acquired hyperspectral data usually have low signal-to-noise ratio due to a narrow bandwidth and local or global shifts of peaks due to instrumental instability or small differences in considering practical measurement conditions. In this work, we propose a methodology based on fused lasso regression that select optimal bands for the biomass prediction model with encouraging sparsity and grouping, which solves the small-sample-size problem by the dimensionality reduction from the sparsity and the noise and peak shift problem by the grouping. The prediction model provided higher accuracy with root-mean-square error (RMSE) of 66.16 t/ha in the cross-validation than other methods; multiple linear analysis, partial least squares regression, and lasso regression. Furthermore, fusion of spectral and spatial information derived from texture index increased the prediction accuracy with RMSE of 62.62 t/ha. This analysis proves efficiency of fused lasso and image texture in biomass estimation of tropical forests.
Rogers, Paul; Stoner, Julie
2016-01-01
Regression models for correlated binary outcomes are commonly fit using a Generalized Estimating Equations (GEE) methodology. GEE uses the Liang and Zeger sandwich estimator to produce unbiased standard error estimators for regression coefficients in large sample settings even when the covariance structure is misspecified. The sandwich estimator performs optimally in balanced designs when the number of participants is large, and there are few repeated measurements. The sandwich estimator is not without drawbacks; its asymptotic properties do not hold in small sample settings. In these situations, the sandwich estimator is biased downwards, underestimating the variances. In this project, a modified form for the sandwich estimator is proposed to correct this deficiency. The performance of this new sandwich estimator is compared to the traditional Liang and Zeger estimator as well as alternative forms proposed by Morel, Pan and Mancl and DeRouen. The performance of each estimator was assessed with 95% coverage probabilities for the regression coefficient estimators using simulated data under various combinations of sample sizes and outcome prevalence values with an Independence (IND), Autoregressive (AR) and Compound Symmetry (CS) correlation structure. This research is motivated by investigations involving rare-event outcomes in aviation data. PMID:26998504
Demitri, Nevine; Zoubir, Abdelhak M
2017-01-01
Glucometers present an important self-monitoring tool for diabetes patients and, therefore, must exhibit high accuracy as well as good usability features. Based on an invasive photometric measurement principle that drastically reduces the volume of the blood sample needed from the patient, we present a framework that is capable of dealing with small blood samples, while maintaining the required accuracy. The framework consists of two major parts: 1) image segmentation; and 2) convergence detection. Step 1 is based on iterative mode-seeking methods to estimate the intensity value of the region of interest. We present several variations of these methods and give theoretical proofs of their convergence. Our approach is able to deal with changes in the number and position of clusters without any prior knowledge. Furthermore, we propose a method based on sparse approximation to decrease the computational load, while maintaining accuracy. Step 2 is achieved by employing temporal tracking and prediction, herewith decreasing the measurement time, and, thus, improving usability. Our framework is tested on several real datasets with different characteristics. We show that we are able to estimate the underlying glucose concentration from much smaller blood samples than is currently state of the art with sufficient accuracy according to the most recent ISO standards and reduce measurement time significantly compared to state-of-the-art methods.
Smith, Blair H; Hannaford, Philip C; Elliott, Alison M; Smith, W Cairns; Chambers, W Alastair
2005-04-01
Sampling for primary care research must strike a balance between efficiency and external validity. For most conditions, even a large population sample will yield a small number of cases, yet other sampling techniques risk problems with extrapolation of findings. To compare the efficiency and external validity of two sampling methods for both an intervention study and epidemiological research in primary care--a convenience sample and a general population sample--comparing the response and follow-up rates, the demographic and clinical characteristics of each sample, and calculating the 'number needed to sample' (NNS) for a hypothetical randomized controlled trial. In 1996, we selected two random samples of adults from 29 general practices in Grampian, for an epidemiological study of chronic pain. One sample of 4175 was identified by an electronic questionnaire that listed patients receiving regular analgesic prescriptions--the 'repeat prescription sample'. The other sample of 5036 was identified from all patients on practice lists--the 'general population sample'. Questionnaires, including demographic, pain and general health measures, were sent to all. A similar follow-up questionnaire was sent in 2000 to all those agreeing to participate in further research. We identified a potential group of subjects for a hypothetical trial in primary care based on a recently published trial (those aged 25-64, with severe chronic back pain, willing to participate in further research). The repeat prescription sample produced better response rates than the general sample overall (86% compared with 82%, P < 0.001), from both genders and from the oldest and youngest age groups. The NNS using convenience sampling was 10 for each member of the final potential trial sample, compared with 55 using general population sampling. There were important differences between the samples in age, marital and employment status, social class and educational level. However, among the potential trial sample, there were no demographic differences. Those from the repeat prescription sample had poorer indices than the general population sample in all pain and health measures. The repeat prescription sampling method was approximately five times more efficient than the general population method. However demographic and clinical differences in the repeat prescription sample might hamper extrapolation of findings to the general population, particularly in an epidemiological study, and demonstrate that simple comparison with age and gender of the target population is insufficient.
Vallée, Julie; Souris, Marc; Fournet, Florence; Bochaton, Audrey; Mobillion, Virginie; Peyronnie, Karine; Salem, Gérard
2007-01-01
Background Geographical objectives and probabilistic methods are difficult to reconcile in a unique health survey. Probabilistic methods focus on individuals to provide estimates of a variable's prevalence with a certain precision, while geographical approaches emphasise the selection of specific areas to study interactions between spatial characteristics and health outcomes. A sample selected from a small number of specific areas creates statistical challenges: the observations are not independent at the local level, and this results in poor statistical validity at the global level. Therefore, it is difficult to construct a sample that is appropriate for both geographical and probability methods. Methods We used a two-stage selection procedure with a first non-random stage of selection of clusters. Instead of randomly selecting clusters, we deliberately chose a group of clusters, which as a whole would contain all the variation in health measures in the population. As there was no health information available before the survey, we selected a priori determinants that can influence the spatial homogeneity of the health characteristics. This method yields a distribution of variables in the sample that closely resembles that in the overall population, something that cannot be guaranteed with randomly-selected clusters, especially if the number of selected clusters is small. In this way, we were able to survey specific areas while minimising design effects and maximising statistical precision. Application We applied this strategy in a health survey carried out in Vientiane, Lao People's Democratic Republic. We selected well-known health determinants with unequal spatial distribution within the city: nationality and literacy. We deliberately selected a combination of clusters whose distribution of nationality and literacy is similar to the distribution in the general population. Conclusion This paper describes the conceptual reasoning behind the construction of the survey sample and shows that it can be advantageous to choose clusters using reasoned hypotheses, based on both probability and geographical approaches, in contrast to a conventional, random cluster selection strategy. PMID:17543100
Vallée, Julie; Souris, Marc; Fournet, Florence; Bochaton, Audrey; Mobillion, Virginie; Peyronnie, Karine; Salem, Gérard
2007-06-01
Geographical objectives and probabilistic methods are difficult to reconcile in a unique health survey. Probabilistic methods focus on individuals to provide estimates of a variable's prevalence with a certain precision, while geographical approaches emphasise the selection of specific areas to study interactions between spatial characteristics and health outcomes. A sample selected from a small number of specific areas creates statistical challenges: the observations are not independent at the local level, and this results in poor statistical validity at the global level. Therefore, it is difficult to construct a sample that is appropriate for both geographical and probability methods. We used a two-stage selection procedure with a first non-random stage of selection of clusters. Instead of randomly selecting clusters, we deliberately chose a group of clusters, which as a whole would contain all the variation in health measures in the population. As there was no health information available before the survey, we selected a priori determinants that can influence the spatial homogeneity of the health characteristics. This method yields a distribution of variables in the sample that closely resembles that in the overall population, something that cannot be guaranteed with randomly-selected clusters, especially if the number of selected clusters is small. In this way, we were able to survey specific areas while minimising design effects and maximising statistical precision. We applied this strategy in a health survey carried out in Vientiane, Lao People's Democratic Republic. We selected well-known health determinants with unequal spatial distribution within the city: nationality and literacy. We deliberately selected a combination of clusters whose distribution of nationality and literacy is similar to the distribution in the general population. This paper describes the conceptual reasoning behind the construction of the survey sample and shows that it can be advantageous to choose clusters using reasoned hypotheses, based on both probability and geographical approaches, in contrast to a conventional, random cluster selection strategy.
Morgan, Sue
2009-07-01
One of the well-known needs of teenagers and young adults with cancer (TYAs) is their quest for knowledge about their illness and its treatment. As well as this they desire to be treated as partners and as an integral part of the team who care for them. This small exploratory study aims to address these issues in the development of a unique project where the young people were given unprecedented access to their own cancer samples and to the professional experts who analyse those samples. This small cohort of 31 TYAs, aged 13-24 years, were invited to look at their samples under a microscope and discuss it with the experts; of these 4 declined. A number of questionnaires, containing both open and closed questions, were given before, immediately post and one month after the viewing session. The results were analysed using SPSS13 and content analysis. The study was able to establish that viewing cancer samples had no measurable effect on the TYAs gaining of new knowledge, sense of control or choices that they made. However, it did show how the young people felt in light of actually 'seeing' their own cancer, and how this seemed to improve the overall perception of their disease. This is a new body of knowledge that requires further, more detailed, rigorous work.
Kitchen, A D; Newham, J A
2011-05-01
Whilst some of the assays used for serological screening of post-mortem blood samples from deceased tissue donors in some countries have been specifically validated by the manufacturer for this purpose, a significant number of those currently in use globally have not. Although specificity has previously been considered a problem in the screening of such samples, we believe that ensuring sensitivity is more important. The aim of this study was to validate a broader range of assays for the screening of post-mortem blood samples from deceased tissue donors. Six microplate immunoassays currently in use within National Health Service Blood and Transplant (NHSBT) for the screening of blood, tissue and stem cell donations were included. Representative samples from confirmed positive donors were titrated in screen negative post-mortem samples in parallel with normal pooled negative serum to determine if there was any inhibition with the post-mortem samples. There were no significant differences seen (P < 0.005) between the dilution curves obtained for the positive samples diluted in post-mortem samples and normal pooled sera. Although small numbers of samples were studied, it can be surmised that the post-mortem blood samples from deceased tissue donors, collected according to United Kingdom guidelines, are a suitable substrate for the assays evaluated. No diminution of reactivity was seen when dilution with sera from deceased donors was compared to dilution using pooled serum from live donors. In the absence of genuine low titre positive post-mortem samples, the use of samples spiked with various levels of target material provides a means of qualifying serological screening assays used by NHSBT for the screening of post-mortem blood samples from deceased tissue donors.
The SDSS-IV MaNGA Sample: Design, Optimization, and Usage Considerations
NASA Astrophysics Data System (ADS)
Wake, David A.; Bundy, Kevin; Diamond-Stanic, Aleksandar M.; Yan, Renbin; Blanton, Michael R.; Bershady, Matthew A.; Sánchez-Gallego, José R.; Drory, Niv; Jones, Amy; Kauffmann, Guinevere; Law, David R.; Li, Cheng; MacDonald, Nicholas; Masters, Karen; Thomas, Daniel; Tinker, Jeremy; Weijmans, Anne-Marie; Brownstein, Joel R.
2017-09-01
We describe the sample design for the SDSS-IV MaNGA survey and present the final properties of the main samples along with important considerations for using these samples for science. Our target selection criteria were developed while simultaneously optimizing the size distribution of the MaNGA integral field units (IFUs), the IFU allocation strategy, and the target density to produce a survey defined in terms of maximizing signal-to-noise ratio, spatial resolution, and sample size. Our selection strategy makes use of redshift limits that only depend on I-band absolute magnitude (M I ), or, for a small subset of our sample, M I and color (NUV - I). Such a strategy ensures that all galaxies span the same range in angular size irrespective of luminosity and are therefore covered evenly by the adopted range of IFU sizes. We define three samples: the Primary and Secondary samples are selected to have a flat number density with respect to M I and are targeted to have spectroscopic coverage to 1.5 and 2.5 effective radii (R e ), respectively. The Color-Enhanced supplement increases the number of galaxies in the low-density regions of color-magnitude space by extending the redshift limits of the Primary sample in the appropriate color bins. The samples cover the stellar mass range 5× {10}8≤slant {M}* ≤slant 3× {10}11 {M}⊙ {h}-2 and are sampled at median physical resolutions of 1.37 and 2.5 kpc for the Primary and Secondary samples, respectively. We provide weights that will statistically correct for our luminosity and color-dependent selection function and IFU allocation strategy, thus correcting the observed sample to a volume-limited sample.
NASA Astrophysics Data System (ADS)
Alexander, Louise; Snape, Joshua F.; Joy, Katherine H.; Downes, Hilary; Crawford, Ian A.
2016-09-01
Lunar mare basalts provide insights into the compositional diversity of the Moon's interior. Basalt fragments from the lunar regolith can potentially sample lava flows from regions of the Moon not previously visited, thus, increasing our understanding of lunar geological evolution. As part of a study of basaltic diversity at the Apollo 12 landing site, detailed petrological and geochemical data are provided here for 13 basaltic chips. In addition to bulk chemistry, we have analyzed the major, minor, and trace element chemistry of mineral phases which highlight differences between basalt groups. Where samples contain olivine, the equilibrium parent melt magnesium number (Mg#; atomic Mg/[Mg + Fe]) can be calculated to estimate parent melt composition. Ilmenite and plagioclase chemistry can also determine differences between basalt groups. We conclude that samples of approximately 1-2 mm in size can be categorized provided that appropriate mineral phases (olivine, plagioclase, and ilmenite) are present. Where samples are fine-grained (grain size <0.3 mm), a "paired samples t-test" can provide a statistical comparison between a particular sample and known lunar basalts. Of the fragments analyzed here, three are found to belong to each of the previously identified olivine and ilmenite basalt suites, four to the pigeonite basalt suite, one is an olivine cumulate, and two could not be categorized because of their coarse grain sizes and lack of appropriate mineral phases. Our approach introduces methods that can be used to investigate small sample sizes (i.e., fines) from future sample return missions to investigate lava flow diversity and petrological significance.
Smith, Nathan Grant; Hart, Trevor A; Kidwai, Ammaar; Vernon, Julia R G; Blais, Martin; Adam, Barry
2017-09-01
Project PRIDE (Promoting Resilience In Discriminatory Environments) is an 8-session small group intervention aimed at reducing negative mental and behavioral health outcomes resulting from minority stress. This study reports the results of a one-armed pilot test of Project PRIDE, which aimed to examine the feasibility and potential for efficacy of the intervention in a sample of 33 gay and bisexual men aged 18 to 25. The intervention appeared feasible to administer in two different sites and all participants who completed posttreatment (n = 22) or follow-up (n = 19) assessments reported high satisfaction with the intervention. Small to large effect sizes were observed for increases in self-esteem; small effect sizes were found for decreases in loneliness and decreases in minority stress variables; and small and medium effect sizes were found for reductions in alcohol use and number of sex partners, respectively. Overall, Project PRIDE appears to be a feasible intervention with promise of efficacy. Copyright © 2017. Published by Elsevier Ltd.
Researchers’ Intuitions About Power in Psychological Research
Bakker, Marjan; Hartgerink, Chris H. J.; Wicherts, Jelte M.; van der Maas, Han L. J.
2016-01-01
Many psychology studies are statistically underpowered. In part, this may be because many researchers rely on intuition, rules of thumb, and prior practice (along with practical considerations) to determine the number of subjects to test. In Study 1, we surveyed 291 published research psychologists and found large discrepancies between their reports of their preferred amount of power and the actual power of their studies (calculated from their reported typical cell size, typical effect size, and acceptable alpha). Furthermore, in Study 2, 89% of the 214 respondents overestimated the power of specific research designs with a small expected effect size, and 95% underestimated the sample size needed to obtain .80 power for detecting a small effect. Neither researchers’ experience nor their knowledge predicted the bias in their self-reported power intuitions. Because many respondents reported that they based their sample sizes on rules of thumb or common practice in the field, we recommend that researchers conduct and report formal power analyses for their studies. PMID:27354203
Researchers' Intuitions About Power in Psychological Research.
Bakker, Marjan; Hartgerink, Chris H J; Wicherts, Jelte M; van der Maas, Han L J
2016-08-01
Many psychology studies are statistically underpowered. In part, this may be because many researchers rely on intuition, rules of thumb, and prior practice (along with practical considerations) to determine the number of subjects to test. In Study 1, we surveyed 291 published research psychologists and found large discrepancies between their reports of their preferred amount of power and the actual power of their studies (calculated from their reported typical cell size, typical effect size, and acceptable alpha). Furthermore, in Study 2, 89% of the 214 respondents overestimated the power of specific research designs with a small expected effect size, and 95% underestimated the sample size needed to obtain .80 power for detecting a small effect. Neither researchers' experience nor their knowledge predicted the bias in their self-reported power intuitions. Because many respondents reported that they based their sample sizes on rules of thumb or common practice in the field, we recommend that researchers conduct and report formal power analyses for their studies. © The Author(s) 2016.
Integrative Analysis of Cancer Diagnosis Studies with Composite Penalization
Liu, Jin; Huang, Jian; Ma, Shuangge
2013-01-01
Summary In cancer diagnosis studies, high-throughput gene profiling has been extensively conducted, searching for genes whose expressions may serve as markers. Data generated from such studies have the “large d, small n” feature, with the number of genes profiled much larger than the sample size. Penalization has been extensively adopted for simultaneous estimation and marker selection. Because of small sample sizes, markers identified from the analysis of single datasets can be unsatisfactory. A cost-effective remedy is to conduct integrative analysis of multiple heterogeneous datasets. In this article, we investigate composite penalization methods for estimation and marker selection in integrative analysis. The proposed methods use the minimax concave penalty (MCP) as the outer penalty. Under the homogeneity model, the ridge penalty is adopted as the inner penalty. Under the heterogeneity model, the Lasso penalty and MCP are adopted as the inner penalty. Effective computational algorithms based on coordinate descent are developed. Numerical studies, including simulation and analysis of practical cancer datasets, show satisfactory performance of the proposed methods. PMID:24578589
Meng, Yilin; Roux, Benoît
2015-08-11
The weighted histogram analysis method (WHAM) is a standard protocol for postprocessing the information from biased umbrella sampling simulations to construct the potential of mean force with respect to a set of order parameters. By virtue of the WHAM equations, the unbiased density of state is determined by satisfying a self-consistent condition through an iterative procedure. While the method works very effectively when the number of order parameters is small, its computational cost grows rapidly in higher dimension. Here, we present a simple and efficient alternative strategy, which avoids solving the self-consistent WHAM equations iteratively. An efficient multivariate linear regression framework is utilized to link the biased probability densities of individual umbrella windows and yield an unbiased global free energy landscape in the space of order parameters. It is demonstrated with practical examples that free energy landscapes that are comparable in accuracy to WHAM can be generated at a small fraction of the cost.
2015-01-01
The weighted histogram analysis method (WHAM) is a standard protocol for postprocessing the information from biased umbrella sampling simulations to construct the potential of mean force with respect to a set of order parameters. By virtue of the WHAM equations, the unbiased density of state is determined by satisfying a self-consistent condition through an iterative procedure. While the method works very effectively when the number of order parameters is small, its computational cost grows rapidly in higher dimension. Here, we present a simple and efficient alternative strategy, which avoids solving the self-consistent WHAM equations iteratively. An efficient multivariate linear regression framework is utilized to link the biased probability densities of individual umbrella windows and yield an unbiased global free energy landscape in the space of order parameters. It is demonstrated with practical examples that free energy landscapes that are comparable in accuracy to WHAM can be generated at a small fraction of the cost. PMID:26574437
Carbon dioxide supersaturation in the surface waters of lakes
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cole, J.J.; Caraco, N.F.; Kling, G.W.
1994-09-09
Data on the partial pressure of carbon dioxide (CO{sub 2}) in the surface waters from a large number of lakes (1835) with a worldwide distribution show that only a small proportion of the 4665 samples analyzed (less than 10 percent) were within {+-}20 percent of equilibrium with the atmosphere and that most samples (87 percent) were supersaturated. The mean partial pressure of CO{sub 2} averaged 1036 microatmospheres, about three times the value in the overlying atmosphere, indicating that lakes are sources rather than sinks of atmospheric CO{sub 2}. On a global scale, the potential efflux of CO{sub 2} from lakesmore » (about 0.14 x 10{sup 15} grams of carbon per year) is about half as large as riverine transport of organic plus inorganic carbon to the ocean. Lakes are a small but potentially important conduit for carbon for terrestrial sources to the atmospheric sink. 18 refs., 2 figs., 1 tab.« less
Small angle x-ray scattering with edge-illumination
NASA Astrophysics Data System (ADS)
Modregger, Peter; Cremona, Tiziana P.; Benarafa, Charaf; Schittny, Johannes C.; Olivo, Alessandro; Endrizzi, Marco
2016-08-01
Sensitivity to sub-pixel sample features has been demonstrated as a valuable capability of phase contrast x-ray imaging. Here, we report on a method to obtain angular-resolved small angle x-ray scattering distributions with edge-illumination- based imaging utilizing incoherent illumination from an x-ray tube. Our approach provides both the three established image modalities (absorption, differential phase and scatter strength), plus a number of additional contrasts related to unresolved sample features. The complementarity of these contrasts is experimentally validated by using different materials in powder form. As a significant application example we show that the extended complementary contrasts could allow the diagnosis of pulmonary emphysema in a murine model. In support of this, we demonstrate that the properties of the retrieved scattering distributions are consistent with the expectation of increased feature sizes related to pulmonary emphysema. Combined with the simplicity of implementation of edge-illumination, these findings suggest a high potential for exploiting extended sub-pixel contrasts in the diagnosis of lung diseases and beyond.
The Efficacy of Yoga as a Form of Treatment for Depression
Bridges, Ledetra; Sharma, Manoj
2017-01-01
The purpose of this article was to systematically review yoga interventions aimed at improving depressive symptoms. A total of 23 interventions published between 2011 and May 2016 were evaluated in this review. Three study designs were used: randomized control trials, quasi-experimental, and pretest/posttest, with majority being randomized control trials. Most of the studies were in the United States. Various yoga schools were used, with the most common being Hatha yoga. The number of participants participating in the studies ranged from 14 to 136, implying that most studies had a small sample. The duration of the intervention period varied greatly, with the majority being 6 weeks or longer. Limitations of the interventions involved the small sample sizes used by the majority of the studies, most studies examining the short-term effect of yoga for depression, and the nonutilization of behavioral theories. Despite the limitations, it can be concluded that the yoga interventions were effective in reducing depression. PMID:28664775
Steele, L. P. [Commonwealth Scientific and Industrial Research Organization (CSIRO), Aspendale, Victoria, Australia; Krummel, P. B. [Commonwealth Scientific and Industrial Research Organization (CSIRO),; Langenfelds, R. L. [Commonwealth Scientific and Industrial Research Organization (CSIRO), Aspendale, Victoria, Australia
2008-01-01
Individual measurements have been obtained from flask air samples returned to the CSIRO GASLAB. Typical sample storage times range from days to weeks for some sites (e.g. Cape Grim, Aircraft over Tasmania and Bass Strait) to as much as one year for Macquarie Island and the Antarctic sites. Experiments carried out to test for changes in sample CO2 mixing ratio during storage have shown significant drifts in some flask types over test periods of several months to years (Cooper et al., 1999). Corrections derived from the test results are applied to network data according to flask type. These measurements indicate a rise in annual average atmospheric CO2 concentration from 357.72 parts per million by volume (ppmv) in 1992 to 383.05 ppmv in 2006, or an increase in annual average of about 1.81 ppmv/year. These flask data may be compared with other flask measurements from the Scripps Institution of Oceanography, available through 2004 in TRENDS; both indicate an annual average increase of 1.72 ppmv/year throuth 2004. Differences may be attributed to different sampling times or days, different numbers of samples, and different curve-fitting techniques used to obtain monthly and annual average numbers from flask data. Measurement error in flask data is believed to be small (Masarie et al., 2001).
Wang, Yuker; Carlton, Victoria EH; Karlin-Neumann, George; Sapolsky, Ronald; Zhang, Li; Moorhead, Martin; Wang, Zhigang C; Richardson, Andrea L; Warren, Robert; Walther, Axel; Bondy, Melissa; Sahin, Aysegul; Krahe, Ralf; Tuna, Musaffe; Thompson, Patricia A; Spellman, Paul T; Gray, Joe W; Mills, Gordon B; Faham, Malek
2009-01-01
Background A major challenge facing DNA copy number (CN) studies of tumors is that most banked samples with extensive clinical follow-up information are Formalin-Fixed Paraffin Embedded (FFPE). DNA from FFPE samples generally underperforms or suffers high failure rates compared to fresh frozen samples because of DNA degradation and cross-linking during FFPE fixation and processing. As FFPE protocols may vary widely between labs and samples may be stored for decades at room temperature, an ideal FFPE CN technology should work on diverse sample sets. Molecular Inversion Probe (MIP) technology has been applied successfully to obtain high quality CN and genotype data from cell line and frozen tumor DNA. Since the MIP probes require only a small (~40 bp) target binding site, we reasoned they may be well suited to assess degraded FFPE DNA. We assessed CN with a MIP panel of 50,000 markers in 93 FFPE tumor samples from 7 diverse collections. For 38 FFPE samples from three collections we were also able to asses CN in matched fresh frozen tumor tissue. Results Using an input of 37 ng genomic DNA, we generated high quality CN data with MIP technology in 88% of FFPE samples from seven diverse collections. When matched fresh frozen tissue was available, the performance of FFPE DNA was comparable to that of DNA obtained from matched frozen tumor (genotype concordance averaged 99.9%), with only a modest loss in performance in FFPE. Conclusion MIP technology can be used to generate high quality CN and genotype data in FFPE as well as fresh frozen samples. PMID:19228381
Lawson, Chris A
2018-09-01
Two experiments examined the extent to which category status influences children's attention to the composition of evidence samples provided by different informants. Children were told about two informants, each of whom presented different samples of evidence, and then were asked to judge which informant they would trust to help them learn something new. The composition of evidence samples was manipulated such that one sample included either a large number (n = 5) or a diverse range of exemplars relative to the other sample, which included either a small number (n = 2) or a homogeneous range of exemplars. Experiment 1 revealed that participants (N = 37; M age = 4.76 years) preferred to place their trust in the informant who presented the large or diverse sample when each informant was labeled "teacher" but exhibited no preference when each informant was labeled "child." Experiment 2 revealed developmental differences in responses when labels and sample composition were pitted against each other. Younger children (n = 32; M age = 3.42 years) consistently trusted the "teacher" regardless of the composition of the sample the informant was said to have provided, whereas older children (n = 30; M age = 5.54 years) consistently trusted the informant who provided the large or diverse sample regardless of whether it was provided by a "teacher" or a "child." These results have important implications for understanding the interplay between children's category knowledge and their evaluation of evidence. Copyright © 2018 Elsevier Inc. All rights reserved.
A cross-sectional study of small mammals for tick-borne pathogen infection in northern Mongolia.
Pulscher, Laura A; Moore, Thomas C; Caddell, Luke; Sukhbaatar, Lkhagvatseren; von Fricken, Michael E; Anderson, Benjamin D; Gonchigoo, Battsetseg; Gray, Gregory C
2018-01-01
Background : Tick-borne pathogens (TBPs) are frequently studied in developed nations but are often neglected in emerging countries. In Mongolia, TBP research is especially sparse, with few research reports focusing upon human and domestic animal disease and tick ecology. However, little information exists on TBPs in small mammals. Methods : In this 2016 cross-sectional pilot study, we sought to uniquely study wildlife for TBPs. We live-trapped small mammals, and tested their whole blood, serum and ear biopsy samples for molecular or serological evidence of Borrelia spp., Rickettsia spp., and Anaplasma spp. /Ehrlichia spp. Results : Of 64 small mammals collected, 56.0%, 39.0% and 0.0% of animals were positive by molecular assays for Borrelia spp., Rickettsia spp., and Anaplasma spp. /Erhlicia spp., respectively. 41.9% were seropositive for A. phagocytophilum and 24.2% of animals were seropositive for Rickettsia rickettsii . Conclusion : This pilot data demonstrates evidence of a number of TBPs among small mammal populations in northern Mongolia and suggests the need to further investigate what role these mammals play in human and domestic animal disease.
Tsui, Dana Wai Yi; Murtaza, Muhammed; Wong, Alvin Seng Cheong; Rueda, Oscar M; Smith, Christopher G; Chandrananda, Dineika; Soo, Ross A; Lim, Hong Liang; Goh, Boon Cher; Caldas, Carlos; Forshew, Tim; Gale, Davina; Liu, Wei; Morris, James; Marass, Francesco; Eisen, Tim; Chin, Tan Min; Rosenfeld, Nitzan
2018-06-01
Tumour heterogeneity leads to the development of multiple resistance mechanisms during targeted therapies. Identifying the dominant driver(s) is critical for treatment decision. We studied the relative dynamics of multiple oncogenic drivers in longitudinal plasma of 50 EGFR -mutant non-small-cell lung cancer patients receiving gefitinib and hydroxychloroquine. We performed digital PCR and targeted sequencing on samples from all patients and shallow whole-genome sequencing on samples from three patients who underwent histological transformation to small-cell lung cancer. In 43 patients with known EGFR mutations from tumour, we identified them accurately in plasma of 41 patients (95%, 41/43). We also found additional mutations, including EGFR T790M (31/50, 62%), TP53 (23/50, 46%), PIK3CA (7/50, 14%) and PTEN (4/50, 8%). Patients with both TP53 and EGFR mutations before treatment had worse overall survival than those with only EGFR Patients who progressed without T790M had worse PFS during TKI continuation and developed alternative alterations, including small-cell lung cancer-associated copy number changes and TP53 mutations, that tracked subsequent treatment responses. Longitudinal plasma analysis can help identify dominant resistance mechanisms, including non-druggable genetic information that may guide clinical management. © 2018 The Authors. Published under the terms of the CC BY 4.0 license.
Occurrence of 1153 organic micropollutants in the aquatic environment of Vietnam.
Chau, H T C; Kadokami, K; Duong, H T; Kong, L; Nguyen, T T; Nguyen, T Q; Ito, Y
2018-03-01
The rapid increase in the number and volume of chemical substances being used in modern society has been accompanied by a large number of potentially hazardous chemicals being found in environmental samples. In Vietnam, the monitoring of chemical substances is mainly limited to a small number of known pollutants in spite of rapid economic growth and urbanization, and there is an urgent need to examine a large number of chemicals to prevent impacts from expanding environmental pollution. However, it is difficult to analyze a large number of chemicals using existing methods, because they are time consuming and expensive. In the present study, we determined 1153 substances to grasp a pollution picture of microcontaminants in the aquatic environment. To achieve this objective, we have used two comprehensive analytical methods: (1) solid-phase extraction (SPE) and LC-TOF-MS analysis, and (2) SPE and GC-MS analysis. We collected 42 samples from northern (the Red River and Hanoi), central (Hue and Danang), and southern (Ho Chi Minh City and Saigon-Dongnai River) Vietnam. One hundred and sixty-five compounds were detected at least once. The compounds detected most frequently (>40 % samples) at μg/L concentrations were sterols (cholesterol, beta-sitosterol, stigmasterol, coprostanol), phthalates (bis(2-ethylhexyl) phthalate and di-n-butyl phthalate), and pharmaceutical and personal care products (caffeine, metformin). These contaminants were detected at almost the same detection frequency as in developed countries. The results reveal that surface waters in Vietnam, particularly in the center of large cities, are polluted by a large number of organic micropollutants, with households and business activities as the major sources. In addition, risk quotients (MEC/PNEC values) for nonylphenol, sulfamethoxazole, ampicillin, acetaminophen, erythromycin and clarithromycin were higher than 1, which indicates a possibility of adverse effects on aquatic ecosystems.
NASA Astrophysics Data System (ADS)
Lemasle, B.; Groenewegen, M. A. T.; Grebel, E. K.; Bono, G.; Fiorentino, G.; François, P.; Inno, L.; Kovtyukh, V. V.; Matsunaga, N.; Pedicelli, S.; Primas, F.; Pritchard, J.; Romaniello, M.; da Silva, R.
2017-12-01
Context. Cepheids are excellent tracers of young stellar populations. They play a crucial role in astrophysics as standard candles. The chemistry of classical Cepheids in the Milky Way is now quite well-known, however despite a much larger sample, the chemical composition of Magellanic Cepheids has been only scarcely investigated. Aims: For the first time, we study the chemical composition of several Cepheids located in the same populous cluster: NGC 1866, in the Large Magellanic Cloud (LMC). To also investigate the chemical composition of Cepheids at lower metallicity, we look at four targets located in the Small Magellanic Cloud (SMC). Our sample allows us to increase the number of Cepheids with known metallicities in the LMC/SMC by 20%/25% and the number of Cepheids with detailed chemical composition in the LMC/SMC by 46%/50%. Methods: We use canonical spectroscopic analysis to determine the chemical composition of Cepheids and provide abundances for a good number of α, iron-peak, and neutron-capture elements. Results: We find that six Cepheids in the LMC cluster NGC 1866 have a very homogeneous chemical composition, also consistent with red giant branch (RGB) stars in the cluster. Period-age relations that include no or average rotation indicate that all the Cepheids in NGC 1866 have a similar age and therefore belong to the same stellar population. Our results are in good agreement with theoretical models accounting for luminosity and radial velocity variations. Using distances based on period-luminosity relations in the near- or mid-infrared, we investigate for the first time the metallicity distribution of the young population in the SMC in the depth direction. Preliminary results show no metallicity gradient along the SMC main body, but our sample is small and does not contain Cepheids in the inner few degrees of the SMC. Based on observations collected at the European Organisation for Astronomical Research in the Southern Hemisphere under ESO programme 082.D-0792(B).
Infants' Auditory Enumeration: Evidence for Analog Magnitudes in the Small Number Range
ERIC Educational Resources Information Center
vanMarle, Kristy; Wynn, Karen
2009-01-01
Vigorous debate surrounds the issue of whether infants use different representational mechanisms to discriminate small and large numbers. We report evidence for ratio-dependent performance in infants' discrimination of small numbers of auditory events, suggesting that infants can use analog magnitudes to represent small values, at least in the…
Replica Exchange Improves Sampling in Low-Resolution Docking Stage of RosettaDock
Zhang, Zhe; Lange, Oliver F.
2013-01-01
Many protein-protein docking protocols are based on a shotgun approach, in which thousands of independent random-start trajectories minimize the rigid-body degrees of freedom. Another strategy is enumerative sampling as used in ZDOCK. Here, we introduce an alternative strategy, ReplicaDock, using a small number of long trajectories of temperature replica exchange. We compare replica exchange sampling as low-resolution stage of RosettaDock with RosettaDock's original shotgun sampling as well as with ZDOCK. A benchmark of 30 complexes starting from structures of the unbound binding partners shows improved performance for ReplicaDock and ZDOCK when compared to shotgun sampling at equal or less computational expense. ReplicaDock and ZDOCK consistently reach lower energies and generate significantly more near-native conformations than shotgun sampling. Accordingly, they both improve typical metrics of prediction quality of complex structures after refinement. Additionally, the refined ReplicaDock ensembles reach significantly lower interface energies and many previously hidden features of the docking energy landscape become visible when ReplicaDock is applied. PMID:24009670
On the Importance of Small Ice Crystals in Tropical Anvil Cirrus
NASA Technical Reports Server (NTRS)
Jensen, E. J.; Lawson, P.; Baker, B.; Pilson, B.; Mo, Q.; Heymsfield, A. J.; Bansemer, A.; Bui, T. P.; McGill, M.; Hlavka, D.;
2009-01-01
In situ measurements of ice crystal concentrations and sizes made with aircraft instrumentation over the past two decades have often indicated the presence of numerous relatively small (< 50 m diameter) crystals in cirrus clouds. Further, these measurements frequently indicate that small crystals account for a large fraction of the extinction in cirrus clouds. The fact that the instruments used to make these measurements, such as the Forward Scattering Spectrometer Probe (FSSP) and the Cloud Aerosol Spectrometer (CAS), ingest ice crystals into the sample volume through inlets has led to suspicion that the indications of numerous small ]crystals could be artifacts of large ]crystal shattering on the instrument inlets. We present new aircraft measurements in anvil cirrus sampled during the Tropical Composition, Cloud, and Climate Coupling (TC4) campaign with the 2 ] Dimensional Stereo (2D ]S) probe, which detects particles as small as 10 m. The 2D ]S has detector "arms" instead of an inlet tube. Since the 2D ]S probe surfaces are much further from the sample volume than is the case for the instruments with inlets, it is expected that 2D ]S will be less susceptible to shattering artifacts. In addition, particle inter ]arrival times are used to identify and remove shattering artifacts that occur even with the 2D ]S probe. The number of shattering artifacts identified by the 2D ]S interarrival time analysis ranges from a negligible contribution to an order of magnitude or more enhancement in apparent ice concentration over the natural ice concentration, depending on the abundance of large crystals and the natural small ]crystal concentration. The 2D ]S measurements in tropical anvil cirrus suggest that natural small ]crystal concentrations are typically one to two orders of magnitude lower than those inferred from CAS. The strong correlation between the CAS/2D ]S ratio of small ]crystal concentrations and large ]crystal concentration suggests that the discrepancy is likely caused by shattering of large crystals on the CAS inlet. We argue that past measurements with CAS in cirrus with large crystals present may contain errors due to crystal shattering, and past conclusions derived from these measurements may need to be revisited. Further, we present correlations between CAS spurious concentration and 2D ]S large ]crystal mass from spatially uniform anvil cirrus sampling periods as an approximate guide for estimating quantitative impact of large ]crystal shattering on CAS concentrations in previous datasets. We use radiative transfer calculations to demonstrate that in the maritime anvil cirrus sampled during TC4, small crystals indicated by 2D ]S contribute relatively little cloud extinction, radiative forcing, or radiative heating in the anvils, regardless of anvil age or vertical location in the clouds. While 2D ]S ice concentrations in fresh anvil cirrus may often exceed 1 cm.3, and are observed to exceed 10 cm.3 in turrets, they are typically 0.1 cm.3 and rarely exceed 1 cm.3 (<1.4% of the time) in aged anvil cirrus. We hypothesize that isolated occurrences of higher ice concentrations in aged anvil cirrus may be caused by ice nucleation driven by either small ]scale convection or gravity waves. It appears that the numerous small crystals detrained from convective updrafts do not persist in the anvil cirrus sampled during TC ]4.
Response Variability in Commercial MOSFET SEE Qualification
George, J. S.; Clymer, D. A.; Turflinger, T. L.; ...
2016-12-01
Single-event effects (SEE) evaluation of five different part types of next generation, commercial trench MOSFETs indicates large part-to-part variation in determining a safe operating area (SOA) for drain-source voltage (V DS) following a test campaign that exposed >50 samples per part type to heavy ions. These results suggest a determination of a SOA using small sample sizes may fail to capture the full extent of the part-to-part variability. An example method is discussed for establishing a Safe Operating Area using a one-sided statistical tolerance limit based on the number of test samples. Finally, burn-in is shown to be a criticalmore » factor in reducing part-to-part variation in part response. Implications for radiation qualification requirements are also explored.« less
Contrasting Size Distributions of Chondrules and Inclusions in Allende CV3
NASA Technical Reports Server (NTRS)
Fisher, Kent R.; Tait, Alastair W.; Simon, Jusin I.; Cuzzi, Jeff N.
2014-01-01
There are several leading theories on the processes that led to the formation of chondrites, e.g., sorting by mass, by X-winds, turbulent concentration, and by photophoresis. The juxtaposition of refractory inclusions (CAIs) and less refractory chondrules is central to these theories and there is much to be learned from their relative size distributions. There have been a number of studies into size distributions of particles in chondrites but only on relatively small scales primarily for chondrules, and rarely for both Calcium Aluminum-rich Inclusions (CAIs) and chondrules in the same sample. We have implemented macro-scale (25 cm diameter sample) and high-resolution microscale sampling of the Allende CV3 chondrite to create a complete data set of size frequencies for CAIs and chondrules.
Response Variability in Commercial MOSFET SEE Qualification
DOE Office of Scientific and Technical Information (OSTI.GOV)
George, J. S.; Clymer, D. A.; Turflinger, T. L.
Single-event effects (SEE) evaluation of five different part types of next generation, commercial trench MOSFETs indicates large part-to-part variation in determining a safe operating area (SOA) for drain-source voltage (V DS) following a test campaign that exposed >50 samples per part type to heavy ions. These results suggest a determination of a SOA using small sample sizes may fail to capture the full extent of the part-to-part variability. An example method is discussed for establishing a Safe Operating Area using a one-sided statistical tolerance limit based on the number of test samples. Finally, burn-in is shown to be a criticalmore » factor in reducing part-to-part variation in part response. Implications for radiation qualification requirements are also explored.« less
Rochelle-Newall, Emma J; Ribolzi, Olivier; Viguier, Marion; Thammahacksa, Chanthamousone; Silvera, Norbert; Latsachack, Keooudone; Dinh, Rinh Pham; Naporn, Piyapong; Sy, Hai Tran; Soulileuth, Bounsamay; Hmaimum, Nikom; Sisouvanh, Pem; Robain, Henri; Janeau, Jean-Louis; Valentin, Christian; Boithias, Laurie; Pierret, Alain
2016-09-08
Lack of access to clean water and adequate sanitation continues to be a major brake on development. Here we present the results of a 12-month investigation into the dynamics of Escherichia coli, a commonly used indicator of faecal contamination in water supplies, in three small, rural catchments in Laos, Thailand and Vietnam. We show that land use and hydrology are major controlling factors of E. coli concentrations in streamwater and that the relative importance of these two factors varies between the dry and wet seasons. In all three catchments, the highest concentrations were observed during the wet season when storm events and overland flow were highest. However, smaller peaks of E. coli concentration were also observed during the dry season. These latter correspond to periods of intense farming activities and small, episodic rain events. Furthermore, vegetation type, through land use and soil surface crusting, combined with mammalian presence play an important role in determining E. coli loads in the streams. Finally, sampling during stormflow revealed the importance of having appropriate sampling protocols if information on maximum contamination levels is required as grab sampling at a fixed time step may miss important peaks in E. coli numbers.
NASA Astrophysics Data System (ADS)
Rochelle-Newall, Emma J.; Ribolzi, Olivier; Viguier, Marion; Thammahacksa, Chanthamousone; Silvera, Norbert; Latsachack, Keooudone; Dinh, Rinh Pham; Naporn, Piyapong; Sy, Hai Tran; Soulileuth, Bounsamay; Hmaimum, Nikom; Sisouvanh, Pem; Robain, Henri; Janeau, Jean-Louis; Valentin, Christian; Boithias, Laurie; Pierret, Alain
2016-09-01
Lack of access to clean water and adequate sanitation continues to be a major brake on development. Here we present the results of a 12-month investigation into the dynamics of Escherichia coli, a commonly used indicator of faecal contamination in water supplies, in three small, rural catchments in Laos, Thailand and Vietnam. We show that land use and hydrology are major controlling factors of E. coli concentrations in streamwater and that the relative importance of these two factors varies between the dry and wet seasons. In all three catchments, the highest concentrations were observed during the wet season when storm events and overland flow were highest. However, smaller peaks of E. coli concentration were also observed during the dry season. These latter correspond to periods of intense farming activities and small, episodic rain events. Furthermore, vegetation type, through land use and soil surface crusting, combined with mammalian presence play an important role in determining E. coli loads in the streams. Finally, sampling during stormflow revealed the importance of having appropriate sampling protocols if information on maximum contamination levels is required as grab sampling at a fixed time step may miss important peaks in E. coli numbers.
Luo, Xian; Li, Liang
2017-11-07
In cellular metabolomics, it is desirable to carry out metabolomic profiling using a small number of cells in order to save time and cost. In some applications (e.g., working with circulating tumor cells in blood), only a limited number of cells are available for analysis. In this report, we describe a method based on high-performance chemical isotope labeling (CIL) nanoflow liquid chromatography mass spectrometry (nanoLC-MS) for high-coverage metabolomic analysis of small numbers of cells (i.e., ≤10000 cells). As an example, 12 C-/ 13 C-dansyl labeling of the metabolites in lysates of 100, 1000, and 10000 MCF-7 breast cancer cells was carried out using a new labeling protocol tailored to handle small amounts of metabolites. Chemical-vapor-assisted ionization in a captivespray interface was optimized for improving metabolite ionization and increasing robustness of nanoLC-MS. Compared to microflow LC-MS, the nanoflow system provided much improved metabolite detectability with a significantly reduced sample amount required for analysis. Experimental duplicate analyses of biological triplicates resulted in the detection of 1620 ± 148, 2091 ± 89 and 2402 ± 80 (n = 6) peak pairs or metabolites in the amine/phenol submetabolome from the 12 C-/ 13 C-dansyl labeled lysates of 100, 1000, and 10000 cells, respectively. About 63-69% of these peak pairs could be either identified using dansyl labeled standard library or mass-matched to chemical structures in human metabolome databases. We envisage the routine applications of this method for high-coverage quantitative cellular metabolomics using a starting material of 10000 cells. Even for analyzing 100 or 1000 cells, although the metabolomic coverage is reduced from the maximal coverage, this method can still detect thousands of metabolites, allowing the analysis of a large fraction of the metabolome and focused analysis of the detectable metabolites.
de Muinck, Eric J; Trosvik, Pål; Gilfillan, Gregor D; Hov, Johannes R; Sundaram, Arvind Y M
2017-07-06
Advances in sequencing technologies and bioinformatics have made the analysis of microbial communities almost routine. Nonetheless, the need remains to improve on the techniques used for gathering such data, including increasing throughput while lowering cost and benchmarking the techniques so that potential sources of bias can be better characterized. We present a triple-index amplicon sequencing strategy to sequence large numbers of samples at significantly lower c ost and in a shorter timeframe compared to existing methods. The design employs a two-stage PCR protocol, incorpo rating three barcodes to each sample, with the possibility to add a fourth-index. It also includes heterogeneity spacers to overcome low complexity issues faced when sequencing amplicons on Illumina platforms. The library preparation method was extensively benchmarked through analysis of a mock community in order to assess biases introduced by sample indexing, number of PCR cycles, and template concentration. We further evaluated the method through re-sequencing of a standardized environmental sample. Finally, we evaluated our protocol on a set of fecal samples from a small cohort of healthy adults, demonstrating good performance in a realistic experimental setting. Between-sample variation was mainly related to batch effects, such as DNA extraction, while sample indexing was also a significant source of bias. PCR cycle number strongly influenced chimera formation and affected relative abundance estimates of species with high GC content. Libraries were sequenced using the Illumina HiSeq and MiSeq platforms to demonstrate that this protocol is highly scalable to sequence thousands of samples at a very low cost. Here, we provide the most comprehensive study of performance and bias inherent to a 16S rRNA gene amplicon sequencing method to date. Triple-indexing greatly reduces the number of long custom DNA oligos required for library preparation, while the inclusion of variable length heterogeneity spacers minimizes the need for PhiX spike-in. This design results in a significant cost reduction of highly multiplexed amplicon sequencing. The biases we characterize highlight the need for highly standardized protocols. Reassuringly, we find that the biological signal is a far stronger structuring factor than the various sources of bias.
Noninvasive Prenatal Testing and Incidental Detection of Occult Maternal Malignancies.
Bianchi, Diana W; Chudova, Darya; Sehnert, Amy J; Bhatt, Sucheta; Murray, Kathryn; Prosen, Tracy L; Garber, Judy E; Wilkins-Haug, Louise; Vora, Neeta L; Warsof, Stephen; Goldberg, James; Ziainia, Tina; Halks-Miller, Meredith
2015-07-14
Understanding the relationship between aneuploidy detection on noninvasive prenatal testing (NIPT) and occult maternal malignancies may explain results that are discordant with the fetal karyotype and improve maternal clinical care. To evaluate massively parallel sequencing data for patterns of copy-number variations that might prospectively identify occult maternal malignancies. Case series identified from 125,426 samples submitted between February 15, 2012, and September 30, 2014, from asymptomatic pregnant women who underwent plasma cell-free DNA sequencing for clinical prenatal aneuploidy screening. Analyses were conducted in a clinical laboratory that performs DNA sequencing. Among the clinical samples, abnormal results were detected in 3757 (3%); these were reported to the ordering physician with recommendations for further evaluation. NIPT for fetal aneuploidy screening (chromosomes 13, 18, 21, X, and Y). Detailed genome-wide bioinformatics analysis was performed on available sequencing data from 8 of 10 women with known cancers. Genome-wide copy-number changes in the original NIPT samples and in subsequent serial samples from individual patients when available are reported. Copy-number changes detected in NIPT sequencing data in the known cancer cases were compared with the types of aneuploidies detected in the overall cohort. From a cohort of 125,426 NIPT results, 3757 (3%) were positive for 1 or more aneuploidies involving chromosomes 13, 18, 21, X, or Y. From this set of 3757 samples, 10 cases of maternal cancer were identified. Detailed clinical and sequencing data were obtained in 8. Maternal cancers most frequently occurred with the rare NIPT finding of more than 1 aneuploidy detected (7 known cancers among 39 cases of multiple aneuploidies by NIPT, 18% [95% CI, 7.5%-33.5%]). All 8 cases that underwent further bioinformatics analysis showed unique patterns of nonspecific copy-number gains and losses across multiple chromosomes. In 1 case, blood was sampled after completion of treatment for colorectal cancer and the abnormal pattern was no longer evident. In this preliminary study, a small number of cases of occult malignancy were subsequently diagnosed among pregnant women whose noninvasive prenatal testing results showed discordance with the fetal karyotype. The clinical importance of these findings will require further research.
Lang, Caroline; Hinchliffe, David; Brendle, Julia; Weirichl, Carmen; Hoy, Steffen
2016-01-01
The influence of different feedstuffs on parameters of digestion and ethology in growing rabbits (duration and number of feed intake, needed masticatory movements, pH value in stomach and small intestine, dry matter content of stomach chyme, level of destruction of cell structures in small intestine) was analysed. 384 five weeks old rabbits were fed with one of three feedstuffs: pellets (mean fibre length = 3 mm), fibre blocks (mean fibre length = 40 mm) and hay-oat-beat-ration (length of hay fibre ≥ 70 mm). The masticatory movements and duration for uptake 0.1 g of feedstuffs were measured in a special observation box. Rabbits in group cages were observed by 24 h video recording and duration and numbers of feed intakes were documented. After 8 weeks rabbits were slaughtered and pH values measured in stomach chyme (and dry matter content) and small intestine. Samples of them were taken and histologically examined (total length of villi and crypts, width of villi and degree of destruction determined by scores from 0 [= no destruction] to 3 [= severe destruction of villi]). Pellets lead to a faster feed intake with a lower number of masticatory movements. This equates a minor feeding time per feed intake and a higher amount of feedstuff in a shorter time. The dry matter content in stomach chyme increases and pH value was significantly higher there, but lower in duodenum. They also showed a significantly higher degree of destruction of villi, a shorter length and a larger width of villi than others.
Remote sensing with simulated unmanned aircraft imagery for precision agriculture applications
Hunt, E. Raymond; Daughtry, Craig S.T.; Mirsky, Steven B.; Hively, W. Dean
2014-01-01
An important application of unmanned aircraft systems (UAS) may be remote-sensing for precision agriculture, because of its ability to acquire images with very small pixel sizes from low altitude flights. The objective of this study was to compare information obtained from two different pixel sizes, one about a meter (the size of a small vegetation plot) and one about a millimeter. Cereal rye (Secale cereale) was planted at the Beltsville Agricultural Research Center for a winter cover crop with fall and spring fertilizer applications, which produced differences in biomass and leaf chlorophyll content. UAS imagery was simulated by placing a Fuji IS-Pro UVIR digital camera at 3-m height looking nadir. An external UV-IR cut filter was used to acquire true-color images; an external red cut filter was used to obtain color-infrared-like images with bands at near-infrared, green, and blue wavelengths. Plot-scale Green Normalized Difference Vegetation Index was correlated with dry aboveground biomass ( ${mbi {r}} = 0.58$ ), whereas the Triangular Greenness Index (TGI) was not correlated with chlorophyll content. We used the SamplePoint program to select 100 pixels systematically; we visually identified the cover type and acquired the digital numbers. The number of rye pixels in each image was better correlated with biomass ( ${mbi {r}} = 0.73$ ), and the average TGI from only leaf pixels was negatively correlated with chlorophyll content ( ${mbi {r}} = -0.72$ ). Thus, better information for crop requirements may be obtained using very small pixel sizes, but new algorithms based on computer vision are needed for analysis. It may not be necessary to geospatially register large numbers of photographs with very small pixel sizes. Instead, images could be analyzed as single plots along field transects.
Influence of prior knowledge of exercise duration on pacing strategies during game-based activities.
Gabbett, Tim J; Walker, Ben; Walker, Shane
2015-04-01
To investigate the influence of prior knowledge of exercise duration on the pacing strategies employed during game-based activities. Twelve semiprofessional team-sport athletes (mean ± SD age 22.8 ± 2.1 y) participated in this study. Players performed 3 small-sided games in random order. In one condition (Control), players were informed that they would play the small-sided game for 12 min and then completed the 12-min game. In a 2nd condition (Deception), players were told that they would play the small-sided game for 6 minutes, but after completing the 6-min game, they were asked to complete another 6 min. In a 3rd condition (Unknown), players were not told how long they would be required to play the small-sided game, but the activity was terminated after 12 min. Movement was recorded using a GPS unit sampling at 10 Hz. Post hoc inspection of video footage was undertaken to count the number of possessions and the number and quality of disposals. Higher initial intensities were observed in the Deception (130.6 ± 3.3 m/min) and Unknown (129.3 ± 2.4 m/min) conditions than the Control condition (123.3 ± 3.4 m/min). Greater amounts of high-speed running occurred during the initial phases of the Deception condition, and more low-speed activity occurred during the Unknown condition. A moderately greater number of total skill involvements occurred in the Unknown condition than the Control condition. These findings suggest that during game-based activities, players alter their pacing strategy based on the anticipated endpoint of the exercise bout.
Plausible Boosting of Millimeter-Galaxies in the COSMOS Field by Intervening Large-Scale Structure
NASA Astrophysics Data System (ADS)
Aretxaga, I.; Wilson, G. W.; Aguilar, E.; Alberts, S.; Scott, K. S.; Scoville, N.; Yun, M. S.; Austermann, J.; Downes, T. D.; Ezawa, H.; Hatsukade, B.; Hughes, D. H.; Kawabe, R.; Kohno, K.; Oshima, T.; Perera, T. A.; Tamura, Y.; Zeballos, M.
2011-10-01
The 0.72 sq. deg. contiguous 1.1mm survey in the central area of the COSMOS field, carried out to a 1σ≍1.26 mJy beam-1 depth with the AzTEC camera mounted on the 10m Atacama Submillimeter Telescope Experiment (ASTE), shows number counts with a significant excess of sources when compared to the number counts derived from the ˜0.5 sq. deg. area sampled at similar depths in the Scuba HAlf Degree Extragalactic Survey (SHADES, Austermann et al. 2010). They are, however, consistent with those derived from fields that were considered too small to characterize the overall blank-field population. We identify differences to be more significant in the S1.1mm ˜> 5 mJy regime, and demonstrate that these excesses in number counts are related to the areas where galaxies at redshifts ˜< 1.1 are more densely clustered. The positions of optical-IR galaxies in the redshift interval 0.6 ˜< z ˜< 0.75 are the ones that show the strongest correlation with the positions of the 1.1mm bright population (S1.mm ˜>5 mJy), a result which does not depend exclusively on the presence of rich clusters within the survey sampled area. The most likely explanation for the observed excess in number counts at 1.1mm is galaxy-galaxy and galaxy-group lensing at moderate amplification levels, that increases in amplitude as one samples larger and larger flux densities.
Multiple hot-deck imputation for network inference from RNA sequencing data.
Imbert, Alyssa; Valsesia, Armand; Le Gall, Caroline; Armenise, Claudia; Lefebvre, Gregory; Gourraud, Pierre-Antoine; Viguerie, Nathalie; Villa-Vialaneix, Nathalie
2018-05-15
Network inference provides a global view of the relations existing between gene expression in a given transcriptomic experiment (often only for a restricted list of chosen genes). However, it is still a challenging problem: even if the cost of sequencing techniques has decreased over the last years, the number of samples in a given experiment is still (very) small compared to the number of genes. We propose a method to increase the reliability of the inference when RNA-seq expression data have been measured together with an auxiliary dataset that can provide external information on gene expression similarity between samples. Our statistical approach, hd-MI, is based on imputation for samples without available RNA-seq data that are considered as missing data but are observed on the secondary dataset. hd-MI can improve the reliability of the inference for missing rates up to 30% and provides more stable networks with a smaller number of false positive edges. On a biological point of view, hd-MI was also found relevant to infer networks from RNA-seq data acquired in adipose tissue during a nutritional intervention in obese individuals. In these networks, novel links between genes were highlighted, as well as an improved comparability between the two steps of the nutritional intervention. Software and sample data are available as an R package, RNAseqNet, that can be downloaded from the Comprehensive R Archive Network (CRAN). alyssa.imbert@inra.fr or nathalie.villa-vialaneix@inra.fr. Supplementary data are available at Bioinformatics online.
Immediate Small Number Perception : Evidence from a New Numerical Carry-Over Procedure
ERIC Educational Resources Information Center
Demeyere, Nele; Humphreys, Glyn W.
2012-01-01
Evidence is presented for the immediate apprehension of exact small quantities. Participants performed a quantification task (are the number of items greater or smaller than?), and carry-over effects were examined between numbers requiring the same response. Carry-over effects between small numbers were strongly affected by repeats of pattern and…
NASA Technical Reports Server (NTRS)
Vilhu, Osmi; Linsky, Jeffrey L.
1990-01-01
Mean coronal temperatures of some active G-K stars were derived from Rev1-processed Einstein-observatory's IPC-spectra. The combined X-ray and transition region emission line data are in rough agreement with static coronal loop models. Although the sample is too small to derive any statistically significant conclusions, it suggests that the mean coronal temperature depends linearly on the inverse Rossby-number, with saturation at short rotation periods.
2013-05-31
21 Figure 15. Example of a Possible Foreign Object Observed in a Small Number of Slides. This Object May Be a Hair, Thread, or Plant Material that...h)anthracene Fluoranthene Fluorene Indeno(1,2,3-cd)pyrene Naphthalene Phenanthrene Pyrene 16 Distribution A. Approved for public release...material during sampling. These were subject to particle analysis as described above in order to estimate the coverage ratio and particle density of
Slimani, Sami; Robyns, Audrey; Jarraud, Sophie; Molmeret, Maëlle; Dusserre, Eric; Mazure, Céline; Facon, Jean Pierre; Lina, Gérard; Etienne, Jerome; Ginevra, Christophe
2012-02-01
A PMA (propidium monoazide) pretreatment protocol, in which PMA is applied directly to membrane filters, was developed for the PCR-based quantification (PMA-qPCR) of viable Legionella pneumophila. Using this method, the amplification of DNA from membrane-damaged L. pneumophila was strongly inhibited for samples containing a small number of dead bacteria. Copyright © 2011 Elsevier B.V. All rights reserved.
Epistasis between neurochemical gene polymorphisms and risk for ADHD
Segurado, Ricardo; Bellgrove, Mark A; Manconi, Francesca; Gill, Michael; Hawi, Ziarah
2011-01-01
A number of genes with function related to synaptic neurochemistry have been genetically associated with attention deficit/hyperactivity disorder. However, susceptibility to the development of common psychiatric disorders by single variants acting alone, can so far only explain a small proportion of the heritability of the phenotype. It has been postulated that the unexplained ‘dark heritability' may at least in part be due to epistatic effects, which may account for the small observed marginal associations, and the difficulties with replication of positive findings. We undertook a comprehensive exploration of pair-wise interactions between genetic variants in 24 candidate genic regions involved in monoaminergic catabolism, anabolism, release, re-uptake and signal transmission in a sample of 177 parent-affected child trios using a case-only design and a case–pseudocontrol design using conditional logistic regression. Marker-pairs thresholded on interaction odds ratio (OR) and P-value are presented. We detected a number of interaction ORs >4.0, including an interesting correlation between markers in the ADRA1B and DBH genes in affected individuals, and several further interesting but smaller effects. These effects are no larger than you would expect by chance under the assumption of independence of all pair-wise relations; however, independence is unlikely. Furthermore, the size of these effects is of interest and attempts to replicate these results in other samples are anticipated. PMID:21368917
Sampling requirements for forage quality characterization of rectangular hay bales
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sheaffer, C.C.; Martin, N.P.; Jewett, J.G.
2000-02-01
Commercial lots of alfalfa (Medicago sativa L.) hay are often bought and sold on the basis of forage quality. Proper sampling is essential to obtain accurate forage quality results for pricing of alfalfa hay, but information about sampling is limited to small, 20- to 40-kg rectangular bales. Their objectives were to determine the within-bale variation in 400-kg rectangular bales and to determine the number and distribution of core samples required to represent the crude protein (CP), acid detergent fiber (ADF), neutral detergent fiber (NDF), and dry matter (DM) concentration in commercial lots of alfalfa hay. Four bales were selected frommore » each of three hay lots and core sampled nine times per side for a total of 54 cores per bale. There was no consistent pattern of forage quality variation within bales. Averaged across lots, any portion of a bale was highly correlated with bale grand means for CP, ADF, NDF, and DM. Three lots of hay were probed six times per bale, one core per bale side from 55, 14, and 14 bales per lot. For determination of CP, ADF, NDF, and DM concentration, total core numbers required to achieve an acceptable standard error (SE) were minimized by sampling once per bale. Bootstrap analysis of data from the most variable hay lot suggested that forage quality of any lot of 400-kg alfalfa hay bales should be adequately represented by 12 bales sampled once per bale.« less
Chaffee, M.A.
1983-01-01
A technique called SCORESUM was developed to display a maximum of multi-element geochemical information on a minimum number of maps for mineral assessment purposes. The technique can be done manually for a small analytical data set or can be done with a computer for a large data set. SCORESUM can be used with highly censored data and can also weight samples so as to minimize the chemical differences of diverse lithologies in different parts of a given study area. The full range of reported analyses for each element of interest in a data set is divided into four categories. Anomaly scores - values of O (background), 1 (weakly anomalous), 2 (moderately anomalous), and 3 (strongly anomalous) - are substituted for all of the analyses falling into each of the four categories. A group of elements based on known or suspected association in altered or mineralized areas is selected for study and the anomaly scores for these elements are summed for each sample site and then plotted on a map. Some of the results of geochemical studies conducted for mineral assessments in two areas are briefly described. The first area, the Mokelumne Wilderness and vicinity, is a relatively small and geologically simple one. The second, the Walker Lake 1?? ?? 2?? quadrangle, is a large area that has extremely complex geology and that contains a number of different mineral deposit environments. These two studies provide examples of how the SCORESUM technique has been used (1) to enhance relatively small but anomalous areas and (2) to delineate and rank areas containing geochemical signatures for specific suites of elements related to certain types of alteration or mineralization. ?? 1983.
Multi-wavelength observations of barred, flocculent galaxies
NASA Astrophysics Data System (ADS)
Ratay, Douglas Lee
Although it is generally accepted that large galaxies form through the assemblage of smaller objects, an explanation for the morphology of galaxies is not available. Any complete theory of galaxy morphology must include production and dissolution mechanisms for galactic bars, rings, nuclear bars, spiral arms, and companions. This theory does not exist because of the lack of detailed data from many types of galaxies in different environments. We have defined a new sample of galaxies which are simultaneously flocculent, barred, and isolated. We have performed optical, near-infrared, and radio (HI) observations of the galaxies in this sample. We measured properties of our galaxies including bar length, bar axis ratio, HI diameter, HI mass, and dynamical mass. We found that our sample group is heterogeneous, and compares well to a standard samples of galaxies. We found two of our galaxies to possess companions, and two others to show evidence of current interactions. This is consistent with other observations indicating that local isolated galaxies do not possess a large number of small companions. We cannot rule out the possibility of very small companions. We find that as a group our sample is slightly less luminous than normal galaxies and may be more likely to be involved in interactions. We conclude that the bar and spiral arm features in our sample are due to processes internal to the galaxies, likely involving the interaction between the galactic disk and halo. We defined a control sample of barred, grand design galaxies to further determine the acceptability of barred, flocculent galaxies as a physically meaningful subset of galaxies.
Identifying Etiological Agents Causing Diarrhea in Low Income Ecuadorian Communities
Vasco, Gabriela; Trueba, Gabriel; Atherton, Richard; Calvopiña, Manuel; Cevallos, William; Andrade, Thamara; Eguiguren, Martha; Eisenberg, Joseph N. S.
2014-01-01
Continued success in decreasing diarrheal disease burden requires targeted interventions. To develop such interventions, it is crucial to understand which pathogens cause diarrhea. Using a case-control design we tested stool samples, collected in both rural and urban Ecuador, for 15 pathogenic microorganisms. Pathogens were present in 51% of case and 27% of control samples from the urban community, and 62% of case and 18% of control samples collected from the rural community. Rotavirus and Shigellae were associated with diarrhea in the urban community; co-infections were more pathogenic than single infection; Campylobacter and Entamoeba histolytica were found in large numbers in cases and controls; and non-typhi Salmonella and enteropathogenic Escherichia coli were not found in any samples. Consistent with the Global Enteric Multicenter Study, focused in south Asia and sub-Saharan Africa, we found that in Ecuador a small group of pathogens accounted for a significant amount of the diarrheal disease burden. PMID:25048373
Panadero, Sonia; Vázquez, José Juan; Martín, Rosa María
2016-06-14
The work analyzes different aspects related to alcohol consumption among homeless people and people at risk of social exclusion. The data was gathered from a representative sample of homeless people in Madrid (n = 188) and a sample of people at risk of social exclusion (n = 164) matched in sex, age, and origin (Spaniards vs. foreigners). The results showed that homeless people present a greater consumption of alcohol and have experienced more problems derived from its consumption than people at risk of social exclusion. Most of the homeless people who had alcohol-related problems had had them prior to their homelessness, and they stated they had poorer health and had experienced a greater number of homelessness episodes. Despite the relevance of problems related to alcohol among our sample, only a small percentage of the sample had participated in treatment programs for alcohol consumption.
Virological Sampling of Inaccessible Wildlife with Drones.
Geoghegan, Jemma L; Pirotta, Vanessa; Harvey, Erin; Smith, Alastair; Buchmann, Jan P; Ostrowski, Martin; Eden, John-Sebastian; Harcourt, Robert; Holmes, Edward C
2018-06-02
There is growing interest in characterizing the viromes of diverse mammalian species, particularly in the context of disease emergence. However, little is known about virome diversity in aquatic mammals, in part due to difficulties in sampling. We characterized the virome of the exhaled breath (or blow) of the Eastern Australian humpback whale ( Megaptera novaeangliae ). To achieve an unbiased survey of virome diversity, a meta-transcriptomic analysis was performed on 19 pooled whale blow samples collected via a purpose-built Unmanned Aerial Vehicle (UAV, or drone) approximately 3 km off the coast of Sydney, Australia during the 2017 winter annual northward migration from Antarctica to northern Australia. To our knowledge, this is the first time that UAVs have been used to sample viruses. Despite the relatively small number of animals surveyed in this initial study, we identified six novel virus species from five viral families. This work demonstrates the potential of UAVs in studies of virus disease, diversity, and evolution.
Grass height and transmission ecology of Echinococcus multilocularis in Tibetan communities, China.
Wang, Qian; Raoul, Francis; Budke, Christine; Craig, Philip S; Xiao, Yong-fu; Vuitton, Dominique A; Campos-Ponce, Maiza; Qiu, Dong-chuan; Pleydell, David; Giraudoux, Patrick
2010-01-05
Alveolar echinococcosis is a major zoonosis of public health significance in western China. Overgrazing was recently assumed as a potential risk factor for transmission of alveolar echinococcosis. The research was designed to further test the overgrazing hypothesis by investigating how overgrazing influenced the burrow density of intermediate host small mammals and how the burrow density of small mammals was associated with dog Echinococcus multilocularis infection. The study sites were chosen by previous studies which found areas where the alveolar echinococcosis was prevalent. The data, including grass height, burrow density of intermediate host small mammals, dog and fox fecal samples as well as Global Positioning System (GPS) position, were collected from field investigations in Shiqu County, Sichuan Province, China. The fecal samples were analyzed using copro-PCR. The worms, teeth, bones and hairs in the fecal samples were visually examined. Single factor and multifactor analyses tools including chi square and generalized linear models were applied to these data. By using grass height as a proxy of grazing pressure in the homogenous pasture, this study found that taller grass in the pasture led to lower small mammals' burrow density (chi(2) = 4.670, P = 0.031, coefficient = -1.570). The Echinococcus multilocularis worm burden in dogs was statistically significantly related to the maximum density of the intermediate host Ochotona spp. (chi(2) = 5.250, P = 0.022, coefficient = 0.028). The prevalence in owned dogs was positively correlated to the number of stray dogs seen within a 200 meter radius (Wald chi(2) = 8.375, P = 0.004, odds ratio = 1.198). Our findings support the hypothesis that overgrazing promotes transmission of alveolar echinococcosis and confirm the role of stray dogs in the transmission of alveolar echinococcosis.
Coxiella burnetii Infections in Small Ruminants and Humans in Switzerland.
Magouras, I; Hunninghaus, J; Scherrer, S; Wittenbrink, M M; Hamburger, A; Stärk, K D C; Schüpbach-Regula, G
2017-02-01
The recent Q fever epidemic in the Netherlands raised concerns about the potential risk of outbreaks in other European countries. In Switzerland, the prevalence of Q fever in animals and humans has not been studied in recent years. In this study, we describe the current situation with respect to Coxiella (C.) burnetii infections in small ruminants and humans in Switzerland, as a basis for future epidemiological investigations and public health risk assessments. Specific objectives of this cross-sectional study were to (i) estimate the seroprevalence of C. burnetii in sheep and goats, (ii) quantify the amount of bacteria shed during abortion and (iii) analyse temporal trends in human C. burnetii infections. The seroprevalence of C. burnetii in small ruminants was determined by commercial ELISA from a representative sample of 100 sheep flocks and 72 goat herds. Herd-level seroprevalence was 5.0% (95% CI: 1.6-11.3) for sheep and 11.1% (95% CI: 4.9-20.7) for goats. Animal-level seroprevalence was 1.8% (95% CI: 0.8-3.4) for sheep and 3.4% (95% CI: 1.7-6) for goats. The quantification of C. burnetii in 97 ovine and caprine abortion samples by real-time PCR indicated shedding of >10 4 bacteria/g in 13.4% of all samples tested. To our knowledge, this is the first study reporting C. burnetii quantities in a large number of small ruminant abortion samples. Annual human Q fever serology data were provided by five major Swiss laboratories. Overall, seroprevalence in humans ranged between 1.7% and 3.5% from 2007 to 2011, and no temporal trends were observed. Interestingly, the two laboratories with significantly higher seroprevalences are located in the regions with the largest goat populations as well as, for one laboratory, with the highest livestock density in Switzerland. However, a direct link between animal and human infection data could not be established in this study. © 2015 Blackwell Verlag GmbH.
Thorlund, Kristian; Imberger, Georgina; Walsh, Michael; Chu, Rong; Gluud, Christian; Wetterslev, Jørn; Guyatt, Gordon; Devereaux, Philip J.; Thabane, Lehana
2011-01-01
Background Meta-analyses including a limited number of patients and events are prone to yield overestimated intervention effect estimates. While many assume bias is the cause of overestimation, theoretical considerations suggest that random error may be an equal or more frequent cause. The independent impact of random error on meta-analyzed intervention effects has not previously been explored. It has been suggested that surpassing the optimal information size (i.e., the required meta-analysis sample size) provides sufficient protection against overestimation due to random error, but this claim has not yet been validated. Methods We simulated a comprehensive array of meta-analysis scenarios where no intervention effect existed (i.e., relative risk reduction (RRR) = 0%) or where a small but possibly unimportant effect existed (RRR = 10%). We constructed different scenarios by varying the control group risk, the degree of heterogeneity, and the distribution of trial sample sizes. For each scenario, we calculated the probability of observing overestimates of RRR>20% and RRR>30% for each cumulative 500 patients and 50 events. We calculated the cumulative number of patients and events required to reduce the probability of overestimation of intervention effect to 10%, 5%, and 1%. We calculated the optimal information size for each of the simulated scenarios and explored whether meta-analyses that surpassed their optimal information size had sufficient protection against overestimation of intervention effects due to random error. Results The risk of overestimation of intervention effects was usually high when the number of patients and events was small and this risk decreased exponentially over time as the number of patients and events increased. The number of patients and events required to limit the risk of overestimation depended considerably on the underlying simulation settings. Surpassing the optimal information size generally provided sufficient protection against overestimation. Conclusions Random errors are a frequent cause of overestimation of intervention effects in meta-analyses. Surpassing the optimal information size will provide sufficient protection against overestimation. PMID:22028777
Mdluli, Fezile; Thamaga-Chitja, Joyce; Schmidt, Stefan
2013-01-01
During October, November and December 2011 (when highest sales of Agri-Hub fresh produce are observed), irrigation water, compost, lettuce and spinach sampled from four different farmer cooperatives supplying the local Agri-Hub in uMbumbulu (KwaZulu-Natal, South Africa) were analyzed monthly for the presence of total and fecal coliforms and Escherichia coli using the most probable number (MPN) technique. The pH values for all irrigation water samples analyzed were within the acceptable range of 6.5–8.5 for agricultural use. Fecal coliform levels were <1,000 MPN per 100 mL irrigation water and <1,000 MPN per g of compost. The vegetables produced by Agri-Hub small-scale farmers met the requirements for total coliforms of <200/g set by the South African Department of Health at the time of sampling. E. coli MPN values for irrigation water and vegetables were below the limit of detection. In addition, the farming practices of 73 farmers were assessed via a survey. The results revealed that more than 40% of farmers used microbiologically safe tap water for irrigation and that trained farmers have a significantly better understanding of the importance of production hygiene than untrained farmers. These results reiterate the importance of interventions that build capacity in the area of food safety and hygiene of small-scale farmers for market access of formal value chains. PMID:24065036
Mdluli, Fezile; Thamaga-Chitja, Joyce; Schmidt, Stefan
2013-09-13
During October, November and December 2011 (when highest sales of Agri-Hub fresh produce are observed), irrigation water, compost, lettuce and spinach sampled from four different farmer cooperatives supplying the local Agri-Hub in uMbumbulu (KwaZulu-Natal, South Africa) were analyzed monthly for the presence of total and fecal coliforms and Escherichia coli using the most probable number (MPN) technique. The pH values for all irrigation water samples analyzed were within the acceptable range of 6.5-8.5 for agricultural use. Fecal coliform levels were <1,000 MPN per 100 mL irrigation water and <1,000 MPN per g of compost. The vegetables produced by Agri-Hub small-scale farmers met the requirements for total coliforms of <200/g set by the South African Department of Health at the time of sampling. E. coli MPN values for irrigation water and vegetables were below the limit of detection. In addition, the farming practices of 73 farmers were assessed via a survey. The results revealed that more than 40% of farmers used microbiologically safe tap water for irrigation and that trained farmers have a significantly better understanding of the importance of production hygiene than untrained farmers. These results reiterate the importance of interventions that build capacity in the area of food safety and hygiene of small-scale farmers for market access of formal value chains.
NASA Astrophysics Data System (ADS)
Mangold, Simon; Doppler, Tobias; Spycher, Simon; Langer, Miriam; Junghans, Marion; Kunz, Manuel; Stamm, Christian; Singer, Heinz
2017-04-01
Agricultural pesticides are regularly found in many surface waters draining agricultural areas. Due to large fluctuations in concentration over time and the potentially high number of pesticides, it is difficult to obtain a complete overview of the real pollution level. This collaborative project between research, federal and cantonal authorities in Switzerland aimed for a comprehensive assessment of pesticide pollution in five small agricultural streams to tackle this knowledge gap. The five streams are located in catchments (1.5 to 9 km2) with intensive agriculture covering a wide range of crops including vegetables, vineyards and orchards. Twelve-hour composite samples were collected continuously from March until the end of August 2015 with automatic sampling devices, yielding 360 samples per site. Using precipitation and water level data, we differentiated between discharge events and low-flow periods. Samples from discharge events where measured individually whereas samples taken during dry weather were pooled for the analysis. This procedure resulted in a complete concentration profile over the entire monitoring period covered by 34 - 60 samples per site. The analysis, using liquid chromatography coupled to high resolution mass spectrometry involved a target screening of about 220 pesticides. The measured concentrations were compared to chronic and acute environmental quality standards (EQS values) resulting in risk quotients RQs, which are the ratios between measured concentrations and the respective EQS values. Despite the small size of the catchments, we observed a large pesticide diversity in all of them with 68 to 103 detected compounds per study area. At all sites, chronic EQS values were exceeded. However, the exposure levels varied substantially among catchments. Maximum chronic RQs per site ranged between 1.1 and 48.8 and the duration of EQS exceedance varied between 2 weeks and 5.5 months. Additionally, the data reveal (very) high concentration peaks reaching up to 40 μg L-1 for single active ingredients. Of 15 compounds measured at high concentrations, several measured concentrations exceeded acute EQS values in three of the five areas for a duration of up to 2.5 months. Concentration peaks were often linked to discharge events but not exclusively. These findings demonstrate that rain driven processes were important causes of the observed concentration levels but that additional (possibly point) sources need to be considered for a comprehensive understanding of pesticide exposure. Overall, the results from these five catchments provide an unique insight into the diversity of pesticide pollution of small streams across a wide range of natural conditions in Switzerland. The spatial differences indicate that the intensity of pesticide use alone cannot explain the level of exposure but point to the influence of landscape characteristics such as topography, the connectivity of field to the stream network or the number of connected farmyards.
Relationship of order and number of siblings to perceived parental attitudes in childhood.
Kitamura, T; Sugawara, M; Shima, S; Toda, M A
1998-06-01
Despite the increasingly recognized link between perceived parenting behavior and the onset of psychopathology in adults, studies of the possible determinants of perceptions of parenting behavior are rare. In a sample of 1,145 pregnant Japanese women, correlations were examined between the numbers and sexes of siblings and perceived rearing practices, as rated by the Parental Bonding Instrument (PBI; Parker, Tupling, & Brown, 1979). The participants with more elder sisters viewed their parents' attitudes as less caring, whereas those with more brothers, particularly younger brothers, viewed their parents' attitudes as less overprotective. However, the proportion of the variance of all the PBI scores explained by different types of siblings was very small.
Characterization of Human Salivary Extracellular RNA by Next-generation Sequencing.
Li, Feng; Kaczor-Urbanowicz, Karolina Elżbieta; Sun, Jie; Majem, Blanca; Lo, Hsien-Chun; Kim, Yong; Koyano, Kikuye; Liu Rao, Shannon; Young Kang, So; Mi Kim, Su; Kim, Kyoung-Mee; Kim, Sung; Chia, David; Elashoff, David; Grogan, Tristan R; Xiao, Xinshu; Wong, David T W
2018-04-23
It was recently discovered that abundant and stable extracellular RNA (exRNA) species exist in bodily fluids. Saliva is an emerging biofluid for biomarker development for noninvasive detection and screening of local and systemic diseases. Use of RNA-Sequencing (RNA-Seq) to profile exRNA is rapidly growing; however, no single preparation and analysis protocol can be used for all biofluids. Specifically, RNA-Seq of saliva is particularly challenging owing to high abundance of bacterial contents and low abundance of salivary exRNA. Given the laborious procedures needed for RNA-Seq library construction, sequencing, data storage, and data analysis, saliva-specific and optimized protocols are essential. We compared different RNA isolation methods and library construction kits for long and small RNA sequencing. The role of ribosomal RNA (rRNA) depletion also was evaluated. The miRNeasy Micro Kit (Qiagen) showed the highest total RNA yield (70.8 ng/mL cell-free saliva) and best small RNA recovery, and the NEBNext library preparation kits resulted in the highest number of detected human genes [5649-6813 at 1 reads per kilobase RNA per million mapped (RPKM)] and small RNAs [482-696 microRNAs (miRNAs) and 190-214 other small RNAs]. The proportion of human RNA-Seq reads was much higher in rRNA-depleted saliva samples (41%) than in samples without rRNA depletion (14%). In addition, the transfer RNA (tRNA)-derived RNA fragments (tRFs), a novel class of small RNAs, were highly abundant in human saliva, specifically tRF-4 (4%) and tRF-5 (15.25%). Our results may help in selection of the best adapted methods of RNA isolation and small and long RNA library constructions for salivary exRNA studies. © 2018 American Association for Clinical Chemistry.
Koreňová, Janka; Rešková, Zuzana; Véghová, Adriana; Kuchta, Tomáš
2015-01-01
Contamination by Staphylococcus aureus of the production environment of three small or medium-sized food-processing factories in Slovakia was investigated on the basis of sub-species molecular identification by multiple locus variable number of tandem repeats analysis (MLVA). On the basis of MLVA profiling, bacterial isolates were assigned to 31 groups. Data from repeated samplings over a period of 3 years facilitated to draw spatial and temporal maps of the contamination routes for individual factories, as well as identification of potential persistent strains. Information obtained by MLVA typing allowed to identify sources and routes of contamination and, subsequently, will allow to optimize the technical and sanitation measures to ensure hygiene.
An empirical study of flight control software reliability
NASA Technical Reports Server (NTRS)
Dunham, J. R.; Pierce, J. L.
1986-01-01
The results of a laboratory experiment in flight control software reliability are reported. The experiment tests a small sample of implementations of a pitch axis control law for a PA28 aircraft with over 14 million pitch commands with varying levels of additive input and feedback noise. The testing which uses the method of n-version programming for error detection surfaced four software faults in one implementation of the control law. The small number of detected faults precluded the conduct of the error burst analyses. The pitch axis problem provides data for use in constructing a model in the prediction of the reliability of software in systems with feedback. The study is undertaken to find means to perform reliability evaluations of flight control software.
Feng, Hongxiang; Zhang, Zhenrong; Qing, Xin; Wang, Xiaowei; Liang, Chaoyang; Liu, Deruo
2016-02-01
Aberrant promoter hypermethylations of tumor suppressor genes are promising markers for lung cancer diagnosis and prognosis. The purpose of this study was to determine methylation status at APC and RAR-β promoters in primary NSCLC, and whether they have any relationship with survival. APC and RAR-β promoter methylation status were determined in 41 NSCLC patients using methylation specific PCR. APC promoter methylation was detectable in 9 (22.0%) tumor samples and 6 (14.6%) corresponding non-tumor samples (P=0.391). RAR-β promoter methylation was detectable in 13 (31.7%) tumor samples and 4 (9.8%) corresponding non-tumor samples (P=0.049) in the NSCLC patients. APC promoter methylation was found to be associated with T stage (P=0.046) and nodal status (P=0.019) in non-tumor samples, and with smoking (P=0.004) in tumor samples. RAR-β promoter methylation was found associated with age (P=0.031) in non-tumor samples and with primary tumor site in tumor samples. Patients with APC promoter methylation in tumor samples showed significantly longer survival than patients without it (Log-rank P=0.014). In a multivariate analysis of prognostic factors, APC methylation in tumor samples was an independent prognostic factor (P=0.012), as were N1 positive lymph node number (P=0.025) and N2 positive lymph node number (P=0.06). Our study shows that RAR-β methylation detected in lung tissue may be used as a predictive marker for NSCLC diagnosis and that APC methylation in tumor sample may be a useful marker for superior survival in NSCLC patients. Copyright © 2015. Published by Elsevier Inc.
Oh, Paul; Lee, Sukho; Kang, Moon Gi
2017-01-01
Recently, several RGB-White (RGBW) color filter arrays (CFAs) have been proposed, which have extra white (W) pixels in the filter array that are highly sensitive. Due to the high sensitivity, the W pixels have better SNR (Signal to Noise Ratio) characteristics than other color pixels in the filter array, especially, in low light conditions. However, most of the RGBW CFAs are designed so that the acquired RGBW pattern image can be converted into the conventional Bayer pattern image, which is then again converted into the final color image by using conventional demosaicing methods, i.e., color interpolation techniques. In this paper, we propose a new RGBW color filter array based on a totally different color interpolation technique, the colorization algorithm. The colorization algorithm was initially proposed for colorizing a gray image into a color image using a small number of color seeds. Here, we adopt this algorithm as a color interpolation technique, so that the RGBW color filter array can be designed with a very large number of W pixels to make the most of the highly sensitive characteristics of the W channel. The resulting RGBW color filter array has a pattern with a large proportion of W pixels, while the small-numbered RGB pixels are randomly distributed over the array. The colorization algorithm makes it possible to reconstruct the colors from such a small number of RGB values. Due to the large proportion of W pixels, the reconstructed color image has a high SNR value, especially higher than those of conventional CFAs in low light condition. Experimental results show that many important information which are not perceived in color images reconstructed with conventional CFAs are perceived in the images reconstructed with the proposed method. PMID:28657602
Oh, Paul; Lee, Sukho; Kang, Moon Gi
2017-06-28
Recently, several RGB-White (RGBW) color filter arrays (CFAs) have been proposed, which have extra white (W) pixels in the filter array that are highly sensitive. Due to the high sensitivity, the W pixels have better SNR (Signal to Noise Ratio) characteristics than other color pixels in the filter array, especially, in low light conditions. However, most of the RGBW CFAs are designed so that the acquired RGBW pattern image can be converted into the conventional Bayer pattern image, which is then again converted into the final color image by using conventional demosaicing methods, i.e., color interpolation techniques. In this paper, we propose a new RGBW color filter array based on a totally different color interpolation technique, the colorization algorithm. The colorization algorithm was initially proposed for colorizing a gray image into a color image using a small number of color seeds. Here, we adopt this algorithm as a color interpolation technique, so that the RGBW color filter array can be designed with a very large number of W pixels to make the most of the highly sensitive characteristics of the W channel. The resulting RGBW color filter array has a pattern with a large proportion of W pixels, while the small-numbered RGB pixels are randomly distributed over the array. The colorization algorithm makes it possible to reconstruct the colors from such a small number of RGB values. Due to the large proportion of W pixels, the reconstructed color image has a high SNR value, especially higher than those of conventional CFAs in low light condition. Experimental results show that many important information which are not perceived in color images reconstructed with conventional CFAs are perceived in the images reconstructed with the proposed method.
76 FR 32069 - Safety Zone; Lorain Independence Day Fireworks, Black River, Lorain, OH
Federal Register 2010, 2011, 2012, 2013, 2014
2011-06-03
... grant or loan recipients, and will not raise any novel legal or policy issues. Small Entities Under the... economic impact on a substantial number of small entities. The term ``small entities'' comprises small... number of small entities. This rule will affect the following entities, some of which may be small...
The Stratigraphy and Evolution of the Lunar Crust
NASA Technical Reports Server (NTRS)
McCallum, I. Stewart
1998-01-01
Reconstruction of stratigraphic relationships in the ancient lunar crust has proved to be a formidable task. The intense bombardment during the first 700 m.y. of lunar history has severely perturbed the original stratigraphy and destroyed the primary textures of all but a few nonmare rocks. However, a knowledge of the crustal stratigraphy as it existed prior to the cataclysmic bombardment about 3.9 Ga is essential to test the major models proposed for crustal origin, i.e., crystal fractionation in a global magmasphere or serial magmatism in a large number of smaller bodies. Despite the large difference in scale implicit in these two models, both require an efficient separation of plagioclase and mafic minerals to form the anorthositic crust and the mafic mantle. Despite the havoc wreaked by the large body impactors, these same impact processes have brought to the lunar surface crystalline samples derived from at least the upper half of the lunar crust, thereby providing an opportunity to reconstruct the stratigraphy in areas sampled by the Apollo missions. As noted, ejecta from the large multiring basins are dominantly, or even exclusively, of crustal origin. Given the most recent determinations of crustal thicknesses, this implies an upper limit to the depth of excavation of about 60 km. Of all the lunar samples studied, a small set has been recognized as "pristine", and within this pristine group, a small fraction have retained some vestiges of primary features formed during the earliest stages of crystallization or recrystallization prior to 4.0 Ga. We have examined a number of these samples that have retained some record of primary crystallization to deduce thermal histories from an analysis of structural, textural, and compositional features in minerals from these samples. Specifically, by quantitative modeling of (1) the growth rate and development of compositional profiles of exsolution lamellae in pyroxenes and (2) the rate of Fe-Mg ordering in orthopyroxenes, we can constrain the cooling rates of appropriate lunar samples. These cooling rates are used to compute depths of burial at the time of crystallization, which enable us to reconstruct parts of the crustal stratigraphy as it existed during the earliest stages of lunar history.
NASA Technical Reports Server (NTRS)
Tsuchiyama, A.; Ebihara, M.; Kimura, M.; Kitajima, F.; Kotsugi, M.; Ito, S.; Nagao, K.; Nakamura, T.; Naraoka, H.; Noguchi, T.;
2011-01-01
The Hayabusa spacecraft arrived at S-type Asteroid 25143 Itokawa in November 2006, and reveal astounding features of the small asteroid (535 x 294 x 209 m). Near-infrared spectral shape indicates that the surface of this body has an olivinerich mineral assemblage potentially similar to that of LL5 or LL6 chondrites with different degrees of space weathering. Based on the surface morphological features observed in high-resolution images of Itokawa s surface, two major types of boulders were distinguished: rounded and angular boulders. Rounded boulders seem to be breccias, while angular boulders seem to have severe impact origin. Although the sample collection did not be made by normal operations, it was considered that some amount of samples, probably small particles of regolith, was collected from MUSES-C regio on the Itokawa s surface. The sample capsule was successfully recovered on the earth on June 13, 2010, and was opened at curation facility of JAXA (Japan Aerospace Exploration Agency), Sagamihara, Japan. A large number of small particles were found in the sample container. Preliminary analysis with SEM/EDX at the curation facility showed that at least more than 1500 grains were identified as rocky particles, and most of them were judged to be of extraterrestrial origin, and definitely from Asteroid Itokawa. Minerals (olivine, low-Ca pyroxene, high-Ca pyroxene, plagioclase, Fe sulfide, Fe-Ni metal, chromite, Ca phosphate), roughly estimated mode the minerals and rough measurement of the chemical compositions of the silicates show that these particles are roughly similar to LL chondrites. Although their size are mostly less than 10 m, some larger particles of about 100 m or larger were also identified. A part of the sample (probably several tens particles) will be selected by Hayabusa sample curation team and examined preliminary in Japan within one year after the sample recovery in prior to detailed analysis phase. Hayabusa Asteroidal Sample Preliminary Examination Team (HASPET) has been preparing for the preliminary examination with close cooperation with the curation team.
Apparatus and process for microbial detection and enumeration
NASA Technical Reports Server (NTRS)
Wilkins, J. R.; Grana, D. (Inventor)
1982-01-01
An apparatus and process for detecting and enumerating specific microorganisms from large volume samples containing small numbers of the microorganisms is presented. The large volume samples are filtered through a membrane filter to concentrate the microorganisms. The filter is positioned between two absorbent pads and previously moistened with a growth medium for the microorganisms. A pair of electrodes are disposed against the filter and the pad electrode filter assembly is retained within a petri dish by retainer ring. The cover is positioned on base of petri dish and sealed at the edges by a parafilm seal prior to being electrically connected via connectors to a strip chart recorder for detecting and enumerating the microorganisms collected on filter.
Podar, Mircea; Shakya, Migun; D'Amore, Rosalinda; ...
2016-01-14
In the last 5 years, the rapid pace of innovations and improvements in sequencing technologies has completely changed the landscape of metagenomic and metagenetic experiments. Therefore, it is critical to benchmark the various methodologies for interrogating the composition of microbial communities, so that we can assess their strengths and limitations. Here, the most common phylogenetic marker for microbial community diversity studies is the 16S ribosomal RNA gene and in the last 10 years the field has moved from sequencing a small number of amplicons and samples to more complex studies where thousands of samples and multiple different gene regions aremore » interrogated.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Okumura, Teppei; Seljak, Uroš; Desjacques, Vincent, E-mail: teppei@ewha.ac.kr, E-mail: useljak@berkeley.edu, E-mail: dvince@physik.uzh.ch
It was recently shown that the power spectrum in redshift space can be written as a sum of cross-power spectra between number weighted velocity moments, of which the lowest are density and momentum density. We investigate numerically the properties of these power spectra for simulated galaxies and dark matter halos and compare them to the dark matter power spectra, generalizing the concept of the bias in density-density power spectra. Because all of the quantities are number weighted this approach is well defined even for sparse systems such as massive halos. This contrasts to the previous approaches to RSD where velocitymore » correlations have been explored, but velocity field is a poorly defined concept for sparse systems. We find that the number density weighting leads to a strong scale dependence of the bias terms for momentum density auto-correlation and cross-correlation with density. This trend becomes more significant for the more biased halos and leads to an enhancement of RSD power relative to the linear theory. Fingers-of-god effects, which in this formalism come from the correlations of the higher order moments beyond the momentum density, lead to smoothing of the power spectrum and can reduce this enhancement of power from the scale dependent bias, but are relatively small for halos with no small scale velocity dispersion. In comparison, for a more realistic galaxy sample with satellites the small scale velocity dispersion generated by satellite motions inside the halos leads to a larger power suppression on small scales, but this depends on the satellite fraction and on the details of how the satellites are distributed inside the halo. We investigate several statistics such as the two-dimensional power spectrum P(k,μ), where μ is the angle between the Fourier mode and line of sight, its multipole moments, its powers of μ{sup 2}, and configuration space statistics. Overall we find that the nonlinear effects in realistic galaxy samples such as luminous red galaxies affect the redshift space clustering on very large scales: for example, the quadrupole moment is affected by 10% for k < 0.1hMpc{sup −1}, which means that these effects need to be understood if we want to extract cosmological information from the redshift space distortions.« less
7 CFR 201.42 - Small containers.
Code of Federal Regulations, 2010 CFR
2010-01-01
... 7 Agriculture 3 2010-01-01 2010-01-01 false Small containers. 201.42 Section 201.42 Agriculture... REGULATIONS Sampling in the Administration of the Act § 201.42 Small containers. In sampling seed in small containers that it is not practical to sample as required in § 201.41, a portion of one unopened container or...
Fink, Susanne; Tsai, Ming-Han; Schnitzler, Paul; Zeier, Martin; Dreger, Peter; Wuchter, Patrick; Bulut, Olcay C; Behrends, Uta; Delecluse, Henri-Jacques
2017-01-01
Transplant recipients frequently exhibit an increased Epstein-Barr virus (EBV) load in the peripheral blood. Here, we quantitated the EBV-infected cells in the peripheral blood of these patients and defined the mode of viral infection, latent or lytic. These data indicated that there is no strong correlation between the number of infected cells and the EBV load (EBVL). This can be explained by a highly variable number of EBV copies per infected cell and by lytic replication in some cells. The plasma of these patients did not contain any free infectious viruses, but contained nevertheless EBV DNA, sometimes in large amounts, that probably originates from cell debris and contributed to the total EBVL. Some of the investigated samples carried a highly variable number of infected cells in active latency, characterized by an expression of the Epstein-Barr nuclear antigens (EBNA2) protein. However, a third of the samples expressed neither EBNA2 nor lytic proteins. Patients with an increased EBVL represent a heterogeneous group of patients whose infection cannot be characterized by this method alone. Precise characterization of the origin of an increased EBVL, in particular, in terms of the number of EBV-infected cells, requires additional investigations including the number of EBV-encoded small RNA-positive cells. © 2016 Steunstichting ESOT.
Vallée, Isabelle; Macé, Pauline; Forbes, Lorry; Scandrett, Brad; Durand, Benoit; Gajadhar, Alvin; Boireau, Pascal
2007-07-01
Routine diagnosis of animal trichinellosis for food safety and trade relies on a method of artificial digestion to free Trichinella muscle larvae from meat for subsequent identification by microscopy. As part of a quality control system, the French National Reference Laboratory (NRL) initiated ring trials to determine the sensitivity of the test performed in the 72 routine diagnostic laboratories in France. A method was devised to obtain calibrated meat samples containing known numbers of capsules with Trichinella spiralis muscle larvae. This method was based on an incomplete artificial digestion of Trichinella-infected mice carcasses to allow the collection of intact Trichinella capsules. Capsules were placed into a meatball of 100 +/- 2 g of pork and horsemeat to produce proficiency samples. Three categories of samples were prepared: small (3 to 5 capsules), medium (7 to 10), and large (12 to 15). The sensitivity was expressed as the percentage of muscle larvae recovered from each proficiency sample. Reproducibility was tested with ring trials organized between two NRLs (France and Canada), and a reference sensitivity of 84.9% was established. National ring trials were then organized in France, with the 72 routine diagnostic laboratories each receiving four proficiency samples per session. After five sessions, an improvement in the digest test sensitivity was observed. Results at the fifth session indicated sensitivities of 78.60% +/- 23.70%, 81.19% +/- 19.59%, and 80.52% +/- 14.71% muscle larvae for small, medium, and large samples, respectively. This study supports the use of proficiency samples to accurately evaluate the performance of routine diagnostic laboratories that conduct digestion tests for animal trichinellosis diagnosis.
Planning multi-arm screening studies within the context of a drug development program
Wason, James M S; Jaki, Thomas; Stallard, Nigel
2013-01-01
Screening trials are small trials used to decide whether an intervention is sufficiently promising to warrant a large confirmatory trial. Previous literature examined the situation where treatments are tested sequentially until one is considered sufficiently promising to take forward to a confirmatory trial. An important consideration for sponsors of clinical trials is how screening trials should be planned to maximize the efficiency of the drug development process. It has been found previously that small screening trials are generally the most efficient. In this paper we consider the design of screening trials in which multiple new treatments are tested simultaneously. We derive analytic formulae for the expected number of patients until a successful treatment is found, and propose methodology to search for the optimal number of treatments, and optimal sample size per treatment. We compare designs in which only the best treatment proceeds to a confirmatory trial and designs in which multiple treatments may proceed to a multi-arm confirmatory trial. We find that inclusion of a large number of treatments in the screening trial is optimal when only one treatment can proceed, and a smaller number of treatments is optimal when more than one can proceed. The designs we investigate are compared on a real-life set of screening designs. Copyright © 2013 John Wiley & Sons, Ltd. PMID:23529936
Predicted trends in the supply and demand of veterinarians in Japan.
Kimura, S; Shinkawa, S; Mago, J; Yamamoto, M; Sakai, M; Sugisaki, T; Karaki, H; Sugiura, K
2008-12-01
Currently in Japan, there are 32,000 active veterinarians, mainly engaged in small and large animal practice and public animal health and public health services. In the face of the notable increase in recent years in the proportion of female students enrolled in veterinary schools and in the number of households with companion animals, a model was developed to predict the supply and demand of veterinarians toward 2040 in Japan. Surveys were conducted on sampled households and veterinarians to estimate input variables used in the supply and demand model. From this data it is predicted that there might be somewhere between a shortage of 1,000 to an over-supply of 3,700 veterinarians engaged in small animal practice in 2040. This, however, will depend on possible changes in the number of visits made to veterinarians by small animal owners and the efficiency of practices in the future. The model also predicts that there will be a shortage of around 1,100 veterinarians in large animal practice in 2040. Considering the many assumptions made to estimate the input variables used in the model, the results of this study do not provide definitive conclusions, but provide a base for discussions on what will be needed in the veterinary profession in the future.
Authorship and sampling practice in selected biomechanics and sports science journals.
Knudson, Duane V
2011-06-01
In some biomedical sciences, changes in patterns of collaboration and authorship have complicated the assignment of credit and responsibility for research. It is unclear if this problem of "promiscuous coauthorship" or "hyperauthorship" (defined as six or more authors) is also apparent in the applied research disciplines within sport and exercise science. This study documented the authorship and sampling of patterns of original research reports in three applied biomechanics (Clinical Biomechanics, Journal of Applied Biomechanics, and Sports Biomechanics) and five similar subdisciplinary journals within sport and exercise science (International Journal of Sports Physiology and Performance, Journal of Sport Rehabilitation, Journal of Teaching Physical Education, Measurement in Physical Education and Exercise Sciences, and Motor Control). Original research reports from the 2009 volumes of these biomechanics and sport and exercise journals were reviewed. Single authorship of papers was rare (2.6%) in these journals, with the mean number of authors ranging from 2.7 to 4.5. Sample sizes and the ratio of sample to authors varied widely, and these variables tended not to be associated with number of authors. Original research reports published in these journals in 2009 tended to be published by small teams of collaborators, so currently there may be few problems with promiscuous coauthorship in these subdisciplines of sport and exercise science.
Kopek, Benjamin G.; Paez-Segala, Maria G.; Shtengel, Gleb; Sochacki, Kem A.; Sun, Mei G.; Wang, Yalin; Xu, C. Shan; van Engelenburg, Schuyler B.; Taraska, Justin W.; Looger, Loren L.; Hess, Harald F.
2017-01-01
Our groups have recently developed related approaches for sample preparation for super-resolution imaging within endogenous cellular environments using correlative light and electron microscopy (CLEM). Four distinct techniques for preparing and acquiring super-resolution CLEM datasets on aldehyde-fixed specimens are provided, including Tokuyasu cryosectioning, whole-cell mount, cell unroofing and platinum replication, and resin embedding and sectioning. Choice of the best protocol for a given application depends on a number of criteria that are discussed in detail. Tokuyasu cryosectioning is relatively rapid but is limited to small, delicate specimens. Whole-cell mount has the simplest sample preparation but is restricted to surface structures. Cell unroofing and platinum replica creates high-contrast, 3-dimensional images of the cytoplasmic surface of the plasma membrane, but is more challenging than whole-cell mount. Resin embedding permits serial sectioning of large samples, but is limited to osmium-resistant probes, and is technically difficult. Expected results from these protocols include super-resolution localization (~10–50 nm) of fluorescent targets within the context of electron microscopy ultrastructure, which can help address cell biological questions. These protocols can be completed in 2–7 days, are compatible with a number of super-resolution imaging protocols, and are broadly applicable across biology. PMID:28384138
Label-free virus detection using silicon photonic microring resonators
McClellan, Melinda S.; Domier, Leslie L; Bailey, Ryan C.
2013-01-01
Viruses represent a continual threat to humans through a number of mechanisms, which include disease, bioterrorism, and destruction of both plant and animal food resources. Many contemporary techniques used for the detection of viruses and viral infections suffer from limitations such as the need for extensive sample preparation or the lengthy window between infection and measurable immune response, for serological methods. In order to develop a method that is fast, cost-effective, and features reduced sample preparation compared to many other virus detection methods, we report the application of silicon photonic microring resonators for the direct, label-free detection of intact viruses in both purified samples as well as in a complex, real-world analytical matrix. As a model system, we demonstrate the quantitative detection of Bean pod mottle virus, a pathogen of great agricultural importance, with a limit of detection of 10 ng/mL. By simply grinding a small amount of leaf sample in buffer with a mortar and pestle, infected leaves can be identified over a healthy control with a total analysis time of less than 45 min. Given the inherent scalability and multiplexing capability of the semiconductor-based technology, we feel that silicon photonic microring resonators are well-positioned as a promising analytical tool for a number of viral detection applications. PMID:22138465
Deepaisarn, S; Tar, P D; Thacker, N A; Seepujak, A; McMahon, A W
2018-01-01
Abstract Motivation Matrix-assisted laser desorption/ionisation time-of-flight mass spectrometry (MALDI) facilitates the analysis of large organic molecules. However, the complexity of biological samples and MALDI data acquisition leads to high levels of variation, making reliable quantification of samples difficult. We present a new analysis approach that we believe is well-suited to the properties of MALDI mass spectra, based upon an Independent Component Analysis derived for Poisson sampled data. Simple analyses have been limited to studying small numbers of mass peaks, via peak ratios, which is known to be inefficient. Conventional PCA and ICA methods have also been applied, which extract correlations between any number of peaks, but we argue makes inappropriate assumptions regarding data noise, i.e. uniform and Gaussian. Results We provide evidence that the Gaussian assumption is incorrect, motivating the need for our Poisson approach. The method is demonstrated by making proportion measurements from lipid-rich binary mixtures of lamb brain and liver, and also goat and cow milk. These allow our measurements and error predictions to be compared to ground truth. Availability and implementation Software is available via the open source image analysis system TINA Vision, www.tina-vision.net. Contact paul.tar@manchester.ac.uk Supplementary information Supplementary data are available at Bioinformatics online. PMID:29091994
Euliss, Ned H.; Mushet, David M.
1999-01-01
We evaluated the influence of intensive agriculture on invertebrate communities of temporary wetlands as indicated by aquatic invertebrate resting eggs, shells, and cases remaining after wetlands dried. To facilitate the comparison, we sampled 19 wetlands within cropland areas and 19 wetlands within grassland areas. We found resting eggs, shells, and cases of significantly more taxa and greater numbers of cladoceran resting eggs (ephippia), planorbid and physid snail shells, and ostracod shells in wetlands within grasslands than in croplands. We also successfully incubated greater numbers of cladocerans and ostracods from soil samples collected from grassland sites. We were unable to detect differences in the viability of cladoceran ephippia between grassland and cropland wetlands, but our sample size was small due to an absence of ephippia in most cropland wetlands sampled; 74% of the cropland wetlands were devoid of cladoceran ephippia whereas ephippia were well represented in nearly all of our grassland sites. Our results corroborate findings of other investigators that prairie pothole wetlands have been negatively impacted by human activ-, ities. Our study demonstrates that aquatic invertebrates of temporary wetlands have, been negatively impacted by intensive agriculture and suggests that future studies need to assess the influence of agricultural practices on wetland-dependant wildlife.