On Orbital Elements of Extrasolar Planetary Candidates and Spectroscopic Binaries
NASA Technical Reports Server (NTRS)
Stepinski, T. F.; Black, D. C.
2001-01-01
We estimate probability densities of orbital elements, periods, and eccentricities, for the population of extrasolar planetary candidates (EPC) and, separately, for the population of spectroscopic binaries (SB) with solar-type primaries. We construct empirical cumulative distribution functions (CDFs) in order to infer probability distribution functions (PDFs) for orbital periods and eccentricities. We also derive a joint probability density for period-eccentricity pairs in each population. Comparison of respective distributions reveals that in all cases EPC and SB populations are, in the context of orbital elements, indistinguishable from each other to a high degree of statistical significance. Probability densities of orbital periods in both populations have P(exp -1) functional form, whereas the PDFs of eccentricities can he best characterized as a Gaussian with a mean of about 0.35 and standard deviation of about 0.2 turning into a flat distribution at small values of eccentricity. These remarkable similarities between EPC and SB must be taken into account by theories aimed at explaining the origin of extrasolar planetary candidates, and constitute an important clue us to their ultimate nature.
Characterizing the Morphology, Distribution, and Formation Geometry of Mercury's Pyroclastic Vents
NASA Astrophysics Data System (ADS)
Jozwiak, L. M.; Head, J. W.; Wilson, L.
2018-05-01
We present a final catalog of pyroclastic vents on Mercury, identifying 104 candidate pyroclastic vents. We then assess the vent distribution, morphologic variation, and probable formation geometries.
VizieR Online Data Catalog: KOI transit probabilities of multi-planet syst. (Brakensiek+, 2016)
NASA Astrophysics Data System (ADS)
Brakensiek, J.; Ragozzine, D.
2016-06-01
Using CORBITS, we computed the transit probabilities of all the KOIs with at least three candidate or confirmed transiting planets and report the results in Table 2 for a variety of inclination distributions. See section 4.6. (1 data file).
Metocean design parameter estimation for fixed platform based on copula functions
NASA Astrophysics Data System (ADS)
Zhai, Jinjin; Yin, Qilin; Dong, Sheng
2017-08-01
Considering the dependent relationship among wave height, wind speed, and current velocity, we construct novel trivariate joint probability distributions via Archimedean copula functions. Total 30-year data of wave height, wind speed, and current velocity in the Bohai Sea are hindcast and sampled for case study. Four kinds of distributions, namely, Gumbel distribution, lognormal distribution, Weibull distribution, and Pearson Type III distribution, are candidate models for marginal distributions of wave height, wind speed, and current velocity. The Pearson Type III distribution is selected as the optimal model. Bivariate and trivariate probability distributions of these environmental conditions are established based on four bivariate and trivariate Archimedean copulas, namely, Clayton, Frank, Gumbel-Hougaard, and Ali-Mikhail-Haq copulas. These joint probability models can maximize marginal information and the dependence among the three variables. The design return values of these three variables can be obtained by three methods: univariate probability, conditional probability, and joint probability. The joint return periods of different load combinations are estimated by the proposed models. Platform responses (including base shear, overturning moment, and deck displacement) are further calculated. For the same return period, the design values of wave height, wind speed, and current velocity obtained by the conditional and joint probability models are much smaller than those by univariate probability. Considering the dependence among variables, the multivariate probability distributions provide close design parameters to actual sea state for ocean platform design.
Pan-European comparison of candidate distributions for climatological drought indices, SPI and SPEI
NASA Astrophysics Data System (ADS)
Stagge, James; Tallaksen, Lena; Gudmundsson, Lukas; Van Loon, Anne; Stahl, Kerstin
2013-04-01
Drought indices are vital to objectively quantify and compare drought severity, duration, and extent across regions with varied climatic and hydrologic regimes. The Standardized Precipitation Index (SPI), a well-reviewed meterological drought index recommended by the WMO, and its more recent water balance variant, the Standardized Precipitation-Evapotranspiration Index (SPEI) both rely on selection of univariate probability distributions to normalize the index, allowing for comparisons across climates. The SPI, considered a universal meteorological drought index, measures anomalies in precipitation, whereas the SPEI measures anomalies in climatic water balance (precipitation minus potential evapotranspiration), a more comprehensive measure of water availability that incorporates temperature. Many reviewers recommend use of the gamma (Pearson Type III) distribution for SPI normalization, while developers of the SPEI recommend use of the three parameter log-logistic distribution, based on point observation validation. Before the SPEI can be implemented at the pan-European scale, it is necessary to further validate the index using a range of candidate distributions to determine sensitivity to distribution selection, identify recommended distributions, and highlight those instances where a given distribution may not be valid. This study rigorously compares a suite of candidate probability distributions using WATCH Forcing Data, a global, historical (1958-2001) climate dataset based on ERA40 reanalysis with 0.5 x 0.5 degree resolution and bias-correction based on CRU-TS2.1 observations. Using maximum likelihood estimation, alternative candidate distributions are fit for the SPI and SPEI across the range of European climate zones. When evaluated at this scale, the gamma distribution for the SPI results in negatively skewed values, exaggerating the index severity of extreme dry conditions, while decreasing the index severity of extreme high precipitation. This bias is particularly notable for shorter aggregation periods (1-6 months) during the summer months in southern Europe (below 45° latitude), and can partially be attributed to distribution fitting difficulties in semi-arid regions where monthly precipitation totals cluster near zero. By contrast, the SPEI has potential for avoiding this fitting difficulty because it is not bounded by zero. However, the recommended log-logistic distribution produces index values with less variation than the standard normal distribution. Among the alternative candidate distributions, the best fit distribution and the distribution parameters vary in space and time, suggesting regional commonalities within hydroclimatic regimes, as discussed further in the presentation.
NASA Astrophysics Data System (ADS)
Zhang, Jiaxin; Shields, Michael D.
2018-01-01
This paper addresses the problem of uncertainty quantification and propagation when data for characterizing probability distributions are scarce. We propose a methodology wherein the full uncertainty associated with probability model form and parameter estimation are retained and efficiently propagated. This is achieved by applying the information-theoretic multimodel inference method to identify plausible candidate probability densities and associated probabilities that each method is the best model in the Kullback-Leibler sense. The joint parameter densities for each plausible model are then estimated using Bayes' rule. We then propagate this full set of probability models by estimating an optimal importance sampling density that is representative of all plausible models, propagating this density, and reweighting the samples according to each of the candidate probability models. This is in contrast with conventional methods that try to identify a single probability model that encapsulates the full uncertainty caused by lack of data and consequently underestimate uncertainty. The result is a complete probabilistic description of both aleatory and epistemic uncertainty achieved with several orders of magnitude reduction in computational cost. It is shown how the model can be updated to adaptively accommodate added data and added candidate probability models. The method is applied for uncertainty analysis of plate buckling strength where it is demonstrated how dataset size affects the confidence (or lack thereof) we can place in statistical estimates of response when data are lacking.
NASA Astrophysics Data System (ADS)
Li, Jia; Wang, Qiang; Yan, Wenjie; Shen, Yi
2015-12-01
Cooperative spectrum sensing exploits the spatial diversity to improve the detection of occupied channels in cognitive radio networks (CRNs). Cooperative compressive spectrum sensing (CCSS) utilizing the sparsity of channel occupancy further improves the efficiency by reducing the number of reports without degrading detection performance. In this paper, we firstly and mainly propose the referred multi-candidate orthogonal matrix matching pursuit (MOMMP) algorithms to efficiently and effectively detect occupied channels at fusion center (FC), where multi-candidate identification and orthogonal projection are utilized to respectively reduce the number of required iterations and improve the probability of exact identification. Secondly, two common but different approaches based on threshold and Gaussian distribution are introduced to realize the multi-candidate identification. Moreover, to improve the detection accuracy and energy efficiency, we propose the matrix construction based on shrinkage and gradient descent (MCSGD) algorithm to provide a deterministic filter coefficient matrix of low t-average coherence. Finally, several numerical simulations validate that our proposals provide satisfactory performance with higher probability of detection, lower probability of false alarm and less detection time.
A Gaussian Model-Based Probabilistic Approach for Pulse Transit Time Estimation.
Jang, Dae-Geun; Park, Seung-Hun; Hahn, Minsoo
2016-01-01
In this paper, we propose a new probabilistic approach to pulse transit time (PTT) estimation using a Gaussian distribution model. It is motivated basically by the hypothesis that PTTs normalized by RR intervals follow the Gaussian distribution. To verify the hypothesis, we demonstrate the effects of arterial compliance on the normalized PTTs using the Moens-Korteweg equation. Furthermore, we observe a Gaussian distribution of the normalized PTTs on real data. In order to estimate the PTT using the hypothesis, we first assumed that R-waves in the electrocardiogram (ECG) can be correctly identified. The R-waves limit searching ranges to detect pulse peaks in the photoplethysmogram (PPG) and to synchronize the results with cardiac beats--i.e., the peaks of the PPG are extracted within the corresponding RR interval of the ECG as pulse peak candidates. Their probabilities of being the actual pulse peak are then calculated using a Gaussian probability function. The parameters of the Gaussian function are automatically updated when a new pulse peak is identified. This update makes the probability function adaptive to variations of cardiac cycles. Finally, the pulse peak is identified as the candidate with the highest probability. The proposed approach is tested on a database where ECG and PPG waveforms are collected simultaneously during the submaximal bicycle ergometer exercise test. The results are promising, suggesting that the method provides a simple but more accurate PTT estimation in real applications.
2010-01-01
As part of our effort to increase survival of drug candidates and to move our medicinal chemistry design to higher probability space for success in the Neuroscience therapeutic area, we embarked on a detailed study of the property space for a collection of central nervous system (CNS) molecules. We carried out a thorough analysis of properties for 119 marketed CNS drugs and a set of 108 Pfizer CNS candidates. In particular, we focused on understanding the relationships between physicochemical properties, in vitro ADME (absorption, distribution, metabolism, and elimination) attributes, primary pharmacology binding efficiencies, and in vitro safety data for these two sets of compounds. This scholarship provides guidance for the design of CNS molecules in a property space with increased probability of success and may lead to the identification of druglike candidates with favorable safety profiles that can successfully test hypotheses in the clinic. PMID:22778836
Probability distribution of extreme share returns in Malaysia
NASA Astrophysics Data System (ADS)
Zin, Wan Zawiah Wan; Safari, Muhammad Aslam Mohd; Jaaman, Saiful Hafizah; Yie, Wendy Ling Shin
2014-09-01
The objective of this study is to investigate the suitable probability distribution to model the extreme share returns in Malaysia. To achieve this, weekly and monthly maximum daily share returns are derived from share prices data obtained from Bursa Malaysia over the period of 2000 to 2012. The study starts with summary statistics of the data which will provide a clue on the likely candidates for the best fitting distribution. Next, the suitability of six extreme value distributions, namely the Gumbel, Generalized Extreme Value (GEV), Generalized Logistic (GLO) and Generalized Pareto (GPA), the Lognormal (GNO) and the Pearson (PE3) distributions are evaluated. The method of L-moments is used in parameter estimation. Based on several goodness of fit tests and L-moment diagram test, the Generalized Pareto distribution and the Pearson distribution are found to be the best fitted distribution to represent the weekly and monthly maximum share returns in Malaysia stock market during the studied period, respectively.
Bayesian structural inference for hidden processes.
Strelioff, Christopher C; Crutchfield, James P
2014-04-01
We introduce a Bayesian approach to discovering patterns in structurally complex processes. The proposed method of Bayesian structural inference (BSI) relies on a set of candidate unifilar hidden Markov model (uHMM) topologies for inference of process structure from a data series. We employ a recently developed exact enumeration of topological ε-machines. (A sequel then removes the topological restriction.) This subset of the uHMM topologies has the added benefit that inferred models are guaranteed to be ε-machines, irrespective of estimated transition probabilities. Properties of ε-machines and uHMMs allow for the derivation of analytic expressions for estimating transition probabilities, inferring start states, and comparing the posterior probability of candidate model topologies, despite process internal structure being only indirectly present in data. We demonstrate BSI's effectiveness in estimating a process's randomness, as reflected by the Shannon entropy rate, and its structure, as quantified by the statistical complexity. We also compare using the posterior distribution over candidate models and the single, maximum a posteriori model for point estimation and show that the former more accurately reflects uncertainty in estimated values. We apply BSI to in-class examples of finite- and infinite-order Markov processes, as well to an out-of-class, infinite-state hidden process.
Bayesian structural inference for hidden processes
NASA Astrophysics Data System (ADS)
Strelioff, Christopher C.; Crutchfield, James P.
2014-04-01
We introduce a Bayesian approach to discovering patterns in structurally complex processes. The proposed method of Bayesian structural inference (BSI) relies on a set of candidate unifilar hidden Markov model (uHMM) topologies for inference of process structure from a data series. We employ a recently developed exact enumeration of topological ɛ-machines. (A sequel then removes the topological restriction.) This subset of the uHMM topologies has the added benefit that inferred models are guaranteed to be ɛ-machines, irrespective of estimated transition probabilities. Properties of ɛ-machines and uHMMs allow for the derivation of analytic expressions for estimating transition probabilities, inferring start states, and comparing the posterior probability of candidate model topologies, despite process internal structure being only indirectly present in data. We demonstrate BSI's effectiveness in estimating a process's randomness, as reflected by the Shannon entropy rate, and its structure, as quantified by the statistical complexity. We also compare using the posterior distribution over candidate models and the single, maximum a posteriori model for point estimation and show that the former more accurately reflects uncertainty in estimated values. We apply BSI to in-class examples of finite- and infinite-order Markov processes, as well to an out-of-class, infinite-state hidden process.
Comparison of spectra using a Bayesian approach. An argument using oil spills as an example.
Li, Jianfeng; Hibbert, D Brynn; Fuller, Steven; Cattle, Julie; Pang Way, Christopher
2005-01-15
The problem of assigning a probability of matching a number of spectra is addressed. The context is in environmental spills when an EPA needs to show that the material from a polluting spill (e.g., oil) is likely to have originated at a particular site (factory, refinery) or from a vehicle (road tanker or ship). Samples are taken from the spill, and candidate sources and are analyzed by spectroscopy (IR, fluorescence) or chromatography (GC or GC/MS). A matching algorithm is applied to pairs of spectra giving a single statistic (R). This can be a point-to-point match giving a correlation coefficient or a Euclidean distance or a derivative of these parameters. The distributions of R for same and different samples are established from existing data. For matching statistics with values in the range {0,1} corresponding to no match (0) to a perfect match (1) a beta distribution can be fitted to most data. The values of R from the match of the spectrum of a spilled oil and of each of a number of suspects are calculated and Bayes' theorem is applied to give a probability of matches between spill sample and each candidate and the probability of no match at all. The method is most effective when simple inspection of the matching parameters does not lead to an obvious conclusion; i.e., there is overlap of the distributions giving rise to dubiety of an assignment. The probability of finding a matching statistic if there were a match to the probability of finding it if there were no match, expressed as a ratio (called the likelihood ratio), is a sensitive and useful parameter to guide the analyst. It is proposed that this approach may be acceptable to a court of law and avoid challenges of apparently subjective opinion of an analyst. Examples of matching the fluorescence and infrared spectra of diesel oils are given.
Scaling properties and universality of first-passage-time probabilities in financial markets
NASA Astrophysics Data System (ADS)
Perelló, Josep; Gutiérrez-Roig, Mario; Masoliver, Jaume
2011-12-01
Financial markets provide an ideal frame for the study of crossing or first-passage time events of non-Gaussian correlated dynamics, mainly because large data sets are available. Tick-by-tick data of six futures markets are herein considered, resulting in fat-tailed first-passage time probabilities. The scaling of the return with its standard deviation collapses the probabilities of all markets examined—and also for different time horizons—into single curves, suggesting that first-passage statistics is market independent (at least for high-frequency data). On the other hand, a very closely related quantity, the survival probability, shows, away from the center and tails of the distribution, a hyperbolic t-1/2 decay typical of a Markovian dynamics, albeit the existence of memory in markets. Modifications of the Weibull and Student distributions are good candidates for the phenomenological description of first-passage time properties under certain regimes. The scaling strategies shown may be useful for risk control and algorithmic trading.
On the detection of very high redshift gamma-ray bursts with Swift
NASA Astrophysics Data System (ADS)
Salvaterra, R.; Campana, S.; Chincarini, G.; Tagliaferri, G.; Covino, S.
2007-09-01
We compute the probability of detecting long gamma-ray bursts (GRBs) at z >= 5 with Swift, assuming that GRBs form preferentially in low-metallicity environments. The model fits both the observed Burst and Transient Source Experiment (BATSE) and Swift GRB differential peak flux distributions well and is consistent with the number of z >= 2.5 detections in the 2-yr Swift data. We find that the probability of observing a burst at z >= 5 becomes larger than 10 per cent for photon fluxes P < 1 ph s-1 cm-2, consistent with the number of confirmed detections. The corresponding fraction of z >= 5 bursts in the Swift catalogue is ~10-30 per cent depending on the adopted metallicity threshold for GRB formation. We propose to use the computed probability as a tool to identify high-redshift GRBs. By jointly considering promptly available information provided by Swift and model results, we can select reliable z >= 5 candidates in a few hours from the BAT detection. We test the procedure against last year Swift data: only three bursts match all our requirements, two being confirmed at z >= 5. Another three possible candidates are picked up by slightly relaxing the adopted criteria. No low-z interloper is found among the six candidates.
75 FR 53925 - Sea Turtle Conservation; Shrimp and Summer Flounder Trawling Requirements
Federal Register 2010, 2011, 2012, 2013, 2014
2010-09-02
... because of the statistical probability the candidate TED may not achieve the standard (i.e., control TED... the test with 4 turtle captures because of the statistical probability the candidate TED may not... because of the statistical probability the candidate TED may not achieve the standard (i.e., [[Page 53930...
PASTIS: Bayesian extrasolar planet validation - I. General framework, models, and performance
NASA Astrophysics Data System (ADS)
Díaz, R. F.; Almenara, J. M.; Santerne, A.; Moutou, C.; Lethuillier, A.; Deleuil, M.
2014-06-01
A large fraction of the smallest transiting planet candidates discovered by the Kepler and CoRoT space missions cannot be confirmed by a dynamical measurement of the mass using currently available observing facilities. To establish their planetary nature, the concept of planet validation has been advanced. This technique compares the probability of the planetary hypothesis against that of all reasonably conceivable alternative false positive (FP) hypotheses. The candidate is considered as validated if the posterior probability of the planetary hypothesis is sufficiently larger than the sum of the probabilities of all FP scenarios. In this paper, we present PASTIS, the Planet Analysis and Small Transit Investigation Software, a tool designed to perform a rigorous model comparison of the hypotheses involved in the problem of planet validation, and to fully exploit the information available in the candidate light curves. PASTIS self-consistently models the transit light curves and follow-up observations. Its object-oriented structure offers a large flexibility for defining the scenarios to be compared. The performance is explored using artificial transit light curves of planets and FPs with a realistic error distribution obtained from a Kepler light curve. We find that data support the correct hypothesis strongly only when the signal is high enough (transit signal-to-noise ratio above 50 for the planet case) and remain inconclusive otherwise. PLAnetary Transits and Oscillations of stars (PLATO) shall provide transits with high enough signal-to-noise ratio, but to establish the true nature of the vast majority of Kepler and CoRoT transit candidates additional data or strong reliance on hypotheses priors is needed.
NASA Astrophysics Data System (ADS)
Lee, Jaeha; Tsutsui, Izumi
2017-05-01
We show that the joint behavior of an arbitrary pair of (generally noncommuting) quantum observables can be described by quasi-probabilities, which are an extended version of the standard probabilities used for describing the outcome of measurement for a single observable. The physical situations that require these quasi-probabilities arise when one considers quantum measurement of an observable conditioned by some other variable, with the notable example being the weak measurement employed to obtain Aharonov's weak value. Specifically, we present a general prescription for the construction of quasi-joint probability (QJP) distributions associated with a given combination of observables. These QJP distributions are introduced in two complementary approaches: one from a bottom-up, strictly operational construction realized by examining the mathematical framework of the conditioned measurement scheme, and the other from a top-down viewpoint realized by applying the results of the spectral theorem for normal operators and their Fourier transforms. It is then revealed that, for a pair of simultaneously measurable observables, the QJP distribution reduces to the unique standard joint probability distribution of the pair, whereas for a noncommuting pair there exists an inherent indefiniteness in the choice of such QJP distributions, admitting a multitude of candidates that may equally be used for describing the joint behavior of the pair. In the course of our argument, we find that the QJP distributions furnish the space of operators in the underlying Hilbert space with their characteristic geometric structures such that the orthogonal projections and inner products of observables can be given statistical interpretations as, respectively, “conditionings” and “correlations”. The weak value Aw for an observable A is then given a geometric/statistical interpretation as either the orthogonal projection of A onto the subspace generated by another observable B, or equivalently, as the conditioning of A given B with respect to the QJP distribution under consideration.
Why Are Product Prices in Online Markets Not Converging?
Mizuno, Takayuki; Watanabe, Tsutomu
2013-01-01
Why are product prices in online markets dispersed in spite of very small search costs? To address this question, we construct a unique dataset from a Japanese price comparison site, which records price quotes offered by e-retailers as well as customers’ clicks on products, which occur when they proceed to purchase the product. The novelty of our approach is that we seek to extract useful information on the source of price dispersion from the shape of price distributions rather than focusing merely on the standard deviation or the coefficient of variation of prices, as previous studies have done. We find that the distribution of prices retailers quote for a particular product at a particular point in time (divided by the lowest price) follows an exponential distribution, showing the presence of substantial price dispersion. For example, 20 percent of all retailers quote prices that are more than 50 percent higher than the lowest price. Next, comparing the probability that customers click on a retailer with a particular rank and the probability that retailers post prices at a particular rank, we show that both decline exponentially with price rank and that the exponents associated with the probabilities are quite close. This suggests that the reason why some retailers set prices at a level substantially higher than the lowest price is that they know that some customers will choose them even at that high price. Based on these findings, we hypothesize that price dispersion in online markets stems from heterogeneity in customers’ preferences over retailers; that is, customers choose a set of candidate retailers based on their preferences, which are heterogeneous across customers, and then pick a particular retailer among the candidates based on the price ranking. PMID:24015219
Why are product prices in online markets not converging?
Mizuno, Takayuki; Watanabe, Tsutomu
2013-01-01
Why are product prices in online markets dispersed in spite of very small search costs? To address this question, we construct a unique dataset from a Japanese price comparison site, which records price quotes offered by e-retailers as well as customers' clicks on products, which occur when they proceed to purchase the product. The novelty of our approach is that we seek to extract useful information on the source of price dispersion from the shape of price distributions rather than focusing merely on the standard deviation or the coefficient of variation of prices, as previous studies have done. We find that the distribution of prices retailers quote for a particular product at a particular point in time (divided by the lowest price) follows an exponential distribution, showing the presence of substantial price dispersion. For example, 20 percent of all retailers quote prices that are more than 50 percent higher than the lowest price. Next, comparing the probability that customers click on a retailer with a particular rank and the probability that retailers post prices at a particular rank, we show that both decline exponentially with price rank and that the exponents associated with the probabilities are quite close. This suggests that the reason why some retailers set prices at a level substantially higher than the lowest price is that they know that some customers will choose them even at that high price. Based on these findings, we hypothesize that price dispersion in online markets stems from heterogeneity in customers' preferences over retailers; that is, customers choose a set of candidate retailers based on their preferences, which are heterogeneous across customers, and then pick a particular retailer among the candidates based on the price ranking.
Dietz, Dennis C.
2014-01-01
A cogent method is presented for computing the expected cost of an appointment schedule where customers are statistically identical, the service time distribution has known mean and variance, and customer no-shows occur with time-dependent probability. The approach is computationally efficient and can be easily implemented to evaluate candidate schedules within a schedule optimization algorithm. PMID:24605070
Improved Membership Probability for Moving Groups: Bayesian and Machine Learning Approaches
NASA Astrophysics Data System (ADS)
Lee, Jinhee; Song, Inseok
2018-01-01
Gravitationally unbound loose stellar associations (i.e., young nearby moving groups: moving groups hereafter) have been intensively explored because they are important in planet and disk formation studies, exoplanet imaging, and age calibration. Among the many efforts devoted to the search for moving group members, a Bayesian approach (e.g.,using the code BANYAN) has become popular recently because of the many advantages it offers. However, the resultant membership probability needs to be carefully adopted because of its sensitive dependence on input models. In this study, we have developed an improved membership calculation tool focusing on the beta-Pic moving group. We made three improvements for building models used in BANYAN II: (1) updating a list of accepted members by re-assessing memberships in terms of position, motion, and age, (2) investigating member distribution functions in XYZ, and (3) exploring field star distribution functions in XYZUVW. Our improved tool can change membership probability up to 70%. Membership probability is critical and must be better defined. For example, our code identifies only one third of the candidate members in SIMBAD that are believed to be kinematically associated with beta-Pic moving group.Additionally, we performed cluster analysis of young nearby stars using an unsupervised machine learning approach. As more moving groups and their members are identified, the complexity and ambiguity in moving group configuration has been increased. To clarify this issue, we analyzed ~4,000 X-ray bright young stellar candidates. Here, we present the preliminary results. By re-identifying moving groups with the least human intervention, we expect to understand the composition of the solar neighborhood. Moreover better defined moving group membership will help us understand star formation and evolution in relatively low density environments; especially for the low-mass stars which will be identified in the coming Gaia release.
A hierarchical two-phase framework for selecting genes in cancer datasets with a neuro-fuzzy system.
Lim, Jongwoo; Wang, Bohyun; Lim, Joon S
2016-04-29
Finding the minimum number of appropriate biomarkers for specific targets such as a lung cancer has been a challenging issue in bioinformatics. We propose a hierarchical two-phase framework for selecting appropriate biomarkers that extracts candidate biomarkers from the cancer microarray datasets and then selects the minimum number of appropriate biomarkers from the extracted candidate biomarkers datasets with a specific neuro-fuzzy algorithm, which is called a neural network with weighted fuzzy membership function (NEWFM). In this context, as the first phase, the proposed framework is to extract candidate biomarkers by using a Bhattacharyya distance method that measures the similarity of two discrete probability distributions. Finally, the proposed framework is able to reduce the cost of finding biomarkers by not receiving medical supplements and improve the accuracy of the biomarkers in specific cancer target datasets.
SAS procedures for designing and analyzing sample surveys
Stafford, Joshua D.; Reinecke, Kenneth J.; Kaminski, Richard M.
2003-01-01
Complex surveys often are necessary to estimate occurrence (or distribution), density, and abundance of plants and animals for purposes of re-search and conservation. Most scientists are familiar with simple random sampling, where sample units are selected from a population of interest (sampling frame) with equal probability. However, the goal of ecological surveys often is to make inferences about populations over large or complex spatial areas where organisms are not homogeneously distributed or sampling frames are in-convenient or impossible to construct. Candidate sampling strategies for such complex surveys include stratified,multistage, and adaptive sampling (Thompson 1992, Buckland 1994).
275 Candidates and 149 Validated Planets Orbiting Bright Stars in K2 Campaigns 0–10
NASA Astrophysics Data System (ADS)
Mayo, Andrew W.; Vanderburg, Andrew; Latham, David W.; Bieryla, Allyson; Morton, Timothy D.; Buchhave, Lars A.; Dressing, Courtney D.; Beichman, Charles; Berlind, Perry; Calkins, Michael L.; Ciardi, David R.; Crossfield, Ian J. M.; Esquerdo, Gilbert A.; Everett, Mark E.; Gonzales, Erica J.; Hirsch, Lea A.; Horch, Elliott P.; Howard, Andrew W.; Howell, Steve B.; Livingston, John; Patel, Rahul; Petigura, Erik A.; Schlieder, Joshua E.; Scott, Nicholas J.; Schumer, Clea F.; Sinukoff, Evan; Teske, Johanna; Winters, Jennifer G.
2018-03-01
Since 2014, NASA’s K2 mission has observed large portions of the ecliptic plane in search of transiting planets and has detected hundreds of planet candidates. With observations planned until at least early 2018, K2 will continue to identify more planet candidates. We present here 275 planet candidates observed during Campaigns 0–10 of the K2 mission that are orbiting stars brighter than 13 mag (in Kepler band) and for which we have obtained high-resolution spectra (R = 44,000). These candidates are analyzed using the vespa package in order to calculate their false-positive probabilities (FPP). We find that 149 candidates are validated with an FPP lower than 0.1%, 39 of which were previously only candidates and 56 of which were previously undetected. The processes of data reduction, candidate identification, and statistical validation are described, and the demographics of the candidates and newly validated planets are explored. We show tentative evidence of a gap in the planet radius distribution of our candidate sample. Comparing our sample to the Kepler candidate sample investigated by Fulton et al., we conclude that more planets are required to quantitatively confirm the gap with K2 candidates or validated planets. This work, in addition to increasing the population of validated K2 planets by nearly 50% and providing new targets for follow-up observations, will also serve as a framework for validating candidates from upcoming K2 campaigns and the Transiting Exoplanet Survey Satellite, expected to launch in 2018.
Unsupervised Spatio-Temporal Data Mining Framework for Burned Area Mapping
NASA Technical Reports Server (NTRS)
Kumar, Vipin (Inventor); Boriah, Shyam (Inventor); Mithal, Varun (Inventor); Khandelwal, Ankush (Inventor)
2016-01-01
A method reduces processing time required to identify locations burned by fire by receiving a feature value for each pixel in an image, each pixel representing a sub-area of a location. Pixels are then grouped based on similarities of the feature values to form candidate burn events. For each candidate burn event, a probability that the candidate burn event is a true burn event is determined based on at least one further feature value for each pixel in the candidate burn event. Candidate burn events that have a probability below a threshold are removed from further consideration as burn events to produce a set of remaining candidate burn events.
Code of Federal Regulations, 2010 CFR
2010-04-01
... beginining study under the programs for which applications are made. (b) Candidates for the fellowship... studies under the program. (c) Candidates shall submit evidence of acceptance, or probable acceptance, for study in programs that will enhance their contributions to their employers. Evidence of probable...
Chrzanowski, Frank
2008-01-01
Two numerical methods, Decision Analysis (DA) and Potential Problem Analysis (PPA) are presented as alternative selection methods to the logical method presented in Part I. In DA properties are weighted and outcomes are scored. The weighted scores for each candidate are totaled and final selection is based on the totals. Higher scores indicate better candidates. In PPA potential problems are assigned a seriousness factor and test outcomes are used to define the probability of occurrence. The seriousness-probability products are totaled and forms with minimal scores are preferred. DA and PPA have never been compared to the logical-elimination method. Additional data were available for two forms of McN-5707 to provide complete preformulation data for five candidate forms. Weight and seriousness factors (independent variables) were obtained from a survey of experienced formulators. Scores and probabilities (dependent variables) were provided independently by Preformulation. The rankings of the five candidate forms, best to worst, were similar for all three methods. These results validate the applicability of DA and PPA for candidate form selection. DA and PPA are particularly applicable in cases where there are many candidate forms and where each form has some degree of unfavorable properties.
Som, Nicholas A.; Goodman, Damon H.; Perry, Russell W.; Hardy, Thomas B.
2016-01-01
Previous methods for constructing univariate habitat suitability criteria (HSC) curves have ranged from professional judgement to kernel-smoothed density functions or combinations thereof. We present a new method of generating HSC curves that applies probability density functions as the mathematical representation of the curves. Compared with previous approaches, benefits of our method include (1) estimation of probability density function parameters directly from raw data, (2) quantitative methods for selecting among several candidate probability density functions, and (3) concise methods for expressing estimation uncertainty in the HSC curves. We demonstrate our method with a thorough example using data collected on the depth of water used by juvenile Chinook salmon (Oncorhynchus tschawytscha) in the Klamath River of northern California and southern Oregon. All R code needed to implement our example is provided in the appendix. Published 2015. This article is a U.S. Government work and is in the public domain in the USA.
YSO jets in the Galactic plane from UWISH2 - V. Jets and outflows in M17
NASA Astrophysics Data System (ADS)
Samal, M. R.; Chen, W. P.; Takami, M.; Jose, J.; Froebrich, D.
2018-07-01
Jets and outflows are the first signposts of stellar birth. Emission in the H2 1-0 S(1) line at 2.122- μm is a powerful tracer of shock excitation in these objects. Here we present the analysis of 2.0 × 0.8 deg2 data from the UK Wide-field Infrared Survey for H2 (UWISH2) in the 1-0 S(1) line to identify and characterize the outflows of the M17 complex. We uncover 48 probable outflows, of which 93 per cent are new discoveries. We identified driving source candidates for 60 per cent of outflows. Among the driving source candidate young stellar objects (YSOs), 90 per cent are protostars and the remaining 10 per cent are Class II YSOs. In comparison with results from other surveys, we suggest that H2 emission fades very quickly as the objects evolve from protostars to pre-main-sequence stars. We fit spectral energy distribution (SED) models to 14 candidate outflow-driving sources and conclude that the outflows of our sample are mostly driven by moderate-mass YSOs that are still actively accreting from their protoplanetary disc. We examined the spatial distribution of the outflows with the gas and dust distribution of the complex and observed that the filamentary dark cloud M17SWex, located on the south-western side of the complex, is associated with a greater number of outflows. We find that our results corroborate previous suggestions that, in the M17 complex, M17SWex is the most active site of star formation. Several of our newly identified outflow candidates are excellent targets for follow-up studies to understand better the very early phase of protostellar evolution.
Biological monitoring of environmental quality: The use of developmental instability
Freeman, D.C.; Emlen, J.M.; Graham, J.H.; Hough, R. A.; Bannon, T.A.
1994-01-01
Distributed robustness is thought to influence the buffering of random phenotypic variation through the scale-free topology of gene regulatory, metabolic, and protein-protein interaction networks. If this hypothesis is true, then the phenotypic response to the perturbation of particular nodes in such a network should be proportional to the number of links those nodes make with neighboring nodes. This suggests a probability distribution approximating an inverse power-law of random phenotypic variation. Zero phenotypic variation, however, is impossible, because random molecular and cellular processes are essential to normal development. Consequently, a more realistic distribution should have a y-intercept close to zero in the lower tail, a mode greater than zero, and a long (fat) upper tail. The double Pareto-lognormal (DPLN) distribution is an ideal candidate distribution. It consists of a mixture of a lognormal body and upper and lower power-law tails.
I show that a conditional probability analysis using a stressor-response model based on a logistic regression provides a useful approach for developing candidate water quality criteria from empirical data, such as the Maryland Biological Streams Survey (MBSS) data.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gagné, Jonathan; Lafrenière, David; Doyon, René
We present Bayesian Analysis for Nearby Young AssociatioNs II (BANYAN II), a modified Bayesian analysis for assessing the membership of later-than-M5 objects to any of several Nearby Young Associations (NYAs). In addition to using kinematic information (from sky position and proper motion), this analysis exploits 2MASS-WISE color-magnitude diagrams in which old and young objects follow distinct sequences. As an improvement over our earlier work, the spatial and kinematic distributions for each association are now modeled as ellipsoids whose axes need not be aligned with the Galactic coordinate axes, and we use prior probabilities matching the expected populations of the NYAsmore » considered versus field stars. We present an extensive contamination analysis to characterize the performance of our new method. We find that Bayesian probabilities are generally representative of contamination rates, except when a parallax measurement is considered. In this case contamination rates become significantly smaller and hence Bayesian probabilities for NYA memberships are pessimistic. We apply this new algorithm to a sample of 158 objects from the literature that are either known to display spectroscopic signs of youth or have unusually red near-infrared colors for their spectral type. Based on our analysis, we identify 25 objects as new highly probable candidates to NYAs, including a new M7.5 bona fide member to Tucana-Horologium, making it the latest-type member. In addition, we reveal that a known L2γ dwarf is co-moving with a bright M5 dwarf, and we show for the first time that two of the currently known ultra red L dwarfs are strong candidates to the AB Doradus moving group. Several objects identified here as highly probable members to NYAs could be free-floating planetary-mass objects if their membership is confirmed.« less
Robotic Mars Sample Return: Risk Assessment and Analysis Report
NASA Technical Reports Server (NTRS)
Lalk, Thomas R.; Spence, Cliff A.
2003-01-01
A comparison of the risk associated with two alternative scenarios for a robotic Mars sample return mission was conducted. Two alternative mission scenarios were identified, the Jet Propulsion Lab (JPL) reference Mission and a mission proposed by Johnson Space Center (JSC). The JPL mission was characterized by two landers and an orbiter, and a Mars orbit rendezvous to retrieve the samples. The JSC mission (Direct/SEP) involves a solar electric propulsion (SEP) return to earth followed by a rendezvous with the space shuttle in earth orbit. A qualitative risk assessment to identify and characterize the risks, and a risk analysis to quantify the risks were conducted on these missions. Technical descriptions of the competing scenarios were developed in conjunction with NASA engineers and the sequence of events for each candidate mission was developed. Risk distributions associated with individual and combinations of events were consolidated using event tree analysis in conjunction with Monte Carlo techniques to develop probabilities of mission success for each of the various alternatives. The results were the probability of success of various end states for each candidate scenario. These end states ranged from complete success through various levels of partial success to complete failure. Overall probability of success for the Direct/SEP mission was determined to be 66% for the return of at least one sample and 58% for the JPL mission for the return of at least one sample cache. Values were also determined for intermediate events and end states as well as for the probability of violation of planetary protection. Overall mission planetary protection event probabilities of occurrence were determined to be 0.002% and 1.3% for the Direct/SEP and JPL Reference missions respectively.
ALMA’s Polarized View of 10 Protostars in the Perseus Molecular Cloud
NASA Astrophysics Data System (ADS)
Cox, Erin G.; Harris, Robert J.; Looney, Leslie W.; Li, Zhi-Yun; Yang, Haifeng; Tobin, John J.; Stephens, Ian
2018-03-01
We present 870 μm ALMA dust polarization observations of 10 young Class 0/I protostars in the Perseus Molecular Cloud. At ∼0.″35 (80 au) resolution, all of our sources show some degree of polarization, with most (9/10) showing significantly extended emission in the polarized continuum. Each source has incredibly intricate polarization signatures. In particular, all three disk-candidates have polarization vectors roughly along the minor axis, which is indicative of polarization produced by dust scattering. On ∼100 au scales, the polarization is at a relatively low level (≲1%) and is quite ordered. In sources with significant envelope emission, the envelope is typically polarized at a much higher (≳5%) level and has a far more disordered morphology. We compute the cumulative probability distributions for both the small (disk-scale) and large (envelope-scale) polarization percentage. We find that the two are intrinsically different, even after accounting for the different detection thresholds in the high/low surface brightness regions. We perform Kolmogorov–Smirnov and Anderson–Darling tests on the distributions of angle offsets of the polarization from the outflow axis. We find disk-candidate sources are different from the non-disk-candidate sources. We conclude that the polarization on the 100 au scale is consistent with the signature of dust scattering for disk-candidates and that the polarization on the envelope-scale in all sources may come from another mechanism, most likely magnetically aligned grains.
We show that a conditional probability analysis that utilizes a stressor-response model based on a logistic regression provides a useful approach for developing candidate water quality criterai from empirical data. The critical step in this approach is transforming the response ...
Supervised Detection of Anomalous Light Curves in Massive Astronomical Catalogs
NASA Astrophysics Data System (ADS)
Nun, Isadora; Pichara, Karim; Protopapas, Pavlos; Kim, Dae-Won
2014-09-01
The development of synoptic sky surveys has led to a massive amount of data for which resources needed for analysis are beyond human capabilities. In order to process this information and to extract all possible knowledge, machine learning techniques become necessary. Here we present a new methodology to automatically discover unknown variable objects in large astronomical catalogs. With the aim of taking full advantage of all information we have about known objects, our method is based on a supervised algorithm. In particular, we train a random forest classifier using known variability classes of objects and obtain votes for each of the objects in the training set. We then model this voting distribution with a Bayesian network and obtain the joint voting distribution among the training objects. Consequently, an unknown object is considered as an outlier insofar it has a low joint probability. By leaving out one of the classes on the training set, we perform a validity test and show that when the random forest classifier attempts to classify unknown light curves (the class left out), it votes with an unusual distribution among the classes. This rare voting is detected by the Bayesian network and expressed as a low joint probability. Our method is suitable for exploring massive data sets given that the training process is performed offline. We tested our algorithm on 20 million light curves from the MACHO catalog and generated a list of anomalous candidates. After analysis, we divided the candidates into two main classes of outliers: artifacts and intrinsic outliers. Artifacts were principally due to air mass variation, seasonal variation, bad calibration, or instrumental errors and were consequently removed from our outlier list and added to the training set. After retraining, we selected about 4000 objects, which we passed to a post-analysis stage by performing a cross-match with all publicly available catalogs. Within these candidates we identified certain known but rare objects such as eclipsing Cepheids, blue variables, cataclysmic variables, and X-ray sources. For some outliers there was no additional information. Among them we identified three unknown variability types and a few individual outliers that will be followed up in order to perform a deeper analysis.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nun, Isadora; Pichara, Karim; Protopapas, Pavlos
The development of synoptic sky surveys has led to a massive amount of data for which resources needed for analysis are beyond human capabilities. In order to process this information and to extract all possible knowledge, machine learning techniques become necessary. Here we present a new methodology to automatically discover unknown variable objects in large astronomical catalogs. With the aim of taking full advantage of all information we have about known objects, our method is based on a supervised algorithm. In particular, we train a random forest classifier using known variability classes of objects and obtain votes for each ofmore » the objects in the training set. We then model this voting distribution with a Bayesian network and obtain the joint voting distribution among the training objects. Consequently, an unknown object is considered as an outlier insofar it has a low joint probability. By leaving out one of the classes on the training set, we perform a validity test and show that when the random forest classifier attempts to classify unknown light curves (the class left out), it votes with an unusual distribution among the classes. This rare voting is detected by the Bayesian network and expressed as a low joint probability. Our method is suitable for exploring massive data sets given that the training process is performed offline. We tested our algorithm on 20 million light curves from the MACHO catalog and generated a list of anomalous candidates. After analysis, we divided the candidates into two main classes of outliers: artifacts and intrinsic outliers. Artifacts were principally due to air mass variation, seasonal variation, bad calibration, or instrumental errors and were consequently removed from our outlier list and added to the training set. After retraining, we selected about 4000 objects, which we passed to a post-analysis stage by performing a cross-match with all publicly available catalogs. Within these candidates we identified certain known but rare objects such as eclipsing Cepheids, blue variables, cataclysmic variables, and X-ray sources. For some outliers there was no additional information. Among them we identified three unknown variability types and a few individual outliers that will be followed up in order to perform a deeper analysis.« less
Methodology for finding and evaluating safe landing sites on small bodies
NASA Astrophysics Data System (ADS)
Rodgers, Douglas J.; Ernst, Carolyn M.; Barnouin, Olivier S.; Murchie, Scott L.; Chabot, Nancy L.
2016-12-01
Here we develop and demonstrate a three-step strategy for finding a safe landing ellipse for a legged spacecraft on a small body such as an asteroid or planetary satellite. The first step, acquisition of a high-resolution terrain model of a candidate landing region, is simulated using existing statistics on block abundances measured at Phobos, Eros, and Itokawa. The synthetic terrain model is generated by randomly placing hemispheric shaped blocks with the empirically determined size-frequency distribution. The resulting terrain is much rockier than typical lunar or martian landing sites. The second step, locating a landing ellipse with minimal hazards, is demonstrated for an assumed approach to landing that uses Autonomous Landing and Hazard Avoidance Technology. The final step, determination of the probability distribution for orientation of the landed spacecraft, is demonstrated for cases of differing regional slope. The strategy described here is both a prototype for finding a landing site during a flight mission and provides tools for evaluating the design of small-body landers. We show that for bodies with Eros-like block distributions, there may be >99% probability of landing stably at a low tilt without blocks impinging on spacecraft structures so as to pose a survival hazard.
Newly discovered Wolf-Rayet and weak emission-line central stars of planetary nebulae
NASA Astrophysics Data System (ADS)
DePew, K.; Parker, Q. A.; Miszalski, B.; De Marco, O.; Frew, D. J.; Acker, A.; Kovacevic, A. V.; Sharp, R. G.
2011-07-01
We present the spectra of 32 previously unpublished confirmed and candidate Wolf-Rayet ([WR]) and weak emission-line (WELS) central stars of planetary nebulae (CSPNe). 18 stars have been discovered in the Macquarie/AAO/Strasbourg Hα (MASH) PN survey sample, and we have also uncovered 14 confirmed and candidate [WR]s and WELS among the CSPNe of previously known PNe. Spectral classifications have been undertaken using both Acker & Neiner and Crowther, De Marco & Barlow schemes. 22 members in this sample are identified as probable [WR]s; the remaining 10 appear to be WELS. Observations undertaken as part of the MASH spectroscopic survey have now increased the number of known [WR]s by ˜30 per cent. This will permit a better analysis of [WR] subclass distribution, metallicity effects and evolutionary sequences in these uncommon objects.
The impacts of precipitation amount simulation on hydrological modeling in Nordic watersheds
NASA Astrophysics Data System (ADS)
Li, Zhi; Brissette, Fancois; Chen, Jie
2013-04-01
Stochastic modeling of daily precipitation is very important for hydrological modeling, especially when no observed data are available. Precipitation is usually modeled by two component model: occurrence generation and amount simulation. For occurrence simulation, the most common method is the first-order two-state Markov chain due to its simplification and good performance. However, various probability distributions have been reported to simulate precipitation amount, and spatiotemporal differences exist in the applicability of different distribution models. Therefore, assessing the applicability of different distribution models is necessary in order to provide more accurate precipitation information. Six precipitation probability distributions (exponential, Gamma, Weibull, skewed normal, mixed exponential, and hybrid exponential/Pareto distributions) are directly and indirectly evaluated on their ability to reproduce the original observed time series of precipitation amount. Data from 24 weather stations and two watersheds (Chute-du-Diable and Yamaska watersheds) in the province of Quebec (Canada) are used for this assessment. Various indices or statistics, such as the mean, variance, frequency distribution and extreme values are used to quantify the performance in simulating the precipitation and discharge. Performance in reproducing key statistics of the precipitation time series is well correlated to the number of parameters of the distribution function, and the three-parameter precipitation models outperform the other models, with the mixed exponential distribution being the best at simulating daily precipitation. The advantage of using more complex precipitation distributions is not as clear-cut when the simulated time series are used to drive a hydrological model. While the advantage of using functions with more parameters is not nearly as obvious, the mixed exponential distribution appears nonetheless as the best candidate for hydrological modeling. The implications of choosing a distribution function with respect to hydrological modeling and climate change impact studies are also discussed.
NASA Technical Reports Server (NTRS)
Schwartz, H.-J.
1976-01-01
The modeling process of a complex system, based on the calculation and optimization of the system parameters, is complicated in that some parameters can be expressed only as probability distributions. In the present paper, a Monte Carlo technique was used to determine the daily range requirements of an electric road vehicle in the United States from probability distributions of trip lengths, frequencies, and average annual mileage data. The analysis shows that a daily range of 82 miles meets to 95% of the car-owner requirements at all times with the exception of long vacation trips. Further, it is shown that the requirement of a daily range of 82 miles can be met by a (intermediate-level) battery technology characterized by an energy density of 30 to 50 Watt-hours per pound. Candidate batteries in this class are nickel-zinc, nickel-iron, and iron-air. These results imply that long-term research goals for battery systems should be focused on lower cost and longer service life, rather than on higher energy densities
Bayesian Inference on Proportional Elections
Brunello, Gabriel Hideki Vatanabe; Nakano, Eduardo Yoshio
2015-01-01
Polls for majoritarian voting systems usually show estimates of the percentage of votes for each candidate. However, proportional vote systems do not necessarily guarantee the candidate with the most percentage of votes will be elected. Thus, traditional methods used in majoritarian elections cannot be applied on proportional elections. In this context, the purpose of this paper was to perform a Bayesian inference on proportional elections considering the Brazilian system of seats distribution. More specifically, a methodology to answer the probability that a given party will have representation on the chamber of deputies was developed. Inferences were made on a Bayesian scenario using the Monte Carlo simulation technique, and the developed methodology was applied on data from the Brazilian elections for Members of the Legislative Assembly and Federal Chamber of Deputies in 2010. A performance rate was also presented to evaluate the efficiency of the methodology. Calculations and simulations were carried out using the free R statistical software. PMID:25786259
DOE Office of Scientific and Technical Information (OSTI.GOV)
Titov, O.; Jauncey, D. L.; Johnston, H. M.
2011-11-15
We present the results of spectroscopic observations of the optical counterparts of 47 southern radio sources from the candidate International Celestial Reference Catalogue as part of a very long baseline interferometry (VLBI) program to strengthen the celestial reference frame, especially in the south. We made the observations with the 3.58 m European Southern Observatory New Technology Telescope. We obtained redshifts for 30 quasars and one radio galaxy, with a further seven objects being probable BL Lac objects with featureless spectra. Of the remainder, four were clear misidentifications with Galactic stars and five had low signal-to-noise spectra and could not bemore » classified. These results, in combination with new VLBI data of the radio sources with redshifts more than 2, add significantly to the existing data needed to refine the distribution of source proper motions over the celestial sphere.« less
Bayesian inference on proportional elections.
Brunello, Gabriel Hideki Vatanabe; Nakano, Eduardo Yoshio
2015-01-01
Polls for majoritarian voting systems usually show estimates of the percentage of votes for each candidate. However, proportional vote systems do not necessarily guarantee the candidate with the most percentage of votes will be elected. Thus, traditional methods used in majoritarian elections cannot be applied on proportional elections. In this context, the purpose of this paper was to perform a Bayesian inference on proportional elections considering the Brazilian system of seats distribution. More specifically, a methodology to answer the probability that a given party will have representation on the chamber of deputies was developed. Inferences were made on a Bayesian scenario using the Monte Carlo simulation technique, and the developed methodology was applied on data from the Brazilian elections for Members of the Legislative Assembly and Federal Chamber of Deputies in 2010. A performance rate was also presented to evaluate the efficiency of the methodology. Calculations and simulations were carried out using the free R statistical software.
Estimation of submarine mass failure probability from a sequence of deposits with age dates
Geist, Eric L.; Chaytor, Jason D.; Parsons, Thomas E.; ten Brink, Uri S.
2013-01-01
The empirical probability of submarine mass failure is quantified from a sequence of dated mass-transport deposits. Several different techniques are described to estimate the parameters for a suite of candidate probability models. The techniques, previously developed for analyzing paleoseismic data, include maximum likelihood and Type II (Bayesian) maximum likelihood methods derived from renewal process theory and Monte Carlo methods. The estimated mean return time from these methods, unlike estimates from a simple arithmetic mean of the center age dates and standard likelihood methods, includes the effects of age-dating uncertainty and of open time intervals before the first and after the last event. The likelihood techniques are evaluated using Akaike’s Information Criterion (AIC) and Akaike’s Bayesian Information Criterion (ABIC) to select the optimal model. The techniques are applied to mass transport deposits recorded in two Integrated Ocean Drilling Program (IODP) drill sites located in the Ursa Basin, northern Gulf of Mexico. Dates of the deposits were constrained by regional bio- and magnetostratigraphy from a previous study. Results of the analysis indicate that submarine mass failures in this location occur primarily according to a Poisson process in which failures are independent and return times follow an exponential distribution. However, some of the model results suggest that submarine mass failures may occur quasiperiodically at one of the sites (U1324). The suite of techniques described in this study provides quantitative probability estimates of submarine mass failure occurrence, for any number of deposits and age uncertainty distributions.
An improved catalog of halo wide binary candidates
DOE Office of Scientific and Technical Information (OSTI.GOV)
Allen, Christine; Monroy-Rodríguez, Miguel A., E-mail: chris@astro.unam.mx
2014-08-01
We present an improved catalog of halo wide binaries compiled from an extensive literature search. Most of our binaries stem from the common proper motion binary catalogs by Allen et al. and Chanamé and Gould, but we have also included binaries from the lists of Ryan and Zapatero-Osorio and Martín. All binaries were carefully checked and their distances and systemic radial velocities are included when available. Probable membership to the halo population was tested by means of reduced proper motion diagrams for 251 candidate halo binaries. After eliminating obvious disk binaries, we ended up with 211 probable halo binaries, 150more » of which have radial velocities available. We compute galactic orbits for these 150 binaries and calculate the time they spend within the galactic disk. Considering the full sample of 251 candidate halo binaries as well as several subsamples, we find that the distribution of angular separations (or expected major semiaxes) follows a power law f(a) ∼ a {sup –1} (Oepik's relation) up to different limits. For the 50 most disk-like binaries, those that spend their entire lives within z = ±500 pc, this limit is found to be 19,000 AU (0.09 pc), while for the 50 most halo-like binaries, those that spend on average only 18% of their lives within z = ±500 pc, the limit is 63,000 AU (0.31 pc). In a companion paper, we employ this catalog to establish limits on the masses of the halo massive perturbers (massive compact halo objects).« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Abulencia, A.; Acosta, D.; Adelman, Jahred A.
2006-02-01
The authors describe a measurement of the top quark mass from events produced in p{bar p} collisions at a center-of-mass energy of 1.96 TeV, using the Collider Detector at Fermilab. They identify t{bar t} candidates where both W bosons from the top quarks decay into leptons (e{nu}, {mu}{nu}, or {tau}{nu}) from a data sample of 360 pb{sup -1}. The top quark mass is reconstructed in each event separately by three different methods, which draw upon simulated distributions of the neutrino pseudorapidity, t{bar t} longitudinal momentum, or neutrino azimuthal angle in order to extract probability distributions for the top quark mass.more » For each method, representative mass distributions, or templates, are constructed from simulated samples of signal and background events, and parameterized to form continuous probability density functions. A likelihood fit incorporating these parameterized templates is then performed on the data sample masses in order to derive a final top quark mass. Combining the three template methods, taking into account correlations in their statistical and systematic uncertainties, results in a top quark mass measurement of 170.1 {+-} 6.0(stat.) {+-} 4.1(syst.) GeV/c{sup 2}.« less
Discovery of a large-scale clumpy structure around the Lynx supercluster at z~ 1.27
NASA Astrophysics Data System (ADS)
Nakata, Fumiaki; Kodama, Tadayuki; Shimasaku, Kazuhiro; Doi, Mamoru; Furusawa, Hisanori; Hamabe, Masaru; Kimura, Masahiko; Komiyama, Yutaka; Miyazaki, Satoshi; Okamura, Sadanori; Ouchi, Masami; Sekiguchi, Maki; Ueda, Yoshihiro; Yagi, Masafumi; Yasuda, Naoki
2005-03-01
We report the discovery of a probable large-scale structure composed of many galaxy clumps around the known twin clusters at z= 1.26 and 1.27 in the Lynx region. Our analysis is based on deep, panoramic, and multicolour imaging, 26.4 × 24.1 arcmin2 in VRi'z' bands with the Suprime-Cam on the 8.2-m Subaru telescope. This unique, deep and wide-field imaging data set allows us for the first time to map out the galaxy distribution in the highest-redshift supercluster known. We apply a photometric redshift technique to extract plausible cluster members at z~ 1.27 down to i'= 26.15 (5σ) corresponding to ~M*+ 2.5 at this redshift. From the two-dimensional distribution of these photometrically selected galaxies, we newly identify seven candidates of galaxy groups or clusters where the surface density of red galaxies is significantly high (>5σ), in addition to the two known clusters. These candidates show clear red colour-magnitude sequences consistent with a passive evolution model, which suggests the existence of additional high-density regions around the Lynx superclusters.
Graham, John H; Robb, Daniel T; Poe, Amy R
2012-01-01
Distributed robustness is thought to influence the buffering of random phenotypic variation through the scale-free topology of gene regulatory, metabolic, and protein-protein interaction networks. If this hypothesis is true, then the phenotypic response to the perturbation of particular nodes in such a network should be proportional to the number of links those nodes make with neighboring nodes. This suggests a probability distribution approximating an inverse power-law of random phenotypic variation. Zero phenotypic variation, however, is impossible, because random molecular and cellular processes are essential to normal development. Consequently, a more realistic distribution should have a y-intercept close to zero in the lower tail, a mode greater than zero, and a long (fat) upper tail. The double Pareto-lognormal (DPLN) distribution is an ideal candidate distribution. It consists of a mixture of a lognormal body and upper and lower power-law tails. If our assumptions are true, the DPLN distribution should provide a better fit to random phenotypic variation in a large series of single-gene knockout lines than other skewed or symmetrical distributions. We fit a large published data set of single-gene knockout lines in Saccharomyces cerevisiae to seven different probability distributions: DPLN, right Pareto-lognormal (RPLN), left Pareto-lognormal (LPLN), normal, lognormal, exponential, and Pareto. The best model was judged by the Akaike Information Criterion (AIC). Phenotypic variation among gene knockouts in S. cerevisiae fits a double Pareto-lognormal (DPLN) distribution better than any of the alternative distributions, including the right Pareto-lognormal and lognormal distributions. A DPLN distribution is consistent with the hypothesis that developmental stability is mediated, in part, by distributed robustness, the resilience of gene regulatory, metabolic, and protein-protein interaction networks. Alternatively, multiplicative cell growth, and the mixing of lognormal distributions having different variances, may generate a DPLN distribution.
NASA Astrophysics Data System (ADS)
Xiang, Min; Qu, Qinqin; Chen, Cheng; Tian, Li; Zeng, Lingkang
2017-11-01
To improve the reliability of communication service in smart distribution grid (SDG), an access selection algorithm based on dynamic network status and different service types for heterogeneous wireless networks was proposed. The network performance index values were obtained in real time by multimode terminal and the variation trend of index values was analyzed by the growth matrix. The index weights were calculated by entropy-weight and then modified by rough set to get the final weights. Combining the grey relational analysis to sort the candidate networks, and the optimum communication network is selected. Simulation results show that the proposed algorithm can implement dynamically access selection in heterogeneous wireless networks of SDG effectively and reduce the network blocking probability.
Analysis of monoamine oxidase A (MAOA) promoter polymorphism in Finnish male alcoholics.
Saito, Takuya; Lachman, Herbert M; Diaz, Libna; Hallikainen, Tero; Kauhanen, Jussi; Salonen, Jukka T; Ryynänen, Olli-Pekka; Karvonen, Matti K; Syvälahti, Erkka; Pohjalainen, Tiina; Hietala, Jarmo; Tiihonen, Jari
2002-03-15
Alterations in monoamine oxidase A (MAOA) expression and enzyme activity may be associated with alcoholism and impulsive behavior. Therefore, functional polymorphisms in the MAOA gene would be good candidates to consider in the interindividual differences that exist in the susceptibility to alcoholism. One variant that has been considered as a candidate in alcoholism is a repeat polymorphism in the MAOA gene promoter. We analyzed a cohort of Finnish males with either type 1 or type 2 alcoholism, as well as controls, for differences in the distribution of MAOA promoter alleles. Based on other studies, we postulated that type 2 alcoholism, which is associated with antisocial behavior, but not type 1 alcoholism, would be correlated with the inheritance of the low promoter activity allele. However, we failed to find a difference in allele distribution in type 1 and type 2 alcoholics. In addition, there was no difference in the allele distribution when each group of alcoholics was compared with controls. However, when both groups of alcoholics were pooled and compared with controls, the difference in allele distribution reached a trend towards significance. Our results suggest a minimal association between the MAOA low activity promoter alleles and alcoholism, regardless of the presence or absence of antisocial behavior. Interestingly, approximately 3% of type 2 alcoholics were found to be heterozygous for the MAOA promoter polymorphism. Since MAOA is X-linked, the heterozygotes are probable cases of Klinefelter's syndrome (47,XXY) suggesting that X-chromosome aneuploidy may increase the risk for developing type 2 alcoholism.
Peak-flow characteristics of Virginia streams
Austin, Samuel H.; Krstolic, Jennifer L.; Wiegand, Ute
2011-01-01
Peak-flow annual exceedance probabilities, also called probability-percent chance flow estimates, and regional regression equations are provided describing the peak-flow characteristics of Virginia streams. Statistical methods are used to evaluate peak-flow data. Analysis of Virginia peak-flow data collected from 1895 through 2007 is summarized. Methods are provided for estimating unregulated peak flow of gaged and ungaged streams. Station peak-flow characteristics identified by fitting the logarithms of annual peak flows to a Log Pearson Type III frequency distribution yield annual exceedance probabilities of 0.5, 0.4292, 0.2, 0.1, 0.04, 0.02, 0.01, 0.005, and 0.002 for 476 streamgaging stations. Stream basin characteristics computed using spatial data and a geographic information system are used as explanatory variables in regional regression model equations for six physiographic regions to estimate regional annual exceedance probabilities at gaged and ungaged sites. Weighted peak-flow values that combine annual exceedance probabilities computed from gaging station data and from regional regression equations provide improved peak-flow estimates. Text, figures, and lists are provided summarizing selected peak-flow sites, delineated physiographic regions, peak-flow estimates, basin characteristics, regional regression model equations, error estimates, definitions, data sources, and candidate regression model equations. This study supersedes previous studies of peak flows in Virginia.
Nesting ecology and nest survival of lesser prairie-chickens on the Southern High Plains of Texas
Grisham, Blake A.; Borsdorf, Philip K.; Boal, Clint W.; Boydston, Kathy K.
2014-01-01
The decline in population and range of lesser prairie-chickens (Tympanuchus pallidicinctus) throughout the central and southern Great Plains has raised concerns considering their candidate status under the United States Endangered Species Act. Baseline ecological data for lesser prairie-chickens are limited, especially for the shinnery oak-grassland communities of Texas. This information is imperative because lesser prairie-chickens in shinnery oak grasslands occur at the extreme southwestern edge of their distribution. This geographic region is characterized by hot, arid climates, less fragmentation, and less anthropogenic development than within the remaining core distribution of the species. Thus, large expanses of open rangeland with less anthropogenic development and a climate that is classified as extreme for ground nesting birds may subsequently influence nest ecology, nest survival, and nest site selection differently compared to the rest of the distribution of the species. We investigated the nesting ecology of 50 radio-tagged lesser prairie-chicken hens from 2008 to 2011 in the shinnery oak-grassland communities in west Texas and found a substantial amount of inter-annual variation in incubation start date and percent of females incubating nests. Prairie-chickens were less likely to nest near unimproved roads and utility poles and in areas with more bare ground and litter. In contrast, hens selected areas dominated by grasses and shrubs and close to stock tanks to nest. Candidate models including visual obstruction best explained daily nest survival; a 5% increase in visual obstruction improved nest survival probability by 10%. The model-averaged probability of a nest surviving the incubation period was 0.43 (SE = 0.006; 95% CI: 0.23, 0.56). Our findings indicate that lesser prairie-chicken reproduction during our study period was dynamic and was correlated with seasonal weather patterns that ultimately promoted greater grass growth earlier in the nesting season that provided visual obstruction from predators.
Risk and utility in portfolio optimization
NASA Astrophysics Data System (ADS)
Cohen, Morrel H.; Natoli, Vincent D.
2003-06-01
Modern portfolio theory (MPT) addresses the problem of determining the optimum allocation of investment resources among a set of candidate assets. In the original mean-variance approach of Markowitz, volatility is taken as a proxy for risk, conflating uncertainty with risk. There have been many subsequent attempts to alleviate that weakness which, typically, combine utility and risk. We present here a modification of MPT based on the inclusion of separate risk and utility criteria. We define risk as the probability of failure to meet a pre-established investment goal. We define utility as the expectation of a utility function with positive and decreasing marginal value as a function of yield. The emphasis throughout is on long investment horizons for which risk-free assets do not exist. Analytic results are presented for a Gaussian probability distribution. Risk-utility relations are explored via empirical stock-price data, and an illustrative portfolio is optimized using the empirical data.
Unification of field theory and maximum entropy methods for learning probability densities
NASA Astrophysics Data System (ADS)
Kinney, Justin B.
2015-09-01
The need to estimate smooth probability distributions (a.k.a. probability densities) from finite sampled data is ubiquitous in science. Many approaches to this problem have been described, but none is yet regarded as providing a definitive solution. Maximum entropy estimation and Bayesian field theory are two such approaches. Both have origins in statistical physics, but the relationship between them has remained unclear. Here I unify these two methods by showing that every maximum entropy density estimate can be recovered in the infinite smoothness limit of an appropriate Bayesian field theory. I also show that Bayesian field theory estimation can be performed without imposing any boundary conditions on candidate densities, and that the infinite smoothness limit of these theories recovers the most common types of maximum entropy estimates. Bayesian field theory thus provides a natural test of the maximum entropy null hypothesis and, furthermore, returns an alternative (lower entropy) density estimate when the maximum entropy hypothesis is falsified. The computations necessary for this approach can be performed rapidly for one-dimensional data, and software for doing this is provided.
Unification of field theory and maximum entropy methods for learning probability densities.
Kinney, Justin B
2015-09-01
The need to estimate smooth probability distributions (a.k.a. probability densities) from finite sampled data is ubiquitous in science. Many approaches to this problem have been described, but none is yet regarded as providing a definitive solution. Maximum entropy estimation and Bayesian field theory are two such approaches. Both have origins in statistical physics, but the relationship between them has remained unclear. Here I unify these two methods by showing that every maximum entropy density estimate can be recovered in the infinite smoothness limit of an appropriate Bayesian field theory. I also show that Bayesian field theory estimation can be performed without imposing any boundary conditions on candidate densities, and that the infinite smoothness limit of these theories recovers the most common types of maximum entropy estimates. Bayesian field theory thus provides a natural test of the maximum entropy null hypothesis and, furthermore, returns an alternative (lower entropy) density estimate when the maximum entropy hypothesis is falsified. The computations necessary for this approach can be performed rapidly for one-dimensional data, and software for doing this is provided.
Hypervelocity star candidates in Gaia DR1/TGAS
NASA Astrophysics Data System (ADS)
Marchetti, T.; Rossi, E. M.; Kordopatis, G.; Brown, A. G. A.; Rimoldi, A.; Starkenburg, E.; Youakim, K.; Ashley, R.
2018-04-01
Hypervelocity stars (HVSs) are characterized by a total velocity in excess of the Galactic escape speed, and with trajectories consistent with coming from the Galactic Centre. We apply a novel data mining routine, an artificial neural network, to discover HVSs in the TGAS subset of the first data release of the Gaia satellite, using only the astrometry of the stars. We find 80 stars with a predicted probability >90% of being HVSs, and we retrieved radial velocities for 47 of those. We discover 14 objects with a total velocity in the Galactic rest frame >400 km s-1, and 5 of these have a probability >50% of being unbound from the Milky Way. Tracing back orbits in different Galactic potentials, we discover 1 HVS candidate, 5 bound HVS candidates, and 5 runaway star candidates with remarkably high velocities, between 400 and 780 km s-1. We wait for future Gaia releases to confirm the goodness of our sample and to increase the number of HVS candidates.
New paradigms for organ allocation and distribution in liver transplantation.
Kalra, Avash; Biggins, Scott W
2018-05-01
The 'Final Rule,' issued by the Health Resources and Service Administration in 2000, mandated that liver allocation policy should be based on disease severity and probability of death, and - among other factors - should be independent of a candidate's residence or listing. As a result, the Organ Procurement Transplantation Network/United Network for Organ Sharing (UNOS) has explored policy changes addressing geographic disparities without compromising outcomes. Major paradigm shifts are underway in U.S. liver allocation policy. New hepatocellular carcinoma exception policy incorporates tumor characteristics associated with posttransplantation outcomes, whereas a National Liver Review Board will promote a standardized process for awarding exception points. Meanwhile, following extensive debate, new allocation policy aims to reduce geographic disparity by broadening sharing to the UNOS region and 150-mile circle around the donor hospital for liver transplant candidates with a calculated model for end-stage liver disease score at least 32. Unnecessary organ travel will be reduced by granting 3 'proximity points' to candidates within the same donation service area (DSA) as a liver donor or within 150 nautical miles of the donor hospital, regardless of DSA or UNOS region. This review provides an evaluation of major policy changes in liver allocation from 2016 to 2018.
The Pitman-Yor Process and an Empirical Study of Choice Behavior
NASA Astrophysics Data System (ADS)
Hisakado, Masato; Sano, Fumiaki; Mori, Shintaro
2018-02-01
This study discusses choice behavior using a voting model in which voters can obtain information from a finite number of previous r voters. Voters vote for a candidate with a probability proportional to the previous vote ratio, which is visible to the voters. We obtain the Pitman sampling formula as the equilibrium distribution of r votes. We present the model as a process of posting on a bulletin board system, 2ch.net, where users can choose one of many threads to create a post. We explore how this choice depends on the last r posts and the distribution of these last r posts across threads. We conclude that the posting process is described by our voting model with analog herders for a small r, which might correspond to the time horizon of users' responses.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, Christine H.; Mittal, Tushar; Kuchner, Marc
During the Spitzer Space Telescope cryogenic mission, Guaranteed Time Observers, Legacy Teams, and General Observers obtained Infrared Spectrograph (IRS) observations of hundreds of debris disk candidates. We calibrated the spectra of 571 candidates, including 64 new IRAS and Multiband Imaging Photometer for Spitzer (MIPS) debris disks candidates, modeled their stellar photospheres, and produced a catalog of excess spectra for unresolved debris disks. For 499 targets with IRS excess but without strong spectral features (and a subset of 420 targets with additional MIPS 70 μm observations), we modeled the IRS (and MIPS data) assuming that the dust thermal emission was well-describedmore » using either a one- or two-temperature blackbody model. We calculated the probability for each model and computed the average probability to select among models. We found that the spectral energy distributions for the majority of objects (∼66%) were better described using a two-temperature model with warm (T {sub gr} ∼ 100-500 K) and cold (T {sub gr} ∼ 50-150 K) dust populations analogous to zodiacal and Kuiper Belt dust, suggesting that planetary systems are common in debris disks and zodiacal dust is common around host stars with ages up to ∼1 Gyr. We found that younger stars generally have disks with larger fractional infrared luminosities and higher grain temperatures and that higher-mass stars have disks with higher grain temperatures. We show that the increasing distance of dust around debris disks is inconsistent with self-stirred disk models, expected if these systems possess planets at 30-150 AU. Finally, we illustrate how observations of debris disks may be used to constrain the radial dependence of material in the minimum mass solar nebula.« less
X-Ray Bursts from the Transient Magnetar Candidate XTE J1810-197
NASA Technical Reports Server (NTRS)
Kouveliotou, Chryssa; Woods, Peter M.; Gavriil, Fotis P.; Kaspi, Victoria M.; Roberts, Mallory S. E.; Ibrahim, Alaa; Markwardt, Craig B.; Swank, Jean H.; Finger, Mark H.
2005-01-01
We have discovered four X-ray bursts, recorded with the Rossi X-ray Timing Explorer Proportional Counter Array between 2003 September and 2004 April, that we show to originate from the transient magnetar candidate XTE 51810-197. The burst morphologies consist of a short spike or multiple spikes lasting approx. 1 s each followed by extended tails of emission where the pulsed flux from XTE 51810-197 is significantly higher. The burst spikes are likely correlated with the pulse maxima, having a chance probability of a random phase distribution of 0.4%. The burst spectra are best fit to a blackbody with temperatures 4-8 keV, considerably harder than the persistent X-ray emission. During the X-ray tails following these bursts, the temperature rapidly cools as the flux declines, maintaining a constant emitting radius after the initial burst peak. The temporal and spectral characteristics of these bursts closely resemble the bursts seen from 1E 1048.1-5937 and a subset of the bursts detected from 1E 2259+586, thus establishing XTE J1810-197 as a magnetar candidate. The bursts detected from these three objects are sufficiently similar to one another, yet si,g&cantly differe2t from those seen from soft gamma repeaters, that they likely represent a new class of bursts from magnetar candidates exclusive (thus far) to the anomalous X-ray pulsar-like sources.
Krishnamurthy, Dilip; Sumaria, Vaidish; Viswanathan, Venkatasubramanian
2018-02-01
Density functional theory (DFT) calculations are being routinely used to identify new material candidates that approach activity near fundamental limits imposed by thermodynamics or scaling relations. DFT calculations are associated with inherent uncertainty, which limits the ability to delineate materials (distinguishability) that possess high activity. Development of error-estimation capabilities in DFT has enabled uncertainty propagation through activity-prediction models. In this work, we demonstrate an approach to propagating uncertainty through thermodynamic activity models leading to a probability distribution of the computed activity and thereby its expectation value. A new metric, prediction efficiency, is defined, which provides a quantitative measure of the ability to distinguish activity of materials and can be used to identify the optimal descriptor(s) ΔG opt . We demonstrate the framework for four important electrochemical reactions: hydrogen evolution, chlorine evolution, oxygen reduction and oxygen evolution. Future studies could utilize expected activity and prediction efficiency to significantly improve the prediction accuracy of highly active material candidates.
NASA Astrophysics Data System (ADS)
Agliozzo, C.; Nikutta, R.; Pignata, G.; Phillips, N. M.; Ingallinera, A.; Buemi, C.; Umana, G.; Leto, P.; Trigilio, C.; Noriega-Crespo, A.; Paladini, R.; Bufano, F.; Cavallaro, F.
2017-04-01
We present new observations of the nebula around the Magellanic candidate Luminous Blue Variable S61. These comprise high-resolution data acquired with the Australia Telescope Compact Array (ATCA), the Atacama Large Millimetre/Submillimetre Array (ALMA), and the VLT Imager and Spectrometer for mid Infrared (VISIR) at the Very Large Telescope. The nebula was detected only in the radio, up to 17 GHz. The 17 GHz ATCA map, with 0.8 arcsec resolution, allowed a morphological comparison with the Hα Hubble Space Telescope image. The radio nebula resembles a spherical shell, as in the optical. The spectral index map indicates that the radio emission is due to free-free transitions in the ionized, optically thin gas, but there are hints of inhomogeneities. We present our new public code RHOCUBE to model 3D density distributions and determine via Bayesian inference the nebula's geometric parameters. We applied the code to model the electron density distribution in the S61 nebula. We found that different distributions fit the data, but all of them converge to the same ionized mass, ˜ 0.1 M⊙, which is an order of magnitude smaller than previous estimates. We show how the nebula models can be used to derive the mass-loss history with high-temporal resolution. The nebula was probably formed through stellar winds, rather than eruptions. From the ALMA and VISIR non-detections, plus the derived extinction map, we deduce that the infrared emission observed by space telescopes must arise from extended, diffuse dust within the ionized region.
Chasing passive galaxies in the early Universe: a critical analysis in CANDELS GOODS-South
NASA Astrophysics Data System (ADS)
Merlin, E.; Fontana, A.; Castellano, M.; Santini, P.; Torelli, M.; Boutsia, K.; Wang, T.; Grazian, A.; Pentericci, L.; Schreiber, C.; Ciesla, L.; McLure, R.; Derriere, S.; Dunlop, J. S.; Elbaz, D.
2018-01-01
We search for passive galaxies at z > 3 in the GOODS-South field, using different techniques based on photometric data, and paying attention to develop methods that are sensitive to objects that have become passive shortly before the epoch of observation. We use CANDELS HST catalogues, ultra-deep Ks data and new IRAC photometry, performing spectral energy distribution fitting using models with abruptly quenched star formation histories. We then single out galaxies which are best fitted by a passively evolving model, and having only low probability (<5 per cent) star-forming solutions. We verify the effects of including nebular lines emission, and we consider possible solutions at different redshifts. The number of selected sources dramatically depends on the models used in the spectral energy distribution (SED) fitting. Without including emission lines and with photometric redshifts fixed at the CANDELS estimate, we single out 30 candidates; the inclusion of nebular lines emission reduces the sample to 10 objects; allowing for solutions at different redshifts, only two galaxies survive as robust candidates. Most of the candidates are not far-infrared emitters, corroborating their association with passive galaxies. Our results translate into an upper limit in the number density of ∼0.173 arcmin2 above the detection limit. However, we conclude that the selection of passive galaxies at z > 3 is still subject to significant uncertainties, being sensitive to assumptions in the SED modelling adopted and to the relatively low S/N of the objects. By means of dedicated simulations, we show that JWST will greatly enhance the accuracy, allowing for a much more robust classification.
Ferreira, Renata G; Jerusalinsky, Leandro; Silva, Thiago César Farias; de Souza Fialho, Marcos; de Araújo Roque, Alan; Fernandes, Adalberto; Arruda, Fátima
2009-10-01
Cebus flavius is a recently rediscovered species and a candidate for the 25 most endangered primate species list. It was hypothesized that the distribution of C. flavius was limited to the Atlantic Forest, while the occurrence of C. libidinosus in the Rio Grande do Norte (RN) Caatinga was inferred, given its occurrence in neighboring states. As a result of a survey in ten areas of the RN Caatinga, this paper reports on four Cebus populations, including the first occurrence of C. flavius in the Caatinga, and an expansion of the northwestern limits of distribution for the species. This C. flavius population may be a rare example of a process of geographic distribution retraction, and is probably the most endangered population of this species. New areas of occurrence of C. libidinosus are also described. Tool use sites were observed in association with reports of the presence of both capuchin species.
NASA Technical Reports Server (NTRS)
Liu, Wilson M.; Padgett, Deborah L.; Terebey, Susan; Angione, John; Rebull, Luisa M.; McCollum, Bruce; Fajardo-Acosta, Sergio; Leisawitz, David
2015-01-01
The Wide-Field Infrared Survey Explorer (WISE) has uncovered a striking cluster of young stellar object (YSO) candidates associated with the L1509 dark cloud in Auriga. The WISE observations, at 3.4, 4.6, 12, and 22 microns, show a number of objects with colors consistent with YSOs, and their spectral energy distributions suggest the presence of circumstellar dust emission, including numerous Class I, flat spectrum, and Class II objects. In general, the YSOs in L1509 are much more tightly clustered than YSOs in other dark clouds in the Taurus-Auriga star forming region, with Class I and flat spectrum objects confined to the densest aggregates, and Class II objects more sparsely distributed. We estimate a most probable distance of 485-700 pc, and possibly as far as the previously estimated distance of 2 kpc.
Mastery of Simple Probability Ideas Among G.C.E. Ordinary Level Mathematics Candidates
ERIC Educational Resources Information Center
Wood, R.; Brown, M.
1976-01-01
Questions concerning probability which appeared on the General Certificate of Education test are analyzed in terms of the processes required to answer them. The relationships between these processes and cognitive development are discussed. (SD)
A randomised approach for NARX model identification based on a multivariate Bernoulli distribution
NASA Astrophysics Data System (ADS)
Bianchi, F.; Falsone, A.; Prandini, M.; Piroddi, L.
2017-04-01
The identification of polynomial NARX models is typically performed by incremental model building techniques. These methods assess the importance of each regressor based on the evaluation of partial individual models, which may ultimately lead to erroneous model selections. A more robust assessment of the significance of a specific model term can be obtained by considering ensembles of models, as done by the RaMSS algorithm. In that context, the identification task is formulated in a probabilistic fashion and a Bernoulli distribution is employed to represent the probability that a regressor belongs to the target model. Then, samples of the model distribution are collected to gather reliable information to update it, until convergence to a specific model. The basic RaMSS algorithm employs multiple independent univariate Bernoulli distributions associated to the different candidate model terms, thus overlooking the correlations between different terms, which are typically important in the selection process. Here, a multivariate Bernoulli distribution is employed, in which the sampling of a given term is conditioned by the sampling of the others. The added complexity inherent in considering the regressor correlation properties is more than compensated by the achievable improvements in terms of accuracy of the model selection process.
Subaru Hyper Suprime-Cam Survey for an optical counterpart of GW170817
NASA Astrophysics Data System (ADS)
Tominaga, Nozomu; Tanaka, Masaomi; Morokuma, Tomoki; Utsumi, Yousuke; Yamaguchi, Masaki S.; Yasuda, Naoki; Tanaka, Masayuki; Yoshida, Michitoshi; Fujiyoshi, Takuya; Furusawa, Hisanori; Kawabata, Koji S.; Lee, Chien-Hsiu; Motohara, Kentaro; Ohsawa, Ryou; Ohta, Kouji; Terai, Tsuyoshi; Abe, Fumio; Aoki, Wako; Asakura, Yuichiro; Barway, Sudhanshu; Bond, Ian A.; Fujisawa, Kenta; Honda, Satoshi; Ioka, Kunihito; Itoh, Youichi; Kawai, Nobuyuki; Kim, Ji Hoon; Koshimoto, Naoki; Matsubayashi, Kazuya; Miyazaki, Shota; Saito, Tomoki; Sekiguchi, Yuichiro; Sumi, Takahiro; Tristram, Paul J.
2018-03-01
We perform a z-band survey for an optical counterpart of the binary neutron star coalescence GW170817 with Subaru/Hyper Suprime-Cam. Our untargeted transient search covers 23.6 deg2 corresponding to the 56.6% credible region of GW170817 and reaches the 50% completeness magnitude of 20.6 mag on average. As a result, we find 60 candidate extragalactic transients, including J-GEM17btc (also known as SSS17a/DLT17ck). While J-GEM17btc is associated with NGC 4993, which is firmly located inside the 3D skymap of GW170817, the other 59 candidates do not have distance information in the GLADE v2 catalog or NASA/IPAC Extragalactic Database. Among 59 of the candidates, 58 are located at the center of extended objects in the Pan-STARRS1 catalog, while one candidate has an offset. We present location, z-band apparent magnitude, and time variability of the candidates and evaluate the probabilities that they are located within the 3D skymap of GW170817. The probability for J-GEM17btc is 64%, which is much higher than for the other 59 candidates (9.3 × 10-3-2.1 × 10-1%). Furthermore, the possibility that at least one of the other 59 candidates is located within the 3D skymap is only 3.2%. Therefore, we conclude that J-GEM17btc is the most likely and distinguished candidate to be the optical counterpart of GW170817.
Two protostar candidates in the bright-rimmed dark cloud LDN 1206
NASA Technical Reports Server (NTRS)
Ressler, Michael E.; Shure, Mark
1991-01-01
The discovery of several near IR objects associated with two IRAS point sources in the LDN 1206 region is reported. IRAS 22272 + 6358A is probably a 'protostar' which is seen only in scattered light at near-IR wavelengths because of heavy obscuration by an almost edge-on circumstellar disk. In contrast, IRAS 22272 + 6358B is directly visible at these wavelengths and is perhaps an object which lies between protostars and T-Tauri stars in its evolution. Both direct and polarimetric K-band images of the region are presented, as well as spectral energy distributions constructed from J, H, K, L, L-prime, and M data and published far-IR and mm data.
A Machine Learning Ensemble Classifier for Early Prediction of Diabetic Retinopathy.
S K, Somasundaram; P, Alli
2017-11-09
The main complication of diabetes is Diabetic retinopathy (DR), retinal vascular disease and it leads to the blindness. Regular screening for early DR disease detection is considered as an intensive labor and resource oriented task. Therefore, automatic detection of DR diseases is performed only by using the computational technique is the great solution. An automatic method is more reliable to determine the presence of an abnormality in Fundus images (FI) but, the classification process is poorly performed. Recently, few research works have been designed for analyzing texture discrimination capacity in FI to distinguish the healthy images. However, the feature extraction (FE) process was not performed well, due to the high dimensionality. Therefore, to identify retinal features for DR disease diagnosis and early detection using Machine Learning and Ensemble Classification method, called, Machine Learning Bagging Ensemble Classifier (ML-BEC) is designed. The ML-BEC method comprises of two stages. The first stage in ML-BEC method comprises extraction of the candidate objects from Retinal Images (RI). The candidate objects or the features for DR disease diagnosis include blood vessels, optic nerve, neural tissue, neuroretinal rim, optic disc size, thickness and variance. These features are initially extracted by applying Machine Learning technique called, t-distributed Stochastic Neighbor Embedding (t-SNE). Besides, t-SNE generates a probability distribution across high-dimensional images where the images are separated into similar and dissimilar pairs. Then, t-SNE describes a similar probability distribution across the points in the low-dimensional map. This lessens the Kullback-Leibler divergence among two distributions regarding the locations of the points on the map. The second stage comprises of application of ensemble classifiers to the extracted features for providing accurate analysis of digital FI using machine learning. In this stage, an automatic detection of DR screening system using Bagging Ensemble Classifier (BEC) is investigated. With the help of voting the process in ML-BEC, bagging minimizes the error due to variance of the base classifier. With the publicly available retinal image databases, our classifier is trained with 25% of RI. Results show that the ensemble classifier can achieve better classification accuracy (CA) than single classification models. Empirical experiments suggest that the machine learning-based ensemble classifier is efficient for further reducing DR classification time (CT).
A selection of giant radio sources from NVSS
Proctor, D. D.
2016-06-01
Results of the application of pattern-recognition techniques to the problem of identifying giant radio sources (GRSs) from the data in the NVSS catalog are presented, and issues affecting the process are explored. Decision-tree pattern-recognition software was applied to training-set source pairs developed from known NVSS large-angular-size radio galaxies. The full training set consisted of 51,195 source pairs, 48 of which were known GRSs for which each lobe was primarily represented by a single catalog component. The source pairs had a maximum separation ofmore » $$20^{\\prime} $$ and a minimum component area of 1.87 square arcmin at the 1.4 mJy level. The importance of comparing the resulting probability distributions of the training and application sets for cases of unknown class ratio is demonstrated. The probability of correctly ranking a randomly selected (GRS, non-GRS) pair from the best of the tested classifiers was determined to be 97.8 ± 1.5%. The best classifiers were applied to the over 870,000 candidate pairs from the entire catalog. Images of higher-ranked sources were visually screened, and a table of over 1600 candidates, including morphological annotation, is presented. These systems include doubles and triples, wide-angle tail and narrow-angle tail, S- or Z-shaped systems, and core-jets and resolved cores. In conclusion, while some resolved-lobe systems are recovered with this technique, generally it is expected that such systems would require a different approach.« less
On the physical nature of globular cluster candidates in the Milky Way bulge
NASA Astrophysics Data System (ADS)
Piatti, Andrés E.
2018-06-01
We present results from 2MASS JKs photometry on the physical reality of recently reported globular cluster (GC) candidates in the Milky Way (MW) bulge. We relied our analysis on photometric membership probabilities that allowed us to distinguish real stellar aggregates from the composite field star population. When building colour-magnitude diagrams and stellar density maps for stars at different membership probability levels, the genuine GC candidate populations are clearly highlighted. We then used the tip of the red giant branch (RGB) as distance estimator, resulting in heliocentric distances that place many of the objects in regions near the MW bulge, where no GC had been previously recognized. Some few GC candidates resulted to be MW halo/disc objects. Metallicities estimated from the standard RGB method are in agreement with the values expected according to the position of the GC candidates in the Galaxy. Finally, we derived, for the first time, their structural parameters. We found that the studied objects have core, half-light, and tidal radii in the ranges spanned by the population of known MW GCs. Their internal dynamical evolutionary stages will be described properly when their masses are estimated.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moorhead, Althea V.; Ford, Eric B.; Morehead, Robert C.
Doppler planet searches have discovered that giant planets follow orbits with a wide range of orbital eccentricities, revolutionizing theories of planet formation. The discovery of hundreds of exoplanet candidates by NASA's Kepler mission enables astronomers to characterize the eccentricity distribution of small exoplanets. Measuring the eccentricity of individual planets is only practical in favorable cases that are amenable to complementary techniques (e.g., radial velocities, transit timing variations, occultation photometry). Yet even in the absence of individual eccentricities, it is possible to study the distribution of eccentricities based on the distribution of transit durations (relative to the maximum transit duration formore » a circular orbit). We analyze the transit duration distribution of Kepler planet candidates. We find that for host stars with T{sub eff} > 5100 K we cannot invert this to infer the eccentricity distribution at this time due to uncertainties and possible systematics in the host star densities. With this limitation in mind, we compare the observed transit duration distribution with models to rule out extreme distributions. If we assume a Rayleigh eccentricity distribution for Kepler planet candidates, then we find best fits with a mean eccentricity of 0.1-0.25 for host stars with T{sub eff} {<=} 5100 K. We compare the transit duration distribution for different subsets of Kepler planet candidates and discuss tentative trends with planetary radius and multiplicity. High-precision spectroscopic follow-up observations for a large sample of host stars will be required to confirm which trends are real and which are the results of systematic errors in stellar radii. Finally, we identify planet candidates that must be eccentric or have a significantly underestimated stellar radius.« less
NASA Astrophysics Data System (ADS)
Doksæter, L.; Olsen, E.; Nøttestad, L.; Fernö, A.
2008-01-01
During Leg 1 of the MAR-ECO expedition on the R.V. G.O. Sars in June 2004 four main species of dolphins were observed along the Mid-Atlantic Ridge from Iceland to the Azores: pilot whale ( Globicephala melas) ( n=326), short-beaked common dolphin ( Delphinus delphis) ( n=273), white-sided dolphin ( Lagenorhynchus acutus) ( n=103), and striped dolphin ( Stenella coeruleoalba) ( n=86). Pilot whales and white-sided dolphins were found in cold (5-16 °C) and less-saline (34.6-35.8‰) water masses in the northern part of the study area, whereas common and striped dolphins inhabited warmer (12-22 °C) and more-saline (34.8-36.7‰) waters in the south. Dolphins tended to aggregate in areas of steep slopes, but actual bottom depth appeared to be less important. Based on spatial correlations between dolphin occurrence and candidate prey organisms recorded acoustically and by midwater trawling, mesopelagic fishes and squids were assumed to be important prey items, with Benthosema glaciale probably being the most important prey for pilot whales and white-sided dolphins, while Lampanyctus macdonaldi, Stomias boa ferox and Chauliodus sloani were probably of particular importance for common dolphins. Cephalopods, especially Gonatus sp. and Teuthowenia megalops were the most likely prey species of pilot whales and striped dolphins, respectively. The difference in physical habitat north and south of the Sub-polar Frontal Zone seemed to have important effects on prey distribution, in turn influencing dolphin distribution.
Attribution and Stereotype Explanations of Non-Visible Handicap Discrimination.
ERIC Educational Resources Information Center
Farrow, Dana
A male epileptic, male nonepileptic, female epileptic, or female nonepileptic job candidate was evaluated for either an auto sales or receptionist job by 112 university students in personnel or behavioral science courses. The female epileptic and the male nonepileptic candidates had significantly higher probabilities of being hired than the other…
Engagement in the electoral processes: scaling laws and the role of political positions.
Mantovani, M C; Ribeiro, H V; Lenzi, E K; Picoli, S; Mendes, R S
2013-08-01
We report on a statistical analysis of the engagement in the electoral processes of all Brazilian cities by considering the number of party memberships and the number of candidates for mayor and councillor. By investigating the relationships between the number of party members and the population of voters, we have found that the functional forms of these relationships are well described by sublinear power laws (allometric scaling) surrounded by a multiplicative log-normal noise. We have observed that this pattern is quite similar to those we previously reported for the relationships between the number of candidates (mayor and councillor) and population of voters [Europhys. Lett. 96, 48001 (2011)], suggesting that similar universal laws may be ruling the engagement in the electoral processes. We also note that the power-law exponents display a clear hierarchy, where the more influential is the political position the smaller is the value of the exponent. We have also investigated the probability distributions of the number of candidates (mayor and councillor), party memberships, and voters. The results indicate that the most influential positions are characterized by distributions with very short tails, while less influential positions display an intermediate power-law decay before showing an exponential-like cutoff. We discuss the possibility that, in addition to the political power of the position, limitations in the number of available seats can also be connected with this changing of behavior. We further believe that our empirical findings point out to an under-representation effect, where the larger the city is, the larger are the obstacles for more individuals to become directly engaged in the electoral process.
Forecasting Lightning at Kennedy Space Center/Cape Canaveral Air Force Station, Florida
NASA Technical Reports Server (NTRS)
Lambert, Winfred; Wheeler, Mark; Roeder, William
2005-01-01
The Applied Meteorology Unit (AMU) developed a set of statistical forecast equations that provide a probability of lightning occurrence on Kennedy Space Center (KSC) I Cape Canaveral Air Force Station (CCAFS) for the day during the warm season (May September). The 45th Weather Squadron (45 WS) forecasters at CCAFS in Florida include a probability of lightning occurrence in their daily 24-hour and weekly planning forecasts, which are briefed at 1100 UTC (0700 EDT). This information is used for general scheduling of operations at CCAFS and KSC. Forecasters at the Spaceflight Meteorology Group also make thunderstorm forecasts for the KSC/CCAFS area during Shuttle flight operations. Much of the current lightning probability forecast at both groups is based on a subjective analysis of model and observational data. The objective tool currently available is the Neumann-Pfeffer Thunderstorm Index (NPTI, Neumann 1971), developed specifically for the KSCICCAFS area over 30 years ago. However, recent studies have shown that 1-day persistence provides a better forecast than the NPTI, indicating that the NPTI needed to be upgraded or replaced. Because they require a tool that provides a reliable estimate of the daily thunderstorm probability forecast, the 45 WS forecasters requested that the AMU develop a new lightning probability forecast tool using recent data and more sophisticated techniques now possible through more computing power than that available over 30 years ago. The equation development incorporated results from two research projects that investigated causes of lightning occurrence near KSCICCAFS and over the Florida peninsula. One proved that logistic regression outperformed the linear regression method used in NPTI, even when the same predictors were used. The other study found relationships between large scale flow regimes and spatial lightning distributions over Florida. Lightning, probabilities based on these flow regimes were used as candidate predictors in the equation development. Fifteen years (1 989-2003) of warm season data were used to develop the forecast equations. The data sources included a local network of cloud-to-ground lightning sensors called the Cloud-to-Ground Lightning Surveillance System (CGLSS), 1200 UTC Florida synoptic soundings, and the 1000 UTC CCAFS sounding. Data from CGLSS were used to determine lightning occurrence for each day. The 1200 UTC soundings were used to calculate the synoptic-scale flow regimes and the 1000 UTC soundings were used to calculate local stability parameters, which were used as candidate predictors of lightning occurrence. Five logistic regression forecast equations were created through careful selection and elimination of the candidate predictors. The resulting equations contain five to six predictors each. Results from four performance tests indicated that the equations showed an increase in skill over several standard forecasting methods, good reliability, an ability to distinguish between non-lightning and lightning days, and good accuracy measures and skill scores. Given the overall good performance the 45 WS requested that the equations be transitioned to operations and added to the current set of tools used to determine the daily lightning probability of occurrence.
Using Elicited Choice Probabilities in Hypothetical Elections to Study Decisions to Vote
Delavande, Adeline; Manski, Charles F.
2015-01-01
This paper demonstrates the feasibility and usefulness of survey research asking respondents to report voting probabilities in hypothetical election scenarios. Posing scenarios enriches the data available for studies of voting decisions, as a researcher can pose many more and varied scenarios than the elections that persons actually face. Multiple scenarios were presented to over 4,000 participants in the American Life Panel (ALP). Each described a hypothetical presidential election, giving characteristics measuring candidate preference, closeness of the election, and the time cost of voting. Persons were asked the probability that they would vote in this election and were willing and able to respond. We analyzed the data through direct study of the variation of voting probabilities with election characteristics and through estimation of a random utility model of voting. Voting time and election closeness were notable determinants of decisions to vote, but not candidate preference. Most findings were corroborated through estimation of a model fit to ALP data on respondents' actual voting behavior in the 2012 election. PMID:25705068
Thomassen, Henri A.; Fuller, Trevon; Asefi-Najafabady, Salvi; Shiplacoff, Julia A. G.; Mulembakani, Prime M.; Blumberg, Seth; Johnston, Sara C.; Kisalu, Neville K.; Kinkela, Timothée L.; Fair, Joseph N.; Wolfe, Nathan D.; Shongo, Robert L.; LeBreton, Matthew; Meyer, Hermann; Wright, Linda L.; Muyembe, Jean-Jacques; Buermann, Wolfgang; Okitolonda, Emile; Hensley, Lisa E.; Lloyd-Smith, James O.; Smith, Thomas B.; Rimoin, Anne W.
2013-01-01
Climate change is predicted to result in changes in the geographic ranges and local prevalence of infectious diseases, either through direct effects on the pathogen, or indirectly through range shifts in vector and reservoir species. To better understand the occurrence of monkeypox virus (MPXV), an emerging Orthopoxvirus in humans, under contemporary and future climate conditions, we used ecological niche modeling techniques in conjunction with climate and remote-sensing variables. We first created spatially explicit probability distributions of its candidate reservoir species in Africa's Congo Basin. Reservoir species distributions were subsequently used to model current and projected future distributions of human monkeypox (MPX). Results indicate that forest clearing and climate are significant driving factors of the transmission of MPX from wildlife to humans under current climate conditions. Models under contemporary climate conditions performed well, as indicated by high values for the area under the receiver operator curve (AUC), and tests on spatially randomly and non-randomly omitted test data. Future projections were made on IPCC 4th Assessment climate change scenarios for 2050 and 2080, ranging from more conservative to more aggressive, and representing the potential variation within which range shifts can be expected to occur. Future projections showed range shifts into regions where MPX has not been recorded previously. Increased suitability for MPX was predicted in eastern Democratic Republic of Congo. Models developed here are useful for identifying areas where environmental conditions may become more suitable for human MPX; targeting candidate reservoir species for future screening efforts; and prioritizing regions for future MPX surveillance efforts. PMID:23935820
Understanding exoplanet populations with simulation-based methods
NASA Astrophysics Data System (ADS)
Morehead, Robert Charles
The Kepler candidate catalog represents an unprecedented sample of exoplanet host stars. This dataset is ideal for probing the populations of exoplanet systems and exploring their architectures. Confirming transiting exoplanets candidates through traditional follow-up methods is challenging, especially for faint host stars. Most of Kepler's validated planets relied on statistical methods to separate true planets from false-positives. Multiple transiting planet systems (MTPS) have been previously shown to have low false-positive rates and over 850 planets in MTPSs have been statistically validated so far. We show that the period-normalized transit duration ratio (xi) offers additional information that can be used to establish the planetary nature of these systems. We briefly discuss the observed distribution of xi for the Q1-Q17 Kepler Candidate Search. We also use xi to develop a Bayesian statistical framework combined with Monte Carlo methods to determine which pairs of planet candidates in an MTPS are consistent with the planet hypothesis for a sample of 862 MTPSs that include candidate planets, confirmed planets, and known false-positives. This analysis proves to be efficient and advantageous in that it only requires catalog-level bulk candidate properties and galactic population modeling to compute the probabilities of a myriad of feasible scenarios composed of background and companion stellar blends in the photometric aperture, without needing additional observational follow-up. Our results agree with the previous results of a low false-positive rate in the Kepler MTPSs. This implies, independently of any other estimates, that most of the MTPSs detected by Kepler are planetary in nature, but that a substantial fraction could be orbiting stars other than then the putative target star, and therefore may be subject to significant error in the inferred planet parameters resulting from unknown or mismeasured stellar host attributes. We also apply approximate Bayesian computation (ABC) using forward simulations of the Kepler planet catalog to simultaneously constrain the distributions of mutual inclination between the planets, orbital eccentricity, the underlying number of planets per planetary system, and the fraction of stars that host planet systems in a subsample of Kepler candidate planets using SimpleABC, a Python package we developed that is a general-purpose framework for ABC analysis. For our investigation into planet architectures, we limit our investigation to candidates in orbits from 10 to 320 days, where the false-positive contamination rate is expected to be low. We test two models, the first is an independent eccentricity ( e) model where mutual inclination and e are drawn from Rayleigh distributions with dispersions sigmaim and sigmae, planets per planetary system is drawn from a Poisson distribution with mean lambda, and the fraction of stars with planetary systems is drawn from two-state categorical distribution parameterized by etap. We also test an Equipartition Model identical to the Independent e Model, except that sigmae is linked to sigmaim by a scaling factor gammae. For the Independent e Model, we find sigmaim = 5.51° +8.00-3.35, sigmae = 0.03+0.05-0.01, lambda = 6.62+7.74 -3.36, and etap = 0.20 +0.18-0.11. For the Equipartition Model, we find sigmaim = 1.15°+0.56-0.33 , gammae = 1.38+1.89 -0.93, lambda = 2.25+0.56-0.29, and etap = 0.56+0.08-0.11 . These results, especially the Equipartition Model, are in good agreement with previous studies. However, deficiencies in our single population models suggest that at least one additional subpopulation of planet systems is needed to explain the Kepler sample, providing more confirmation of the so-called "Kepler Dichotomy".
NASA Astrophysics Data System (ADS)
DeMarco, Adam Ward
The turbulent motions with the atmospheric boundary layer exist over a wide range of spatial and temporal scales and are very difficult to characterize. Thus, to explore the behavior of such complex flow enviroments, it is customary to examine their properties from a statistical perspective. Utilizing the probability density functions of velocity and temperature increments, deltau and deltaT, respectively, this work investigates their multiscale behavior to uncover the unique traits that have yet to be thoroughly studied. Utilizing diverse datasets, including idealized, wind tunnel experiments, atmospheric turbulence field measurements, multi-year ABL tower observations, and mesoscale models simulations, this study reveals remarkable similiarities (and some differences) between the small and larger scale components of the probability density functions increments fields. This comprehensive analysis also utilizes a set of statistical distributions to showcase their ability to capture features of the velocity and temperature increments' probability density functions (pdfs) across multiscale atmospheric motions. An approach is proposed for estimating their pdfs utilizing the maximum likelihood estimation (MLE) technique, which has never been conducted utilizing atmospheric data. Using this technique, we reveal the ability to estimate higher-order moments accurately with a limited sample size, which has been a persistent concern for atmospheric turbulence research. With the use robust Goodness of Fit (GoF) metrics, we quantitatively reveal the accuracy of the distributions to the diverse dataset. Through this analysis, it is shown that the normal inverse Gaussian (NIG) distribution is a prime candidate to be used as an estimate of the increment pdfs fields. Therefore, using the NIG model and its parameters, we display the variations in the increments over a range of scales revealing some unique scale-dependent qualities under various stability and ow conditions. This novel approach can provide a method of characterizing increment fields with the sole use of only four pdf parameters. Also, we investigate the capability of the current state-of-the-art mesoscale atmospheric models to predict the features and highlight the potential for use for future model development. With the knowledge gained in this study, a number of applications can benefit by using our methodology, including the wind energy and optical wave propagation fields.
NASA Astrophysics Data System (ADS)
Gromov, Yu Yu; Minin, Yu V.; Ivanova, O. G.; Morozova, O. N.
2018-03-01
Multidimensional discrete distributions of probabilities of independent random values were received. Their one-dimensional distribution is widely used in probability theory. Producing functions of those multidimensional distributions were also received.
First optical candidate for a recovered classical nova in a globular cluster - Nova 1938 in M14
NASA Technical Reports Server (NTRS)
Shara, Michael M.; Potter, Michael; Moffat, Anthony F. J.; Hogg, Helen S.; Wehlau, Amelia
1986-01-01
U, B, V, R, and H-alpha CCD frames of the field of the nova which appeared in the globular cluster M14 in 1938 have been compared with the nova discovery images. On the basis of positional coincidence, brightness, and blue color, a candidate nova was identified and its right ascension and declination to within 1 arcsec each. Confirmation of the candidate and detailed study of the quiescent nova will probably require Hubble Space Telescope observations.
Hobby, Kirsten; Gallagher, Richard T; Caldwell, Patrick; Wilson, Ian D
2009-01-01
This work describes the identification of 'isotopically enriched' metabolites of 4-cyanoaniline using the unique features of the software package 'Spectral Simplicity'. The software is capable of creating the theoretical mass spectra for partially isotope-enriched compounds, and subsequently performing an elemental composition analysis to give the elemental formula for the 'isotopically enriched' metabolite. A novel mass spectral correlation method, called 'FuzzyFit', was employed. 'FuzzyFit' utilises the expected experimental distribution of errors in both mass accuracy and isotope pattern and enables discrimination between statistically probable and improbable candidate formulae. The software correctly determined the molecular formulae of ten previously described metabolites of 4-cyanoaniline confirming the technique of partial isotope enrichment can produce results analogous to standard methodologies. Six previously unknown species were also identified, based on the presence of the unique 'designer' isotope ratio. Three of the unknowns were tentatively identified as N-acetylglutamine, O-methyl-N acetylglucuronide and a putative fatty acid conjugate. The discovery of a significant number of unknown species of a model drug with a comprehensive history of investigation highlights the potential for enhancement to the analytical process by the use of 'designer' isotope ratio compounds. The 'FuzzyFit' methodology significantly aided the elucidation of candidate formulae, by provision of a vastly simplified candidate formula data set. Copyright (c) 2008 John Wiley & Sons, Ltd.
Probability and surprisal in auditory comprehension of morphologically complex words.
Balling, Laura Winther; Baayen, R Harald
2012-10-01
Two auditory lexical decision experiments document for morphologically complex words two points at which the probability of a target word given the evidence shifts dramatically. The first point is reached when morphologically unrelated competitors are no longer compatible with the evidence. Adapting terminology from Marslen-Wilson (1984), we refer to this as the word's initial uniqueness point (UP1). The second point is the complex uniqueness point (CUP) introduced by Balling and Baayen (2008), at which morphologically related competitors become incompatible with the input. Later initial as well as complex uniqueness points predict longer response latencies. We argue that the effects of these uniqueness points arise due to the large surprisal (Levy, 2008) carried by the phonemes at these uniqueness points, and provide independent evidence that how cumulative surprisal builds up in the course of the word co-determines response latencies. The presence of effects of surprisal, both at the initial uniqueness point of complex words, and cumulatively throughout the word, challenges the Shortlist B model of Norris and McQueen (2008), and suggests that a Bayesian approach to auditory comprehension requires complementation from information theory in order to do justice to the cognitive cost of updating probability distributions over lexical candidates. Copyright © 2012 Elsevier B.V. All rights reserved.
Finding strong lenses in CFHTLS using convolutional neural networks
NASA Astrophysics Data System (ADS)
Jacobs, C.; Glazebrook, K.; Collett, T.; More, A.; McCarthy, C.
2017-10-01
We train and apply convolutional neural networks, a machine learning technique developed to learn from and classify image data, to Canada-France-Hawaii Telescope Legacy Survey (CFHTLS) imaging for the identification of potential strong lensing systems. An ensemble of four convolutional neural networks was trained on images of simulated galaxy-galaxy lenses. The training sets consisted of a total of 62 406 simulated lenses and 64 673 non-lens negative examples generated with two different methodologies. An ensemble of trained networks was applied to all of the 171 deg2 of the CFHTLS wide field image data, identifying 18 861 candidates including 63 known and 139 other potential lens candidates. A second search of 1.4 million early-type galaxies selected from the survey catalogue as potential deflectors, identified 2465 candidates including 117 previously known lens candidates, 29 confirmed lenses/high-quality lens candidates, 266 novel probable or potential lenses and 2097 candidates we classify as false positives. For the catalogue-based search we estimate a completeness of 21-28 per cent with respect to detectable lenses and a purity of 15 per cent, with a false-positive rate of 1 in 671 images tested. We predict a human astronomer reviewing candidates produced by the system would identify 20 probable lenses and 100 possible lenses per hour in a sample selected by the robot. Convolutional neural networks are therefore a promising tool for use in the search for lenses in current and forthcoming surveys such as the Dark Energy Survey and the Large Synoptic Survey Telescope.
NASA Astrophysics Data System (ADS)
Utsumi, Yousuke; Tominaga, Nozomu; Tanaka, Masaomi; Morokuma, Tomoki; Yoshida, Michitoshi; Asakura, Yuichiro; Finet, François; Furusawa, Hisanori; Kawabata, Koji S.; Liu, Wei; Matsubayashi, Kazuya; Moritani, Yuki; Motohara, Kentaro; Nakata, Fumiaki; Ohta, Kouji; Terai, Tsuyoshi; Uemura, Makoto; Yasuda, Naoki
2018-01-01
We present the results of detailed analysis of an optical imaging survey conducted using the Subaru/Hyper Suprime-Cam (HSC) that aimed to identify an optical counterpart to the gravitational wave event GW151226. In half a night, the i- and z-band imaging survey by HSC covered 63.5 deg2 of the error region, which contains about 7% of the LIGO localization probability, and the same field was observed in three different epochs. The detectable magnitude of the candidates in a differenced image is evaluated as i ˜ 23.2 mag for the requirement of at least two 5 σ detections, and 1744 candidates are discovered. Assuming a kilonova as an optical counterpart, we compare the optical properties of the candidates with model predictions. A red and rapidly declining light curve condition enables the discrimination of a kilonova from other transients, and a small number of candidates satisfy this condition. The presence of stellar-like counterparts in the reference frame suggests that the surviving candidates are likely to be flare stars. The fact that most of those candidates are in the galactic plane, |b| < 5°, supports this interpretation. We also check whether the candidates are associated with the nearby GLADE galaxies, which reduces the number of contaminants even with a looser color cut. When a better probability map (with localization accuracy of ˜50 deg2) is available, kilonova searches of up to approximately 200 Mpc will become feasible by conducting immediate follow-up observations with an interval of 3-6 d.
Substellar objects in nearby young clusters (SONYC). VIII. Substellar population in Lupus 3
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mužić, Koraljka; Scholz, Alexander; Geers, Vincent C.
2014-04-20
SONYC—Substellar Objects in Nearby Young Clusters—is a survey program to investigate the frequency and properties of substellar objects in nearby star-forming regions. We present a new imaging and spectroscopic survey conducted in the young (∼1 Myr), nearby (∼200 pc) star-forming region Lupus 3. Deep optical and near-infrared images were obtained with MOSAIC-II and NEWFIRM at the CTIO 4 m telescope, covering ∼1.4 deg{sup 2} on the sky. The i-band completeness limit of 20.3 mag is equivalent to 0.009-0.02 M {sub ☉}, for A{sub V} ≤ 5. Photometry and 11-12 yr baseline proper motions were used to select candidate low-mass membersmore » of Lupus 3. We performed a spectroscopic follow-up of 123 candidates, using VIMOS at the Very Large Telescope, and we identify 7 probable members, among which 4 have spectral type later than M6.0 and T {sub eff} ≤ 3000 K, i.e., are probably substellar in nature. Two of the new probable members of Lupus 3 appear underluminous for their spectral class and exhibit emission line spectrum with strong H{sub α} or forbidden lines associated with active accretion. We derive a relation between the spectral type and effective temperature: T {sub eff} = (4120 ± 175) – (172 ± 26) × SpT, where SpT refers to the M spectral subtype between 1 and 9. Combining our results with the previous works on Lupus 3, we show that the spectral type distribution is consistent with that in other star-forming regions, as well as the derived star-to-brown dwarf ratio of 2.0-3.3. We compile a census of all spectroscopically confirmed low-mass members with spectral type M0 or later.« less
The MACHO Project: Microlensing Results from 5.7 Years of LMC Observations
NASA Astrophysics Data System (ADS)
Becker, A. C.; Alcock, C.; Allsman, R. A.; Alves, D.; Axelrod, T. S.; Bennett, D. P.; Cook, K. H.; Drake, A. J.; Freeman, K. C.; Geha, M.; Griest, K.; Lehner, M. J.; Marshall, S. L.; Minniti, D.; Nelson, C. A.; Peterson, B. A.; Popowski, P.; Pratt, M. R.; Quinn, P. J.; Rodgers, A. W.; Stubbs, C. W.; Sutherland, W.; Tomaney, A. B.; Vandehei, T.; Welch, D. L.; MACHO Collaboration
1999-12-01
We report on our search for microlensing towards the Large Magellanic Cloud (LMC). Analysis of 5.7 years of photometry on 11.9 million stars in the LMC reveals 17 candidate microlensing events. A careful treatment of our detection efficiency shows that this is significantly more than the 2 to 4 events expected from lensing by known stellar populations. The timescales (t) of the events range from 34 to 230 days. The spatial distribution of events is probably inconsistent with LMC/LMC disk self-lensing, but consistent with an extended lens distribution such as a Milky Way or LMC halo. The optical depth results appear consistent with, but are lower than, our previous 2.1 year results. Besides a factor of 3.4 increase in exposure, this new work also includes an improved efficiency determination, improved likelihood analysis, and more thorough testing of robustness and systematic errors, especially in regards to the treatment of potential backgrounds to microlensing, such as supernovae in galaxies behind the LMC.
Tygert, Mark
2010-09-21
We discuss several tests for determining whether a given set of independent and identically distributed (i.i.d.) draws does not come from a specified probability density function. The most commonly used are Kolmogorov-Smirnov tests, particularly Kuiper's variant, which focus on discrepancies between the cumulative distribution function for the specified probability density and the empirical cumulative distribution function for the given set of i.i.d. draws. Unfortunately, variations in the probability density function often get smoothed over in the cumulative distribution function, making it difficult to detect discrepancies in regions where the probability density is small in comparison with its values in surrounding regions. We discuss tests without this deficiency, complementing the classical methods. The tests of the present paper are based on the plain fact that it is unlikely to draw a random number whose probability is small, provided that the draw is taken from the same distribution used in calculating the probability (thus, if we draw a random number whose probability is small, then we can be confident that we did not draw the number from the same distribution used in calculating the probability).
Jarman, Kristin H [Richland, WA; Cannon, William R [Richland, WA; Jarman, Kenneth D [Richland, WA; Heredia-Langner, Alejandro [Richland, WA
2011-07-12
Peptides are identified from a list of candidates using collision-induced dissociation tandem mass spectrometry data. A probabilistic model for the occurrence of spectral peaks corresponding to frequently observed partial peptide fragment ions is applied. As part of the identification procedure, a probability score is produced that indicates the likelihood of any given candidate being the correct match. The statistical significance of the score is known without necessarily having reference to the actual identity of the peptide. In one form of the invention, a genetic algorithm is applied to candidate peptides using an objective function that takes into account the number of shifted peaks appearing in the candidate spectrum relative to the test spectrum.
Yurtkuran, Alkın; Emel, Erdal
2016-01-01
The artificial bee colony (ABC) algorithm is a popular swarm based technique, which is inspired from the intelligent foraging behavior of honeybee swarms. This paper proposes a new variant of ABC algorithm, namely, enhanced ABC with solution acceptance rule and probabilistic multisearch (ABC-SA) to address global optimization problems. A new solution acceptance rule is proposed where, instead of greedy selection between old solution and new candidate solution, worse candidate solutions have a probability to be accepted. Additionally, the acceptance probability of worse candidates is nonlinearly decreased throughout the search process adaptively. Moreover, in order to improve the performance of the ABC and balance the intensification and diversification, a probabilistic multisearch strategy is presented. Three different search equations with distinctive characters are employed using predetermined search probabilities. By implementing a new solution acceptance rule and a probabilistic multisearch approach, the intensification and diversification performance of the ABC algorithm is improved. The proposed algorithm has been tested on well-known benchmark functions of varying dimensions by comparing against novel ABC variants, as well as several recent state-of-the-art algorithms. Computational results show that the proposed ABC-SA outperforms other ABC variants and is superior to state-of-the-art algorithms proposed in the literature.
Ancestral Asian source(s) of new world Y-chromosome founder haplotypes.
Karafet, T M; Zegura, S L; Posukh, O; Osipova, L; Bergen, A; Long, J; Goldman, D; Klitz, W; Harihara, S; de Knijff, P; Wiebe, V; Griffiths, R C; Templeton, A R; Hammer, M F
1999-01-01
Haplotypes constructed from Y-chromosome markers were used to trace the origins of Native Americans. Our sample consisted of 2,198 males from 60 global populations, including 19 Native American and 15 indigenous North Asian groups. A set of 12 biallelic polymorphisms gave rise to 14 unique Y-chromosome haplotypes that were unevenly distributed among the populations. Combining multiallelic variation at two Y-linked microsatellites (DYS19 and DXYS156Y) with the unique haplotypes results in a total of 95 combination haplotypes. Contra previous findings based on Y- chromosome data, our new results suggest the possibility of more than one Native American paternal founder haplotype. We postulate that, of the nine unique haplotypes found in Native Americans, haplotypes 1C and 1F are the best candidates for major New World founder haplotypes, whereas haplotypes 1B, 1I, and 1U may either be founder haplotypes and/or have arrived in the New World via recent admixture. Two of the other four haplotypes (YAP+ haplotypes 4 and 5) are probably present because of post-Columbian admixture, whereas haplotype 1G may have originated in the New World, and the Old World source of the final New World haplotype (1D) remains unresolved. The contrasting distribution patterns of the two major candidate founder haplotypes in Asia and the New World, as well as the results of a nested cladistic analysis, suggest the possibility of more than one paternal migration from the general region of Lake Baikal to the Americas. PMID:10053017
DOE Office of Scientific and Technical Information (OSTI.GOV)
Malo, Lison; Artigau, Étienne; Doyon, René
2014-06-10
Based on high-resolution spectra obtained with PHOENIX at Gemini-South, CRIRES at VLT-UT1, and ESPaDOnS at the Canada-France-Hawaii Telescope, we present new measurements of the radial and projected rotational velocities of 219 low-mass stars. The target likely membership was initially established using the Bayesian analysis tool recently presented in Malo et al., taking into account only the position, proper motion, and photometry of the stars to assess their membership probability. In the present study, we include radial velocity as an additional input to our analysis, and in doing so we confirm the high membership probability for 130 candidates: 27 in βmore » Pictoris, 22 in Tucana-Horologium, 25 in Columba, 7 in Carina, 18 in Argus and 18 in AB Doradus, and 13 with an ambiguous membership. Our analysis also confirms the membership of 57 stars proposed in the literature. A subsample of 16 candidates was observed at 3 or more epochs, allowing us to discover 6 new spectroscopic binaries. The fraction of binaries in our sample is 25%, consistent with values in the literature. Of the stars in our sample, 20% show projected rotational velocities (vsin i) higher than 30 km s{sup –1} and therefore are considered as fast rotators. A parallax and other youth indicators are still needed to fully confirm the 130 highly probable candidates identified here as new bona fide members. Finally, based on the X-ray emission of bona fide and highly probable group members, we show that for low-mass stars in the 12-120 Myr age range, the X-ray luminosity is an excellent indicator of youth and better than the more traditionally used R {sub X} parameter, the ratio of X-ray to bolometric luminosity.« less
NASA Astrophysics Data System (ADS)
Shioiri, Tetsu; Asari, Naoki; Sato, Junichi; Sasage, Kosuke; Yokokura, Kunio; Homma, Mitsutaka; Suzuki, Katsumi
To investigate the reliability of equipment of vacuum insulation, a study was carried out to clarify breakdown probability distributions in vacuum gap. Further, a double-break vacuum circuit breaker was investigated for breakdown probability distribution. The test results show that the breakdown probability distribution of the vacuum gap can be represented by a Weibull distribution using a location parameter, which shows the voltage that permits a zero breakdown probability. The location parameter obtained from Weibull plot depends on electrode area. The shape parameter obtained from Weibull plot of vacuum gap was 10∼14, and is constant irrespective non-uniform field factor. The breakdown probability distribution after no-load switching can be represented by Weibull distribution using a location parameter. The shape parameter after no-load switching was 6∼8.5, and is constant, irrespective of gap length. This indicates that the scatter of breakdown voltage was increased by no-load switching. If the vacuum circuit breaker uses a double break, breakdown probability at low voltage becomes lower than single-break probability. Although potential distribution is a concern in the double-break vacuum cuicuit breaker, its insulation reliability is better than that of the single-break vacuum interrupter even if the bias of the vacuum interrupter's sharing voltage is taken into account.
The Distribution of Wolf-Rayet Stars in NGC 6744
NASA Astrophysics Data System (ADS)
Sandford, Emily; Bibby, J. L.; Zurek, D.; Crowther, P.
2013-01-01
We undertake a survey of the Wolf-Rayet population of NGC 6744, a spiral galaxy located at an estimated distance of 11.6 Mpc, in order to determine whether the distribution of this population is consistent with the distributions of various ccSNe subtypes. 242 Wolf-Rayet candidate sources are identified, 40% of which are only detected in narrow-band helium II imaging, not in broad-band imaging. The spatial distribution of WR candidates is compared to the distributions of ccSNe subtype populations in the broad-band B filter and the narrow-band Hα filter. WR stars appear to follow the Type Ic distribution in the faintest 30% of the galaxy in the B image, but follow the Ib or II distribution in the brightest regions, possibly due to the difficulty of detecting WR stars in the brightest regions. WR candidates are found to be closely associated with HII regions; however, the treatment of the residual background strongly affects the distribution, and this result must be investigated further.
Garriguet, Didier
2016-04-01
Estimates of the prevalence of adherence to physical activity guidelines in the population are generally the result of averaging individual probability of adherence based on the number of days people meet the guidelines and the number of days they are assessed. Given this number of active and inactive days (days assessed minus days active), the conditional probability of meeting the guidelines that has been used in the past is a Beta (1 + active days, 1 + inactive days) distribution assuming the probability p of a day being active is bounded by 0 and 1 and averages 50%. A change in the assumption about the distribution of p is required to better match the discrete nature of the data and to better assess the probability of adherence when the percentage of active days in the population differs from 50%. Using accelerometry data from the Canadian Health Measures Survey, the probability of adherence to physical activity guidelines is estimated using a conditional probability given the number of active and inactive days distributed as a Betabinomial(n, a + active days , β + inactive days) assuming that p is randomly distributed as Beta(a, β) where the parameters a and β are estimated by maximum likelihood. The resulting Betabinomial distribution is discrete. For children aged 6 or older, the probability of meeting physical activity guidelines 7 out of 7 days is similar to published estimates. For pre-schoolers, the Betabinomial distribution yields higher estimates of adherence to the guidelines than the Beta distribution, in line with the probability of being active on any given day. In estimating the probability of adherence to physical activity guidelines, the Betabinomial distribution has several advantages over the previously used Beta distribution. It is a discrete distribution and maximizes the richness of accelerometer data.
McKenna, Matthew T.; Wang, Shijun; Nguyen, Tan B.; Burns, Joseph E.; Petrick, Nicholas; Summers, Ronald M.
2012-01-01
Computer-aided detection (CAD) systems have been shown to improve the diagnostic performance of CT colonography (CTC) in the detection of premalignant colorectal polyps. Despite the improvement, the overall system is not optimal. CAD annotations on true lesions are incorrectly dismissed, and false positives are misinterpreted as true polyps. Here, we conduct an observer performance study utilizing distributed human intelligence in the form of anonymous knowledge workers (KWs) to investigate human performance in classifying polyp candidates under different presentation strategies. We evaluated 600 polyp candidates from 50 patients, each case having at least one polyp • 6 mm, from a large database of CTC studies. Each polyp candidate was labeled independently as a true or false polyp by 20 KWs and an expert radiologist. We asked each labeler to determine whether the candidate was a true polyp after looking at a single 3D-rendered image of the candidate and after watching a video fly-around of the candidate. We found that distributed human intelligence improved significantly when presented with the additional information in the video fly-around. We noted that performance degraded with increasing interpretation time and increasing difficulty, but distributed human intelligence performed better than our CAD classifier for “easy” and “moderate” polyp candidates. Further, we observed numerous parallels between the expert radiologist and the KWs. Both showed similar improvement in classification moving from single-image to video interpretation. Additionally, difficulty estimates obtained from the KWs using an expectation maximization algorithm correlated well with the difficulty rating assigned by the expert radiologist. Our results suggest that distributed human intelligence is a powerful tool that will aid in the development of CAD for CTC. PMID:22705287
McKenna, Matthew T; Wang, Shijun; Nguyen, Tan B; Burns, Joseph E; Petrick, Nicholas; Summers, Ronald M
2012-08-01
Computer-aided detection (CAD) systems have been shown to improve the diagnostic performance of CT colonography (CTC) in the detection of premalignant colorectal polyps. Despite the improvement, the overall system is not optimal. CAD annotations on true lesions are incorrectly dismissed, and false positives are misinterpreted as true polyps. Here, we conduct an observer performance study utilizing distributed human intelligence in the form of anonymous knowledge workers (KWs) to investigate human performance in classifying polyp candidates under different presentation strategies. We evaluated 600 polyp candidates from 50 patients, each case having at least one polyp ≥6 mm, from a large database of CTC studies. Each polyp candidate was labeled independently as a true or false polyp by 20 KWs and an expert radiologist. We asked each labeler to determine whether the candidate was a true polyp after looking at a single 3D-rendered image of the candidate and after watching a video fly-around of the candidate. We found that distributed human intelligence improved significantly when presented with the additional information in the video fly-around. We noted that performance degraded with increasing interpretation time and increasing difficulty, but distributed human intelligence performed better than our CAD classifier for "easy" and "moderate" polyp candidates. Further, we observed numerous parallels between the expert radiologist and the KWs. Both showed similar improvement in classification moving from single-image to video interpretation. Additionally, difficulty estimates obtained from the KWs using an expectation maximization algorithm correlated well with the difficulty rating assigned by the expert radiologist. Our results suggest that distributed human intelligence is a powerful tool that will aid in the development of CAD for CTC. Copyright © 2012. Published by Elsevier B.V.
Discovery of a large-scale clumpy structure of the Lynx supercluster at z[similar]1.27
NASA Astrophysics Data System (ADS)
Nakata, Fumiaki; Kodama, Tadayuki; Shimasaku, Kazuhiro; Doi, Mamoru; Furusawa, Hisanori; Hamabe, Masaru; Kimura, Masahiko; Komiyama, Yutaka; Miyazaki, Satoshi; Okamura, Sadanori; Ouchi, Masami; Sekiguchi, Maki; Yagi, Masafumi; Yasuda, Naoki
2004-07-01
We report the discovery of a probable large-scale structure composed of many galaxy clumps around the known twin clusters at z=1.26 and z=1.27 in the Lynx region. Our analysis is based on deep, panoramic, and multi-colour imaging with the Suprime-Cam on the 8.2 m Subaru telescope. We apply a photometric redshift technique to extract plausible cluster members at z˜1.27 down to ˜ M*+2.5. From the 2-D distribution of these photometrically selected galaxies, we newly identify seven candidates of galaxy groups or clusters where the surface density of red galaxies is significantly high (>5σ), in addition to the two known clusters, comprising the largest most distant supercluster ever identified.
El medio interestelar en los alrededores de la region HII Sh2-183
NASA Astrophysics Data System (ADS)
Cichowolski, S.; Cappa, C. E.; Blanco, A.; Eppens, L.; Ertini, K.; Leiva, M. M.
2017-10-01
We present a multiwavelength study of the HII region Sh2-183, located at (,) = (123.3,+3.0) at a distance of 7.0 1.5 kpc from the Sun. Based on the radio continuum data we estimated the amount of ionized gas, the electronic density, and the number of ionizing photons needed to keep the region ionized, which is important since the star/s responsible of the region was/were not detected yet. On the other hand, based on IRAS data we have analyzed the dust temperature and distribution. The Hi line data allowed the detection of a shell-like structure surrounding the ionized gas and the CO data revealed the presence of 6 molecular clouds probably related to Sh2-183, which harbor several young stellar object candidates.
Approach to the Highly Sensitized Kidney Transplant Candidate
Vranic, Gayle M.
2016-01-01
For patients with ESRD, kidney transplant offers significant survival and quality-of-life advantages compared with dialysis. But for patients seeking transplant who are highly sensitized, wait times have traditionally been long and options limited. The approach to the highly sensitized candidate for kidney transplant has changed substantially over time owing to new advances in desensitization, options for paired donor exchange (PDE), and changes to the deceased-donor allocation system. Initial evaluation should focus on determining living-donor availability because a compatible living donor is always the best option. However, for most highly sensitized candidates this scenario is unlikely. For candidates with an incompatible donor, PDE can improve the prospects of finding a compatible living donor but for many highly sensitized patients the probability of finding a match in the relatively small pools of donors in PDE programs is limited. Desensitization of a living donor/recipient pair with low levels of incompatibility is another reasonable approach. But for pairs with high levels of pathologic HLA antibodies, outcomes after desensitization for the patient and allograft are less optimal. Determining the degree of sensitization by calculated panel-reactive antibody (cPRA) is critical in counseling the highly sensitized patient on expected wait times to deceased-donor transplant. For candidates with a high likelihood of finding a compatible deceased donor in a reasonable time frame, waiting for a kidney is a good strategy. For the candidate without a living donor and with a low probability of finding a deceased-donor match, desensitization on the waiting list can be considered. The approach to the highly sensitized kidney transplant candidate must be individualized and requires careful discussion among the transplant center, patient, and referring nephrologist. PMID:26915916
Link, W.A.; Barker, R.J.
2008-01-01
Judicious choice of candidate generating distributions improves efficiency of the Metropolis-Hastings algorithm. In Bayesian applications, it is sometimes possible to identify an approximation to the target posterior distribution; this approximate posterior distribution is a good choice for candidate generation. These observations are applied to analysis of the Cormack?Jolly?Seber model and its extensions.
Spectral Models of Neutron Star Magnetospheres
NASA Technical Reports Server (NTRS)
Romani, Roger W.
1997-01-01
We revisit the association of unidentified Galactic plane EGRET sources with tracers of recent massive star formation and death. Up-to-date catalogs of OB associations, SNR's, young pulsars, H2 regions and young open clusters were used in finding counterparts for a recent list of EGRET sources. It has been argued for some time that EGRET source positions are correlated with SNR's and OB associations as a class; we extend such analyses by finding additional counterparts and assessing the probability of individual source identifications. Among the several scenarios relating EGRET sources to massive stars, we focus on young neutron stars as the origin of the gamma-ray emission. The characteristics of the candidate identifications are compared to the known gamma-ray pulsar sample and to detailed Galactic population syntheses using our outer gap pulsar model of gamma-ray emission. Both the spatial distribution and luminosity function of the candidates are in good agreement with the model predictions; we infer that young pulsars can account for the bulk of the excess low latitude EGRET sources. We show that with this identification, the gamma-ray point sources provide an important new window into the history of recent massive star death in the solar neighborhood.
On the possibility of singlet fission in crystalline quaterrylene
NASA Astrophysics Data System (ADS)
Wang, Xiaopeng; Liu, Xingyu; Cook, Cameron; Schatschneider, Bohdan; Marom, Noa
2018-05-01
Singlet fission (SF), the spontaneous down-conversion of a singlet exciton into two triplet excitons residing on neighboring molecules, is a promising route to improve organic photovoltaic (OPV) device efficiencies by harvesting two charge carriers from one photon. However, only a few materials have been discovered that exhibit intermolecular SF in the solid state, most of which are acene derivatives. Recently, there has been a growing interest in rylenes as potential SF materials. We use many-body perturbation theory in the GW approximation and the Bethe-Salpeter equation to investigate the possibility of intermolecular SF in crystalline perylene and quaterrylene. A new method is presented for determining the percent charge transfer (%CT) character of an exciton wave-function from double-Bader analysis. This enables relating exciton probability distributions to crystal packing. Based on comparison to known and predicted SF materials with respect to the energy conservation criterion (ES-2ET) and %CT, crystalline quaterrylene is a promising candidate for intermolecular SF. Furthermore, quaterrylene is attractive for OPV applications, thanks to its high stability and narrow optical gap. Perylene is not expected to exhibit SF; however, it is a promising candidate for harvesting sub-gap photons by triplet-triplet annihilation.
On the possibility of singlet fission in crystalline quaterrylene.
Wang, Xiaopeng; Liu, Xingyu; Cook, Cameron; Schatschneider, Bohdan; Marom, Noa
2018-05-14
Singlet fission (SF), the spontaneous down-conversion of a singlet exciton into two triplet excitons residing on neighboring molecules, is a promising route to improve organic photovoltaic (OPV) device efficiencies by harvesting two charge carriers from one photon. However, only a few materials have been discovered that exhibit intermolecular SF in the solid state, most of which are acene derivatives. Recently, there has been a growing interest in rylenes as potential SF materials. We use many-body perturbation theory in the GW approximation and the Bethe-Salpeter equation to investigate the possibility of intermolecular SF in crystalline perylene and quaterrylene. A new method is presented for determining the percent charge transfer (%CT) character of an exciton wave-function from double-Bader analysis. This enables relating exciton probability distributions to crystal packing. Based on comparison to known and predicted SF materials with respect to the energy conservation criterion (E S -2E T ) and %CT, crystalline quaterrylene is a promising candidate for intermolecular SF. Furthermore, quaterrylene is attractive for OPV applications, thanks to its high stability and narrow optical gap. Perylene is not expected to exhibit SF; however, it is a promising candidate for harvesting sub-gap photons by triplet-triplet annihilation.
BATSE Gamma-Ray Burst Line Search. IV. Line Candidates from the Visual Search
NASA Astrophysics Data System (ADS)
Band, D. L.; Ryder, S.; Ford, L. A.; Matteson, J. L.; Palmer, D. M.; Teegarden, B. J.; Briggs, M. S.; Paciesas, W. S.; Pendleton, G. N.; Preece, R. D.
1996-02-01
We evaluate the significance of the line candidates identified by a visual search of burst spectra from BATSE's Spectroscopy Detectors. None of the candidates satisfy our detection criteria: an F-test probability less than 10-4 for a feature in one detector and consistency among the detectors that viewed the burst. Most of the candidates are not very significant and are likely to be fluctuations. Because of the expectation of finding absorption lines, the search was biased toward absorption features. We do not have a quantitative measure of the completeness of the search, which would enable a comparison with previous missions. Therefore, a more objective computerized search has begun.
Gravitational lensing by an ensemble of isothermal galaxies
NASA Technical Reports Server (NTRS)
Katz, Neal; Paczynski, Bohdan
1987-01-01
Calculation of 28,000 models of gravitational lensing of a distant quasar by an ensemble of randomly placed galaxies, each having a singular isothermal mass distribuiton, is reported. The average surface mass density was 0.2 of the critical value in all models. It is found that the surface mass density averaged over the area of the smallest circle that encompasses the multiple images is 0.82, only slightly smaller than expected from a simple analytical model of Turner et al. (1984). The probability of getting multiple images is also as large as expected analytically. Gravitational lensing is dominated by the matter in the beam; i.e., by the beam convergence. The cases where the multiple imaging is due to asymmetry in mass distribution (i.e., due to shear) are very rare. Therefore, the observed gravitational-lens candidates for which no lensing object has been detected between the images cannot be a result of asymmetric mass distribution outside the images, at least in a model with randomly distributed galaxies. A surprisingly large number of large separations between the multiple images is found: up to 25 percent of multiple images have their angular separation 2 to 4 times larger than expected in a simple analytical model.
Deficit of Wide Binaries in the η Chamaeleontis Young Cluster
NASA Astrophysics Data System (ADS)
Brandeker, Alexis; Jayawardhana, Ray; Khavari, Parandis; Haisch, Karl E., Jr.; Mardones, Diego
2006-12-01
We have carried out a sensitive high-resolution imaging survey of stars in the young (6-8 Myr), nearby (97 pc) compact cluster around η Chamaeleontis to search for stellar and substellar companions. Our data were obtained using the NACO adaptive optics system on the ESO Very Large Telescope (VLT). Given its youth and proximity, any substellar companions are expected to be luminous, especially in the near-infrared, and thus easier to detect next to their parent stars. Here, we present VLT NACO adaptive optics imaging with companion detection limits for 17 η Cha cluster members, and follow-up VLT ISAAC near-infrared spectroscopy for companion candidates. The widest binary detected is ~0.2", corresponding to the projected separation 20 AU, despite our survey being sensitive down to substellar companions outside 0.3", and planetary-mass objects outside 0.5". This implies that the stellar companion probability outside 0.3" and the brown dwarf companion probability outside 0.5" are less than 0.16 with 95% confidence. We compare the wide binary frequency of η Cha to that of the similarly aged TW Hydrae association and estimate the statistical likelihood that the wide binary probability is equal in both groups to be less than 2×10-4. Even though the η Cha cluster is relatively dense, stellar encounters in its present configuration cannot account for the relative deficit of wide binaries. We thus conclude that the difference in wide binary probability in these two groups provides strong evidence for multiplicity properties being dependent on environment. In two appendices we derive the projected separation probability distribution for binaries, used to constrain physical separations from observed projected separations, and summarize statistical tools useful for multiplicity studies.
Dinov, Ivo D; Siegrist, Kyle; Pearl, Dennis K; Kalinin, Alexandr; Christou, Nicolas
2016-06-01
Probability distributions are useful for modeling, simulation, analysis, and inference on varieties of natural processes and physical phenomena. There are uncountably many probability distributions. However, a few dozen families of distributions are commonly defined and are frequently used in practice for problem solving, experimental applications, and theoretical studies. In this paper, we present a new computational and graphical infrastructure, the Distributome , which facilitates the discovery, exploration and application of diverse spectra of probability distributions. The extensible Distributome infrastructure provides interfaces for (human and machine) traversal, search, and navigation of all common probability distributions. It also enables distribution modeling, applications, investigation of inter-distribution relations, as well as their analytical representations and computational utilization. The entire Distributome framework is designed and implemented as an open-source, community-built, and Internet-accessible infrastructure. It is portable, extensible and compatible with HTML5 and Web2.0 standards (http://Distributome.org). We demonstrate two types of applications of the probability Distributome resources: computational research and science education. The Distributome tools may be employed to address five complementary computational modeling applications (simulation, data-analysis and inference, model-fitting, examination of the analytical, mathematical and computational properties of specific probability distributions, and exploration of the inter-distributional relations). Many high school and college science, technology, engineering and mathematics (STEM) courses may be enriched by the use of modern pedagogical approaches and technology-enhanced methods. The Distributome resources provide enhancements for blended STEM education by improving student motivation, augmenting the classical curriculum with interactive webapps, and overhauling the learning assessment protocols.
Dinov, Ivo D.; Siegrist, Kyle; Pearl, Dennis K.; Kalinin, Alexandr; Christou, Nicolas
2015-01-01
Probability distributions are useful for modeling, simulation, analysis, and inference on varieties of natural processes and physical phenomena. There are uncountably many probability distributions. However, a few dozen families of distributions are commonly defined and are frequently used in practice for problem solving, experimental applications, and theoretical studies. In this paper, we present a new computational and graphical infrastructure, the Distributome, which facilitates the discovery, exploration and application of diverse spectra of probability distributions. The extensible Distributome infrastructure provides interfaces for (human and machine) traversal, search, and navigation of all common probability distributions. It also enables distribution modeling, applications, investigation of inter-distribution relations, as well as their analytical representations and computational utilization. The entire Distributome framework is designed and implemented as an open-source, community-built, and Internet-accessible infrastructure. It is portable, extensible and compatible with HTML5 and Web2.0 standards (http://Distributome.org). We demonstrate two types of applications of the probability Distributome resources: computational research and science education. The Distributome tools may be employed to address five complementary computational modeling applications (simulation, data-analysis and inference, model-fitting, examination of the analytical, mathematical and computational properties of specific probability distributions, and exploration of the inter-distributional relations). Many high school and college science, technology, engineering and mathematics (STEM) courses may be enriched by the use of modern pedagogical approaches and technology-enhanced methods. The Distributome resources provide enhancements for blended STEM education by improving student motivation, augmenting the classical curriculum with interactive webapps, and overhauling the learning assessment protocols. PMID:27158191
Red, redder, reddest: SCUBA-2 imaging of colour-selected Herschel sources
NASA Astrophysics Data System (ADS)
Duivenvoorden, S.; Oliver, S.; Scudder, J. M.; Greenslade, J.; Riechers, D. A.; Wilkins, S. M.; Buat, V.; Chapman, S. C.; Clements, D. L.; Cooray, A.; Coppin, K. E. K.; Dannerbauer, H.; De Zotti, G.; Dunlop, J. S.; Eales, S. A.; Efstathiou, A.; Farrah, D.; Geach, J. E.; Holland, W. S.; Hurley, P. D.; Ivison, R. J.; Marchetti, L.; Petitpas, G.; Sargent, M. T.; Scott, D.; Symeonidis, M.; Vaccari, M.; Vieira, J. D.; Wang, L.; Wardlow, J.; Zemcov, M.
2018-06-01
High-redshift, luminous, dusty star-forming galaxies (DSFGs) constrain the extremity of galaxy formation theories. The most extreme are discovered through follow-up on candidates in large area surveys. Here, we present extensive 850 μm SCUBA-2 follow-up observations of 188 red DSFG candidates from the Herschel Multitiered Extragalactic Survey (HerMES) Large Mode Survey, covering 274 deg2. We detected 87 per cent with a signal-to-noise ratio >3 at 850 μm. We introduce a new method for incorporating the confusion noise in our spectral energy distribution fitting by sampling correlated flux density fluctuations from a confusion limited map. The new 850 μm data provide a better constraint on the photometric redshifts of the candidates, with photometric redshift errors decreasing from σz/(1 + z) ≈ 0.21 to 0.15. Comparison spectroscopic redshifts also found little bias (<(z - zspec)/(1 + zspec)> = 0.08). The mean photometric redshift is found to be 3.6 with a dispersion of 0.4 and we identify 21 DSFGs with a high probability of lying at z > 4. After simulating our selection effects we find number counts are consistent with phenomenological galaxy evolution models. There is a statistically significant excess of WISE-1 and SDSS sources near our red galaxies, giving a strong indication that lensing may explain some of the apparently extreme objects. Nevertheless, our sample includes examples of galaxies with the highest star formation rates in the Universe (≫103 M⊙ yr-1).
Random Partition Distribution Indexed by Pairwise Information
Dahl, David B.; Day, Ryan; Tsai, Jerry W.
2017-01-01
We propose a random partition distribution indexed by pairwise similarity information such that partitions compatible with the similarities are given more probability. The use of pairwise similarities, in the form of distances, is common in some clustering algorithms (e.g., hierarchical clustering), but we show how to use this type of information to define a prior partition distribution for flexible Bayesian modeling. A defining feature of the distribution is that it allocates probability among partitions within a given number of subsets, but it does not shift probability among sets of partitions with different numbers of subsets. Our distribution places more probability on partitions that group similar items yet keeps the total probability of partitions with a given number of subsets constant. The distribution of the number of subsets (and its moments) is available in closed-form and is not a function of the similarities. Our formulation has an explicit probability mass function (with a tractable normalizing constant) so the full suite of MCMC methods may be used for posterior inference. We compare our distribution with several existing partition distributions, showing that our formulation has attractive properties. We provide three demonstrations to highlight the features and relative performance of our distribution. PMID:29276318
A brief introduction to probability.
Di Paola, Gioacchino; Bertani, Alessandro; De Monte, Lavinia; Tuzzolino, Fabio
2018-02-01
The theory of probability has been debated for centuries: back in 1600, French mathematics used the rules of probability to place and win bets. Subsequently, the knowledge of probability has significantly evolved and is now an essential tool for statistics. In this paper, the basic theoretical principles of probability will be reviewed, with the aim of facilitating the comprehension of statistical inference. After a brief general introduction on probability, we will review the concept of the "probability distribution" that is a function providing the probabilities of occurrence of different possible outcomes of a categorical or continuous variable. Specific attention will be focused on normal distribution that is the most relevant distribution applied to statistical analysis.
Electoral Susceptibility and Entropically Driven Interactions
NASA Astrophysics Data System (ADS)
Caravan, Bassir; Levine, Gregory
2013-03-01
In the United States electoral system the election is usually decided by the electoral votes cast by a small number of ``swing states'' where the two candidates historically have roughly equal probabilities of winning. The effective value of a swing state is determined not only by the number of its electoral votes but by the frequency of its appearance in the set of winning partitions of the electoral college. Since the electoral vote values of swing states are not identical, the presence or absence of a state in a winning partition is generally correlated with the frequency of appearance of other states and, hence, their effective values. We quantify the effective value of states by an electoral susceptibility, χj, the variation of the winning probability with the ``cost'' of changing the probability of winning state j. Associating entropy with the logarithm of the number of appearances of a state within the set of winning partitions, the entropy per state (in effect, the chemical potential) is not additive and the states may be said to ``interact.'' We study χj for a simple model with a Zipf's law type distribution of electoral votes. We show that the susceptibility for small states is largest in ``one-sided'' electoral contests and smallest in close contests. This research was supported by Department of Energy DE-FG02-08ER64623, Research Corporation CC6535 (GL) and HHMI Scholar Program (BC)
Yura, Harold T; Hanson, Steen G
2012-04-01
Methods for simulation of two-dimensional signals with arbitrary power spectral densities and signal amplitude probability density functions are disclosed. The method relies on initially transforming a white noise sample set of random Gaussian distributed numbers into a corresponding set with the desired spectral distribution, after which this colored Gaussian probability distribution is transformed via an inverse transform into the desired probability distribution. In most cases the method provides satisfactory results and can thus be considered an engineering approach. Several illustrative examples with relevance for optics are given.
Deriving injury risk curves using survival analysis from biomechanical experiments.
Yoganandan, Narayan; Banerjee, Anjishnu; Hsu, Fang-Chi; Bass, Cameron R; Voo, Liming; Pintar, Frank A; Gayzik, F Scott
2016-10-03
Injury risk curves from biomechanical experimental data analysis are used in automotive studies to improve crashworthiness and advance occupant safety. Metrics such as acceleration and deflection coupled with outcomes such as fractures and anatomical disruptions from impact tests are used in simple binary regression models. As an improvement, the International Standards Organization suggested a different approach. It was based on survival analysis. While probability curves for side-impact-induced thorax and abdominal injuries and frontal impact-induced foot-ankle-leg injuries are developed using this approach, deficiencies are apparent. The objective of this study is to present an improved, robust and generalizable methodology in an attempt to resolve these issues. It includes: (a) statistical identification of the most appropriate independent variable (metric) from a pool of candidate metrics, measured and or derived during experimentation and analysis processes, based on the highest area under the receiver operator curve, (b) quantitative determination of the most optimal probability distribution based on the lowest Akaike information criterion, (c) supplementing the qualitative/visual inspection method for comparing the selected distribution with a non-parametric distribution with objective measures, (d) identification of overly influential observations using different methods, and (e) estimation of confidence intervals using techniques more appropriate to the underlying survival statistical model. These clear and quantified details can be easily implemented with commercial/open source packages. They can be used in retrospective analysis and prospective design of experiments, and in applications to different loading scenarios such as underbody blast events. The feasibility of the methodology is demonstrated using post mortem human subject experiments and 24 metrics associated with thoracic/abdominal injuries in side-impacts. Published by Elsevier Ltd.
The global impact distribution of Near-Earth objects
NASA Astrophysics Data System (ADS)
Rumpf, Clemens; Lewis, Hugh G.; Atkinson, Peter M.
2016-02-01
Asteroids that could collide with the Earth are listed on the publicly available Near-Earth object (NEO) hazard web sites maintained by the National Aeronautics and Space Administration (NASA) and the European Space Agency (ESA). The impact probability distribution of 69 potentially threatening NEOs from these lists that produce 261 dynamically distinct impact instances, or Virtual Impactors (VIs), were calculated using the Asteroid Risk Mitigation and Optimization Research (ARMOR) tool in conjunction with OrbFit. ARMOR projected the impact probability of each VI onto the surface of the Earth as a spatial probability distribution. The projection considers orbit solution accuracy and the global impact probability. The method of ARMOR is introduced and the tool is validated against two asteroid-Earth collision cases with objects 2008 TC3 and 2014 AA. In the analysis, the natural distribution of impact corridors is contrasted against the impact probability distribution to evaluate the distributions' conformity with the uniform impact distribution assumption. The distribution of impact corridors is based on the NEO population and orbital mechanics. The analysis shows that the distribution of impact corridors matches the common assumption of uniform impact distribution and the result extends the evidence base for the uniform assumption from qualitative analysis of historic impact events into the future in a quantitative way. This finding is confirmed in a parallel analysis of impact points belonging to a synthetic population of 10,006 VIs. Taking into account the impact probabilities introduced significant variation into the results and the impact probability distribution, consequently, deviates markedly from uniformity. The concept of impact probabilities is a product of the asteroid observation and orbit determination technique and, thus, represents a man-made component that is largely disconnected from natural processes. It is important to consider impact probabilities because such information represents the best estimate of where an impact might occur.
A Causal Analysis of Observed Declines in Managed Honey Bees (Apis mellifera)
Staveley, Jane P.; Law, Sheryl A.; Fairbrother, Anne; Menzie, Charles A.
2013-01-01
The European honey bee (Apis mellifera) is a highly valuable, semi-free-ranging managed agricultural species. While the number of managed hives has been increasing, declines in overwinter survival, and the onset of colony collapse disorder in 2006, precipitated a large amount of research on bees' health in an effort to isolate the causative factors. A workshop was convened during which bee experts were introduced to a formal causal analysis approach to compare 39 candidate causes against specified criteria to evaluate their relationship to the reduced overwinter survivability observed since 2006 of commercial bees used in the California almond industry. Candidate causes were categorized as probable, possible, or unlikely; several candidate causes were categorized as indeterminate due to lack of information. Due to time limitations, a full causal analysis was not completed at the workshop. In this article, examples are provided to illustrate the process and provide preliminary findings, using three candidate causes. Varroa mites plus viruses were judged to be a “probable cause” of the reduced survival, while nutrient deficiency was judged to be a “possible cause.” Neonicotinoid pesticides were judged to be “unlikely” as the sole cause of this reduced survival, although they could possibly be a contributing factor. PMID:24363549
Smith, D.R.; Rogala, J.T.; Gray, B.R.; Zigler, S.J.; Newton, T.J.
2011-01-01
Reliable estimates of abundance are needed to assess consequences of proposed habitat restoration and enhancement projects on freshwater mussels in the Upper Mississippi River (UMR). Although there is general guidance on sampling techniques for population assessment of freshwater mussels, the actual performance of sampling designs can depend critically on the population density and spatial distribution at the project site. To evaluate various sampling designs, we simulated sampling of populations, which varied in density and degree of spatial clustering. Because of logistics and costs of large river sampling and spatial clustering of freshwater mussels, we focused on adaptive and non-adaptive versions of single and two-stage sampling. The candidate designs performed similarly in terms of precision (CV) and probability of species detection for fixed sample size. Both CV and species detection were determined largely by density, spatial distribution and sample size. However, designs did differ in the rate that occupied quadrats were encountered. Occupied units had a higher probability of selection using adaptive designs than conventional designs. We used two measures of cost: sample size (i.e. number of quadrats) and distance travelled between the quadrats. Adaptive and two-stage designs tended to reduce distance between sampling units, and thus performed better when distance travelled was considered. Based on the comparisons, we provide general recommendations on the sampling designs for the freshwater mussels in the UMR, and presumably other large rivers.
Reranking candidate gene models with cross-species comparison for improved gene prediction
Liu, Qian; Crammer, Koby; Pereira, Fernando CN; Roos, David S
2008-01-01
Background Most gene finders score candidate gene models with state-based methods, typically HMMs, by combining local properties (coding potential, splice donor and acceptor patterns, etc). Competing models with similar state-based scores may be distinguishable with additional information. In particular, functional and comparative genomics datasets may help to select among competing models of comparable probability by exploiting features likely to be associated with the correct gene models, such as conserved exon/intron structure or protein sequence features. Results We have investigated the utility of a simple post-processing step for selecting among a set of alternative gene models, using global scoring rules to rerank competing models for more accurate prediction. For each gene locus, we first generate the K best candidate gene models using the gene finder Evigan, and then rerank these models using comparisons with putative orthologous genes from closely-related species. Candidate gene models with lower scores in the original gene finder may be selected if they exhibit strong similarity to probable orthologs in coding sequence, splice site location, or signal peptide occurrence. Experiments on Drosophila melanogaster demonstrate that reranking based on cross-species comparison outperforms the best gene models identified by Evigan alone, and also outperforms the comparative gene finders GeneWise and Augustus+. Conclusion Reranking gene models with cross-species comparison improves gene prediction accuracy. This straightforward method can be readily adapted to incorporate additional lines of evidence, as it requires only a ranked source of candidate gene models. PMID:18854050
NASA Astrophysics Data System (ADS)
Borucki, William J.; Koch, David G.; Basri, Gibor; Batalha, Natalie; Brown, Timothy M.; Bryson, Stephen T.; Caldwell, Douglas; Christensen-Dalsgaard, Jørgen; Cochran, William D.; DeVore, Edna; Dunham, Edward W.; Gautier, Thomas N., III; Geary, John C.; Gilliland, Ronald; Gould, Alan; Howell, Steve B.; Jenkins, Jon M.; Latham, David W.; Lissauer, Jack J.; Marcy, Geoffrey W.; Rowe, Jason; Sasselov, Dimitar; Boss, Alan; Charbonneau, David; Ciardi, David; Doyle, Laurance; Dupree, Andrea K.; Ford, Eric B.; Fortney, Jonathan; Holman, Matthew J.; Seager, Sara; Steffen, Jason H.; Tarter, Jill; Welsh, William F.; Allen, Christopher; Buchhave, Lars A.; Christiansen, Jessie L.; Clarke, Bruce D.; Das, Santanu; Désert, Jean-Michel; Endl, Michael; Fabrycky, Daniel; Fressin, Francois; Haas, Michael; Horch, Elliott; Howard, Andrew; Isaacson, Howard; Kjeldsen, Hans; Kolodziejczak, Jeffery; Kulesa, Craig; Li, Jie; Lucas, Philip W.; Machalek, Pavel; McCarthy, Donald; MacQueen, Phillip; Meibom, Søren; Miquel, Thibaut; Prsa, Andrej; Quinn, Samuel N.; Quintana, Elisa V.; Ragozzine, Darin; Sherry, William; Shporer, Avi; Tenenbaum, Peter; Torres, Guillermo; Twicken, Joseph D.; Van Cleve, Jeffrey; Walkowicz, Lucianne; Witteborn, Fred C.; Still, Martin
2011-07-01
On 2011 February 1 the Kepler mission released data for 156,453 stars observed from the beginning of the science observations on 2009 May 2 through September 16. There are 1235 planetary candidates with transit-like signatures detected in this period. These are associated with 997 host stars. Distributions of the characteristics of the planetary candidates are separated into five class sizes: 68 candidates of approximately Earth-size (R p < 1.25 R ⊕), 288 super-Earth-size (1.25 R ⊕ <= R p < 2 R ⊕), 662 Neptune-size (2 R ⊕ <= R p < 6 R ⊕), 165 Jupiter-size (6 R ⊕ <= R p < 15 R ⊕), and 19 up to twice the size of Jupiter (15 R ⊕ <= R p < 22 R ⊕). In the temperature range appropriate for the habitable zone, 54 candidates are found with sizes ranging from Earth-size to larger than that of Jupiter. Six are less than twice the size of the Earth. Over 74% of the planetary candidates are smaller than Neptune. The observed number versus size distribution of planetary candidates increases to a peak at two to three times the Earth-size and then declines inversely proportional to the area of the candidate. Our current best estimates of the intrinsic frequencies of planetary candidates, after correcting for geometric and sensitivity biases, are 5% for Earth-size candidates, 8% for super-Earth-size candidates, 18% for Neptune-size candidates, 2% for Jupiter-size candidates, and 0.1% for very large candidates; a total of 0.34 candidates per star. Multi-candidate, transiting systems are frequent; 17% of the host stars have multi-candidate systems, and 34% of all the candidates are part of multi-candidate systems.
Impact of screening for metabolic syndrome on the evaluation of obese living kidney donors.
Marcusa, Daniel P; Schaubel, Douglas E; Woodside, Kenneth J; Sung, Randall S
2018-01-01
We report our experience with metabolic syndrome screening for obese living kidney donor candidates to mitigate the long-term risk of CKD. We retrospectively reviewed 814 obese (BMI≥30) and 993 nonobese living kidney donor evaluations over 12 years. Using logistic regression, we explored interactions between social/clinical variables and candidate acceptance before and after policy implementation. Obese donor candidate acceptance decreased after metabolic syndrome screening began (56.3%, 46.3%, p < 0.01), while nonobese candidate acceptance remained similar (59.6%, 59.2%, p = 0.59). Adjusting for age, gender, race, BMI, and number of prior evaluations, acceptance of obese candidates decreased significantly more than nonobese (p = 0.025). In candidates without metabolic syndrome, there was no significant change in how age, sex, race, or BMI affected a donor candidate's probability of acceptance. Metabolic syndrome screening is a simple stratification tool for centers with liberal absolute BMI cut-offs to exclude potentially higher-risk obese candidates. Copyright © 2017 Elsevier Inc. All rights reserved.
Semantic Web Ontology and Data Integration: a Case Study in Aiding Psychiatric Drug Repurposing.
Liang, Chen; Sun, Jingchun; Tao, Cui
2015-01-01
There remain significant difficulties selecting probable candidate drugs from existing databases. We describe an ontology-oriented approach to represent the nexus between genes, drugs, phenotypes, symptoms, and diseases from multiple information sources. We also report a case study in which we attempted to explore candidate drugs effective for bipolar disorder and epilepsy. We constructed an ontology incorporating knowledge between the two diseases and performed semantic reasoning tasks with the ontology. The results suggested 48 candidate drugs that hold promise for further breakthrough. The evaluation demonstrated the validity our approach. Our approach prioritizes the candidate drugs that have potential associations among genes, phenotypes and symptoms, and thus facilitates the data integration and drug repurposing in psychiatric disorders.
Kepler Reliability and Occurrence Rates
NASA Astrophysics Data System (ADS)
Bryson, Steve
2016-10-01
The Kepler mission has produced tables of exoplanet candidates (``KOI table''), as well as tables of transit detections (``TCE table''), hosted at the Exoplanet Archive (http://exoplanetarchive.ipac.caltech.edu). Transit detections in the TCE table that are plausibly due to a transiting object are selected for inclusion in the KOI table. KOI table entries that have not been identified as false positives (FPs) or false alarms (FAs) are classified as planet candidates (PCs, Mullally et al. 2015). A subset of PCs have been confirmed as planetary transits with greater than 99% probability, but most PCs have <99% probability of being true planets. The fraction of PCs that are true transiting planets is the PC reliability rate. The overall PC population is believed to have a reliability rate >90% (Morton & Johnson 2011).
Mining of high utility-probability sequential patterns from uncertain databases
Zhang, Binbin; Fournier-Viger, Philippe; Li, Ting
2017-01-01
High-utility sequential pattern mining (HUSPM) has become an important issue in the field of data mining. Several HUSPM algorithms have been designed to mine high-utility sequential patterns (HUPSPs). They have been applied in several real-life situations such as for consumer behavior analysis and event detection in sensor networks. Nonetheless, most studies on HUSPM have focused on mining HUPSPs in precise data. But in real-life, uncertainty is an important factor as data is collected using various types of sensors that are more or less accurate. Hence, data collected in a real-life database can be annotated with existing probabilities. This paper presents a novel pattern mining framework called high utility-probability sequential pattern mining (HUPSPM) for mining high utility-probability sequential patterns (HUPSPs) in uncertain sequence databases. A baseline algorithm with three optional pruning strategies is presented to mine HUPSPs. Moroever, to speed up the mining process, a projection mechanism is designed to create a database projection for each processed sequence, which is smaller than the original database. Thus, the number of unpromising candidates can be greatly reduced, as well as the execution time for mining HUPSPs. Substantial experiments both on real-life and synthetic datasets show that the designed algorithm performs well in terms of runtime, number of candidates, memory usage, and scalability for different minimum utility and minimum probability thresholds. PMID:28742847
Statistical evaluation of vibration analysis techniques
NASA Technical Reports Server (NTRS)
Milner, G. Martin; Miller, Patrice S.
1987-01-01
An evaluation methodology is presented for a selection of candidate vibration analysis techniques applicable to machinery representative of the environmental control and life support system of advanced spacecraft; illustrative results are given. Attention is given to the statistical analysis of small sample experiments, the quantification of detection performance for diverse techniques through the computation of probability of detection versus probability of false alarm, and the quantification of diagnostic performance.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kovalev, Andrew N.
The authors describe a measurement of the top quark mass using events with two charged leptons collected by the CDF II Detector from pmore » $$\\bar{p}$$ collisions with √s = 1.96 TeV at the Fermilab Tevatron. The posterior probability distribution of the top quark pole mass is calculated using the differential cross-section for the t$$\\bar{t}$$ production and decay expressed with respect to observed leptons and jets momenta. The presence of background events in the collected sample is modeled using calculations of the differential cross-sections for major background processes. This measurement represents the first application of this method to events with two charged leptons. In a data sample with integrated luminosity of 340 pb -1, they observe 33 candidate events and measure M top = 165.2 ± 61. stat ± 3.4 syst GeV/c 2.« less
ARCHITECTURE AND DYNAMICS OF KEPLER'S CANDIDATE MULTIPLE TRANSITING PLANET SYSTEMS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lissauer, Jack J.; Jenkins, Jon M.; Borucki, William J.
About one-third of the {approx}1200 transiting planet candidates detected in the first four months of Kepler data are members of multiple candidate systems. There are 115 target stars with two candidate transiting planets, 45 with three, 8 with four, and 1 each with five and six. We characterize the dynamical properties of these candidate multi-planet systems. The distribution of observed period ratios shows that the vast majority of candidate pairs are neither in nor near low-order mean-motion resonances. Nonetheless, there are small but statistically significant excesses of candidate pairs both in resonance and spaced slightly too far apart to bemore » in resonance, particularly near the 2:1 resonance. We find that virtually all candidate systems are stable, as tested by numerical integrations that assume a nominal mass-radius relationship. Several considerations strongly suggest that the vast majority of these multi-candidate systems are true planetary systems. Using the observed multiplicity frequencies, we find that a single population of planetary systems that matches the higher multiplicities underpredicts the number of singly transiting systems. We provide constraints on the true multiplicity and mutual inclination distribution of the multi-candidate systems, revealing a population of systems with multiple super-Earth-size and Neptune-size planets with low to moderate mutual inclinations.« less
Nonadditive entropies yield probability distributions with biases not warranted by the data.
Pressé, Steve; Ghosh, Kingshuk; Lee, Julian; Dill, Ken A
2013-11-01
Different quantities that go by the name of entropy are used in variational principles to infer probability distributions from limited data. Shore and Johnson showed that maximizing the Boltzmann-Gibbs form of the entropy ensures that probability distributions inferred satisfy the multiplication rule of probability for independent events in the absence of data coupling such events. Other types of entropies that violate the Shore and Johnson axioms, including nonadditive entropies such as the Tsallis entropy, violate this basic consistency requirement. Here we use the axiomatic framework of Shore and Johnson to show how such nonadditive entropy functions generate biases in probability distributions that are not warranted by the underlying data.
ProbOnto: ontology and knowledge base of probability distributions.
Swat, Maciej J; Grenon, Pierre; Wimalaratne, Sarala
2016-09-01
Probability distributions play a central role in mathematical and statistical modelling. The encoding, annotation and exchange of such models could be greatly simplified by a resource providing a common reference for the definition of probability distributions. Although some resources exist, no suitably detailed and complex ontology exists nor any database allowing programmatic access. ProbOnto, is an ontology-based knowledge base of probability distributions, featuring more than 80 uni- and multivariate distributions with their defining functions, characteristics, relationships and re-parameterization formulas. It can be used for model annotation and facilitates the encoding of distribution-based models, related functions and quantities. http://probonto.org mjswat@ebi.ac.uk Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.
Ma, Chihua; Luciani, Timothy; Terebus, Anna; Liang, Jie; Marai, G Elisabeta
2017-02-15
Visualizing the complex probability landscape of stochastic gene regulatory networks can further biologists' understanding of phenotypic behavior associated with specific genes. We present PRODIGEN (PRObability DIstribution of GEne Networks), a web-based visual analysis tool for the systematic exploration of probability distributions over simulation time and state space in such networks. PRODIGEN was designed in collaboration with bioinformaticians who research stochastic gene networks. The analysis tool combines in a novel way existing, expanded, and new visual encodings to capture the time-varying characteristics of probability distributions: spaghetti plots over one dimensional projection, heatmaps of distributions over 2D projections, enhanced with overlaid time curves to display temporal changes, and novel individual glyphs of state information corresponding to particular peaks. We demonstrate the effectiveness of the tool through two case studies on the computed probabilistic landscape of a gene regulatory network and of a toggle-switch network. Domain expert feedback indicates that our visual approach can help biologists: 1) visualize probabilities of stable states, 2) explore the temporal probability distributions, and 3) discover small peaks in the probability landscape that have potential relation to specific diseases.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yilmaz, Şeyda, E-mail: seydayilmaz@ktu.edu.tr; Bayrak, Erdem, E-mail: erdmbyrk@gmail.com; Bayrak, Yusuf, E-mail: bayrak@ktu.edu.tr
In this study we examined and compared the three different probabilistic distribution methods for determining the best suitable model in probabilistic assessment of earthquake hazards. We analyzed a reliable homogeneous earthquake catalogue between a time period 1900-2015 for magnitude M ≥ 6.0 and estimated the probabilistic seismic hazard in the North Anatolian Fault zone (39°-41° N 30°-40° E) using three distribution methods namely Weibull distribution, Frechet distribution and three-parameter Weibull distribution. The distribution parameters suitability was evaluated Kolmogorov-Smirnov (K-S) goodness-of-fit test. We also compared the estimated cumulative probability and the conditional probabilities of occurrence of earthquakes for different elapsed timemore » using these three distribution methods. We used Easyfit and Matlab software to calculate these distribution parameters and plotted the conditional probability curves. We concluded that the Weibull distribution method was the most suitable than other distribution methods in this region.« less
Time of arrival in quantum and Bohmian mechanics
NASA Astrophysics Data System (ADS)
Leavens, C. R.
1998-08-01
In a recent paper Grot, Rovelli, and Tate (GRT) [Phys. Rev. A 54, 4676 (1996)] derived an expression for the probability distribution π(TX) of intrinsic arrival times T(X) at position x=X for a quantum particle with initial wave function ψ(x,t=0) freely evolving in one dimension. This was done by quantizing the classical expression for the time of arrival of a free particle at X, assuming a particular choice of operator ordering, and then regulating the resulting time of arrival operator. For the special case of a minimum-uncertainty-product wave packet at t=0 with average wave number
A reversible-jump Markov chain Monte Carlo algorithm for 1D inversion of magnetotelluric data
NASA Astrophysics Data System (ADS)
Mandolesi, Eric; Ogaya, Xenia; Campanyà, Joan; Piana Agostinetti, Nicola
2018-04-01
This paper presents a new computer code developed to solve the 1D magnetotelluric (MT) inverse problem using a Bayesian trans-dimensional Markov chain Monte Carlo algorithm. MT data are sensitive to the depth-distribution of rock electric conductivity (or its reciprocal, resistivity). The solution provided is a probability distribution - the so-called posterior probability distribution (PPD) for the conductivity at depth, together with the PPD of the interface depths. The PPD is sampled via a reversible-jump Markov Chain Monte Carlo (rjMcMC) algorithm, using a modified Metropolis-Hastings (MH) rule to accept or discard candidate models along the chains. As the optimal parameterization for the inversion process is generally unknown a trans-dimensional approach is used to allow the dataset itself to indicate the most probable number of parameters needed to sample the PPD. The algorithm is tested against two simulated datasets and a set of MT data acquired in the Clare Basin (County Clare, Ireland). For the simulated datasets the correct number of conductive layers at depth and the associated electrical conductivity values is retrieved, together with reasonable estimates of the uncertainties on the investigated parameters. Results from the inversion of field measurements are compared with results obtained using a deterministic method and with well-log data from a nearby borehole. The PPD is in good agreement with the well-log data, showing as a main structure a high conductive layer associated with the Clare Shale formation. In this study, we demonstrate that our new code go beyond algorithms developend using a linear inversion scheme, as it can be used: (1) to by-pass the subjective choices in the 1D parameterizations, i.e. the number of horizontal layers in the 1D parameterization, and (2) to estimate realistic uncertainties on the retrieved parameters. The algorithm is implemented using a simple MPI approach, where independent chains run on isolated CPU, to take full advantage of parallel computer architectures. In case of a large number of data, a master/slave appoach can be used, where the master CPU samples the parameter space and the slave CPUs compute forward solutions.
Incorporating Skew into RMS Surface Roughness Probability Distribution
NASA Technical Reports Server (NTRS)
Stahl, Mark T.; Stahl, H. Philip.
2013-01-01
The standard treatment of RMS surface roughness data is the application of a Gaussian probability distribution. This handling of surface roughness ignores the skew present in the surface and overestimates the most probable RMS of the surface, the mode. Using experimental data we confirm the Gaussian distribution overestimates the mode and application of an asymmetric distribution provides a better fit. Implementing the proposed asymmetric distribution into the optical manufacturing process would reduce the polishing time required to meet surface roughness specifications.
The Estimation of Tree Posterior Probabilities Using Conditional Clade Probability Distributions
Larget, Bret
2013-01-01
In this article I introduce the idea of conditional independence of separated subtrees as a principle by which to estimate the posterior probability of trees using conditional clade probability distributions rather than simple sample relative frequencies. I describe an algorithm for these calculations and software which implements these ideas. I show that these alternative calculations are very similar to simple sample relative frequencies for high probability trees but are substantially more accurate for relatively low probability trees. The method allows the posterior probability of unsampled trees to be calculated when these trees contain only clades that are in other sampled trees. Furthermore, the method can be used to estimate the total probability of the set of sampled trees which provides a measure of the thoroughness of a posterior sample. [Bayesian phylogenetics; conditional clade distributions; improved accuracy; posterior probabilities of trees.] PMID:23479066
NASA Astrophysics Data System (ADS)
Verkhodanov, O. V.; Verkhodanova, N. V.; Ulakhovich, O. S.; Solovyov, D. I.; Khabibullina, M. L.
2018-01-01
Based on the data from the Westerbork Northern Sky Survey performed at a frequency of 325 MHz in the range of right ascensions 0h ≤ α < 2h and declinations 29° < δ < 78° and using multi-frequency Planck maps, we selected candidate objects with the Sunyaev-Zeldovich effect. The list of the most probable candidates includes 381 sources. It is shown that the search for such objects can be accelerated by using a priori data on the negative level of fluctuations in the CMB map with removed low multipoles in the direction to radio sources.
NASA Astrophysics Data System (ADS)
Gernez, Pierre; Stramski, Dariusz; Darecki, Miroslaw
2011-07-01
Time series measurements of fluctuations in underwater downward irradiance, Ed, within the green spectral band (532 nm) show that the probability distribution of instantaneous irradiance varies greatly as a function of depth within the near-surface ocean under sunny conditions. Because of intense light flashes caused by surface wave focusing, the near-surface probability distributions are highly skewed to the right and are heavy tailed. The coefficients of skewness and excess kurtosis at depths smaller than 1 m can exceed 3 and 20, respectively. We tested several probability models, such as lognormal, Gumbel, Fréchet, log-logistic, and Pareto, which are potentially suited to describe the highly skewed heavy-tailed distributions. We found that the models cannot approximate with consistently good accuracy the high irradiance values within the right tail of the experimental distribution where the probability of these values is less than 10%. This portion of the distribution corresponds approximately to light flashes with Ed > 1.5?, where ? is the time-averaged downward irradiance. However, the remaining part of the probability distribution covering all irradiance values smaller than the 90th percentile can be described with a reasonable accuracy (i.e., within 20%) with a lognormal model for all 86 measurements from the top 10 m of the ocean included in this analysis. As the intensity of irradiance fluctuations decreases with depth, the probability distribution tends toward a function symmetrical around the mean like the normal distribution. For the examined data set, the skewness and excess kurtosis assumed values very close to zero at a depth of about 10 m.
Sayeski, Kristin L; Earle, Gentry A; Eslinger, R Paige; Whitenton, Jessy N
2017-04-01
Matching phonemes (speech sounds) to graphemes (letters and letter combinations) is an important aspect of decoding (translating print to speech) and encoding (translating speech to print). Yet, many teacher candidates do not receive explicit training in phoneme-grapheme correspondence. Difficulty with accurate phoneme production and/or lack of understanding of sound-symbol correspondence can make it challenging for teachers to (a) identify student errors on common assessments and (b) serve as a model for students when teaching beginning reading or providing remedial reading instruction. For students with dyslexia, lack of teacher proficiency in this area is particularly problematic. This study examined differences between two learning conditions (massed and distributed practice) on teacher candidates' development of phoneme-grapheme correspondence knowledge and skills. An experimental, pretest-posttest-delayed test design was employed with teacher candidates (n = 52) to compare a massed practice condition (one, 60-min session) to a distributed practice condition (four, 15-min sessions distributed over 4 weeks) for learning phonemes associated with letters and letter combinations. Participants in the distributed practice condition significantly outperformed participants in the massed practice condition on their ability to correctly produce phonemes associated with different letters and letter combinations. Implications for teacher preparation are discussed.
Predicting the probability of slip in gait: methodology and distribution study.
Gragg, Jared; Yang, James
2016-01-01
The likelihood of a slip is related to the available and required friction for a certain activity, here gait. Classical slip and fall analysis presumed that a walking surface was safe if the difference between the mean available and required friction coefficients exceeded a certain threshold. Previous research was dedicated to reformulating the classical slip and fall theory to include the stochastic variation of the available and required friction when predicting the probability of slip in gait. However, when predicting the probability of a slip, previous researchers have either ignored the variation in the required friction or assumed the available and required friction to be normally distributed. Also, there are no published results that actually give the probability of slip for various combinations of required and available frictions. This study proposes a modification to the equation for predicting the probability of slip, reducing the previous equation from a double-integral to a more convenient single-integral form. Also, a simple numerical integration technique is provided to predict the probability of slip in gait: the trapezoidal method. The effect of the random variable distributions on the probability of slip is also studied. It is shown that both the required and available friction distributions cannot automatically be assumed as being normally distributed. The proposed methods allow for any combination of distributions for the available and required friction, and numerical results are compared to analytical solutions for an error analysis. The trapezoidal method is shown to be highly accurate and efficient. The probability of slip is also shown to be sensitive to the input distributions of the required and available friction. Lastly, a critical value for the probability of slip is proposed based on the number of steps taken by an average person in a single day.
Integrated-Circuit Pseudorandom-Number Generator
NASA Technical Reports Server (NTRS)
Steelman, James E.; Beasley, Jeff; Aragon, Michael; Ramirez, Francisco; Summers, Kenneth L.; Knoebel, Arthur
1992-01-01
Integrated circuit produces 8-bit pseudorandom numbers from specified probability distribution, at rate of 10 MHz. Use of Boolean logic, circuit implements pseudorandom-number-generating algorithm. Circuit includes eight 12-bit pseudorandom-number generators, outputs are uniformly distributed. 8-bit pseudorandom numbers satisfying specified nonuniform probability distribution are generated by processing uniformly distributed outputs of eight 12-bit pseudorandom-number generators through "pipeline" of D flip-flops, comparators, and memories implementing conditional probabilities on zeros and ones.
On the combinatorics of sparsification.
Huang, Fenix Wd; Reidys, Christian M
2012-10-22
We study the sparsification of dynamic programming based on folding algorithms of RNA structures. Sparsification is a method that improves significantly the computation of minimum free energy (mfe) RNA structures. We provide a quantitative analysis of the sparsification of a particular decomposition rule, Λ∗. This rule splits an interval of RNA secondary and pseudoknot structures of fixed topological genus. Key for quantifying sparsifications is the size of the so called candidate sets. Here we assume mfe-structures to be specifically distributed (see Assumption 1) within arbitrary and irreducible RNA secondary and pseudoknot structures of fixed topological genus. We then present a combinatorial framework which allows by means of probabilities of irreducible sub-structures to obtain the expectation of the Λ∗-candidate set w.r.t. a uniformly random input sequence. We compute these expectations for arc-based energy models via energy-filtered generating functions (GF) in case of RNA secondary structures as well as RNA pseudoknot structures. Furthermore, for RNA secondary structures we also analyze a simplified loop-based energy model. Our combinatorial analysis is then compared to the expected number of Λ∗-candidates obtained from the folding mfe-structures. In case of the mfe-folding of RNA secondary structures with a simplified loop-based energy model our results imply that sparsification provides a significant, constant improvement of 91% (theory) to be compared to an 96% (experimental, simplified arc-based model) reduction. However, we do not observe a linear factor improvement. Finally, in case of the "full" loop-energy model we can report a reduction of 98% (experiment). Sparsification was initially attributed a linear factor improvement. This conclusion was based on the so called polymer-zeta property, which stems from interpreting polymer chains as self-avoiding walks. Subsequent findings however reveal that the O(n) improvement is not correct. The combinatorial analysis presented here shows that, assuming a specific distribution (see Assumption 1), of mfe-structures within irreducible and arbitrary structures, the expected number of Λ∗-candidates is Θ(n2). However, the constant reduction is quite significant, being in the range of 96%. We furthermore show an analogous result for the sparsification of the Λ∗-decomposition rule for RNA pseudoknotted structures of genus one. Finally we observe that the effect of sparsification is sensitive to the employed energy model.
NASA Astrophysics Data System (ADS)
Yan, Fuhan; Li, Zhaofeng; Jiang, Yichuan
2016-05-01
The issues of modeling and analyzing diffusion in social networks have been extensively studied in the last few decades. Recently, many studies focus on uncertain diffusion process. The uncertainty of diffusion process means that the diffusion probability is unpredicted because of some complex factors. For instance, the variety of individuals' opinions is an important factor that can cause uncertainty of diffusion probability. In detail, the difference between opinions can influence the diffusion probability, and then the evolution of opinions will cause the uncertainty of diffusion probability. It is known that controlling the diffusion process is important in the context of viral marketing and political propaganda. However, previous methods are hardly feasible to control the uncertain diffusion process of individual opinion. In this paper, we present suitable strategy to control this diffusion process based on the approximate estimation of the uncertain factors. We formulate a model in which the diffusion probability is influenced by the distance between opinions, and briefly discuss the properties of the diffusion model. Then, we present an optimization problem at the background of voting to show how to control this uncertain diffusion process. In detail, it is assumed that each individual can choose one of the two candidates or abstention based on his/her opinion. Then, we present strategy to set suitable initiators and their opinions so that the advantage of one candidate will be maximized at the end of diffusion. The results show that traditional influence maximization algorithms are not applicable to this problem, and our algorithm can achieve expected performance.
Prediction of Waitlist Mortality in Adult Heart Transplant Candidates: The Candidate Risk Score.
Jasseron, Carine; Legeai, Camille; Jacquelinet, Christian; Leprince, Pascal; Cantrelle, Christelle; Audry, Benoît; Porcher, Raphael; Bastien, Olivier; Dorent, Richard
2017-09-01
The cardiac allocation system in France is currently based on urgency and geography. Medical urgency is defined by therapies without considering objective patient mortality risk factors. This study aimed to develop a waitlist mortality risk score from commonly available candidate variables. The study included all patients, aged 16 years or older, registered on the national registry CRISTAL for first single-organ heart transplantation between January 2010 and December 2014. This population was randomly divided in a 2:1 ratio into derivation and validation cohorts. The association of variables at listing with 1-year waitlist death or delisting for worsening medical condition was assessed within the derivation cohort. The predictors were used to generate a candidate risk score (CRS). Validation of the CRS was performed in the validation cohort. Concordance probability estimation (CPE) was used to evaluate the discriminative capacity of the models. During the study period, 2333 patients were newly listed. The derivation (n =1 555) and the validation cohorts (n = 778) were similar. Short-term mechanical circulatory support, natriuretic peptide decile, glomerular filtration rate, and total bilirubin level were included in a simplified model and incorporated into the score. The Concordance probability estimation of the CRS was 0.73 in the derivation cohort and 0.71 in the validation cohort. The correlation between observed and expected 1-year waitlist mortality in the validation cohort was 0.87. The candidate risk score provides an accurate objective prediction of waitlist mortality. It is currently being used to develop a modified cardiac allocation system in France.
Bivariate normal, conditional and rectangular probabilities: A computer program with applications
NASA Technical Reports Server (NTRS)
Swaroop, R.; Brownlow, J. D.; Ashwworth, G. R.; Winter, W. R.
1980-01-01
Some results for the bivariate normal distribution analysis are presented. Computer programs for conditional normal probabilities, marginal probabilities, as well as joint probabilities for rectangular regions are given: routines for computing fractile points and distribution functions are also presented. Some examples from a closed circuit television experiment are included.
Using Weighted Entropy to Rank Chemicals in Quantitative High Throughput Screening Experiments
Shockley, Keith R.
2014-01-01
Quantitative high throughput screening (qHTS) experiments can simultaneously produce concentration-response profiles for thousands of chemicals. In a typical qHTS study, a large chemical library is subjected to a primary screen in order to identify candidate hits for secondary screening, validation studies or prediction modeling. Different algorithms, usually based on the Hill equation logistic model, have been used to classify compounds as active or inactive (or inconclusive). However, observed concentration-response activity relationships may not adequately fit a sigmoidal curve. Furthermore, it is unclear how to prioritize chemicals for follow-up studies given the large uncertainties that often accompany parameter estimates from nonlinear models. Weighted Shannon entropy can address these concerns by ranking compounds according to profile-specific statistics derived from estimates of the probability mass distribution of response at the tested concentration levels. This strategy can be used to rank all tested chemicals in the absence of a pre-specified model structure or the approach can complement existing activity call algorithms by ranking the returned candidate hits. The weighted entropy approach was evaluated here using data simulated from the Hill equation model. The procedure was then applied to a chemical genomics profiling data set interrogating compounds for androgen receptor agonist activity. PMID:24056003
Assessment of source probabilities for potential tsunamis affecting the U.S. Atlantic coast
Geist, E.L.; Parsons, T.
2009-01-01
Estimating the likelihood of tsunamis occurring along the U.S. Atlantic coast critically depends on knowledge of tsunami source probability. We review available information on both earthquake and landslide probabilities from potential sources that could generate local and transoceanic tsunamis. Estimating source probability includes defining both size and recurrence distributions for earthquakes and landslides. For the former distribution, source sizes are often distributed according to a truncated or tapered power-law relationship. For the latter distribution, sources are often assumed to occur in time according to a Poisson process, simplifying the way tsunami probabilities from individual sources can be aggregated. For the U.S. Atlantic coast, earthquake tsunami sources primarily occur at transoceanic distances along plate boundary faults. Probabilities for these sources are constrained from previous statistical studies of global seismicity for similar plate boundary types. In contrast, there is presently little information constraining landslide probabilities that may generate local tsunamis. Though there is significant uncertainty in tsunami source probabilities for the Atlantic, results from this study yield a comparative analysis of tsunami source recurrence rates that can form the basis for future probabilistic analyses.
Critical Thinking Tendencies among Teacher Candidates
ERIC Educational Resources Information Center
Genc, Salih Zeki
2008-01-01
The study aims to determine critical thinking tendencies among teacher candidates. 720 students from primary school teaching department (Primary School Teaching Programme, Science Teaching Programme and Pre-School Teaching Programme) form the sample of the study. When the gender and age distributions were investigated, 253 candidates are males and…
Multinomial mixture model with heterogeneous classification probabilities
Holland, M.D.; Gray, B.R.
2011-01-01
Royle and Link (Ecology 86(9):2505-2512, 2005) proposed an analytical method that allowed estimation of multinomial distribution parameters and classification probabilities from categorical data measured with error. While useful, we demonstrate algebraically and by simulations that this method yields biased multinomial parameter estimates when the probabilities of correct category classifications vary among sampling units. We address this shortcoming by treating these probabilities as logit-normal random variables within a Bayesian framework. We use Markov chain Monte Carlo to compute Bayes estimates from a simulated sample from the posterior distribution. Based on simulations, this elaborated Royle-Link model yields nearly unbiased estimates of multinomial and correct classification probability estimates when classification probabilities are allowed to vary according to the normal distribution on the logit scale or according to the Beta distribution. The method is illustrated using categorical submersed aquatic vegetation data. ?? 2010 Springer Science+Business Media, LLC.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Friar, James Lewis; Goldman, Terrance J.; Pérez-Mercader, J.
In this paper, we apply the Law of Total Probability to the construction of scale-invariant probability distribution functions (pdf's), and require that probability measures be dimensionless and unitless under a continuous change of scales. If the scale-change distribution function is scale invariant then the constructed distribution will also be scale invariant. Repeated application of this construction on an arbitrary set of (normalizable) pdf's results again in scale-invariant distributions. The invariant function of this procedure is given uniquely by the reciprocal distribution, suggesting a kind of universality. Finally, we separately demonstrate that the reciprocal distribution results uniquely from requiring maximum entropymore » for size-class distributions with uniform bin sizes.« less
Positive phase space distributions and uncertainty relations
NASA Technical Reports Server (NTRS)
Kruger, Jan
1993-01-01
In contrast to a widespread belief, Wigner's theorem allows the construction of true joint probabilities in phase space for distributions describing the object system as well as for distributions depending on the measurement apparatus. The fundamental role of Heisenberg's uncertainty relations in Schroedinger form (including correlations) is pointed out for these two possible interpretations of joint probability distributions. Hence, in order that a multivariate normal probability distribution in phase space may correspond to a Wigner distribution of a pure or a mixed state, it is necessary and sufficient that Heisenberg's uncertainty relation in Schroedinger form should be satisfied.
The Top 10 List of Gravitational Lens Candidates from the HUBBLE SPACE TELESCOPE Medium Deep Survey
NASA Astrophysics Data System (ADS)
Ratnatunga, Kavan U.; Griffiths, Richard E.; Ostrander, Eric J.
1999-05-01
A total of 10 good candidates for gravitational lensing have been discovered in the WFPC2 images from the Hubble Space Telescope (HST) Medium Deep Survey (MDS) and archival primary observations. These candidate lenses are unique HST discoveries, i.e., they are faint systems with subarcsecond separations between the lensing objects and the lensed source images. Most of them are difficult objects for ground-based spectroscopic confirmation or for measurement of the lens and source redshifts. Seven are ``strong lens'' candidates that appear to have multiple images of the source. Three are cases in which the single image of the source galaxy has been significantly distorted into an arc. The first two quadruply lensed candidates were reported by Ratnatunga et al. We report on the subsequent eight candidates and describe them with simple models based on the assumption of singular isothermal potentials. Residuals from the simple models for some of the candidates indicate that a more complex model for the potential will probably be required to explain the full structural detail of the observations once they are confirmed to be lenses. We also discuss the effective survey area that was searched for these candidate lens objects.
The population of faint Jupiter family comets near the Earth
NASA Astrophysics Data System (ADS)
Fernández, Julio A.; Morbidelli, Alessandro
2006-11-01
We study the population of faint Jupiter family comets (JFCs) that approach the Earth (perihelion distances q<1.3 AU) by applying a debiasing technique to the observed sample. We found for the debiased cumulative luminosity function (CLF) of absolute total magnitudes H a bimodal distribution in which brighter comets ( H≲9) follow a linear relation with a steep slope α=0.65±0.14, while fainter comets follow a much shallower slope α=0.25±0.06 down to H˜18. The slope can be pushed up to α=0.35±0.09 if a second break in the H distribution to a much shallower slope is introduced at H˜16. We estimate a population of about 10 3 faint JFCs with q<1.3 AU and 10
Ubiquity of Benford's law and emergence of the reciprocal distribution
Friar, James Lewis; Goldman, Terrance J.; Pérez-Mercader, J.
2016-04-07
In this paper, we apply the Law of Total Probability to the construction of scale-invariant probability distribution functions (pdf's), and require that probability measures be dimensionless and unitless under a continuous change of scales. If the scale-change distribution function is scale invariant then the constructed distribution will also be scale invariant. Repeated application of this construction on an arbitrary set of (normalizable) pdf's results again in scale-invariant distributions. The invariant function of this procedure is given uniquely by the reciprocal distribution, suggesting a kind of universality. Finally, we separately demonstrate that the reciprocal distribution results uniquely from requiring maximum entropymore » for size-class distributions with uniform bin sizes.« less
NASA Astrophysics Data System (ADS)
Li, Y.; Gong, H.; Zhu, L.; Guo, L.; Gao, M.; Zhou, C.
2016-12-01
Continuous over-exploitation of groundwater causes dramatic drawdown, and leads to regional land subsidence in the Huairou Emergency Water Resources region, which is located in the up-middle part of the Chaobai river basin of Beijing. Owing to the spatial heterogeneity of strata's lithofacies of the alluvial fan, ground deformation has no significant positive correlation with groundwater drawdown, and one of the challenges ahead is to quantify the spatial distribution of strata's lithofacies. The transition probability geostatistics approach provides potential for characterizing the distribution of heterogeneous lithofacies in the subsurface. Combined the thickness of clay layer extracted from the simulation, with deformation field acquired from PS-InSAR technology, the influence of strata's lithofacies on land subsidence can be analyzed quantitatively. The strata's lithofacies derived from borehole data were generalized into four categories and their probability distribution in the observe space was mined by using the transition probability geostatistics, of which clay was the predominant compressible material. Geologically plausible realizations of lithofacies distribution were produced, accounting for complex heterogeneity in alluvial plain. At a particular probability level of more than 40 percent, the volume of clay defined was 55 percent of the total volume of strata's lithofacies. This level, equaling nearly the volume of compressible clay derived from the geostatistics, was thus chosen to represent the boundary between compressible and uncompressible material. The method incorporates statistical geological information, such as distribution proportions, average lengths and juxtaposition tendencies of geological types, mainly derived from borehole data and expert knowledge, into the Markov chain model of transition probability. Some similarities of patterns were indicated between the spatial distribution of deformation field and clay layer. In the area with roughly similar water table decline, locations in the subsurface having a higher probability for the existence of compressible material occur more than that in the location with a lower probability. Such estimate of spatial probability distribution is useful to analyze the uncertainty of land subsidence.
The exact probability distribution of the rank product statistics for replicated experiments.
Eisinga, Rob; Breitling, Rainer; Heskes, Tom
2013-03-18
The rank product method is a widely accepted technique for detecting differentially regulated genes in replicated microarray experiments. To approximate the sampling distribution of the rank product statistic, the original publication proposed a permutation approach, whereas recently an alternative approximation based on the continuous gamma distribution was suggested. However, both approximations are imperfect for estimating small tail probabilities. In this paper we relate the rank product statistic to number theory and provide a derivation of its exact probability distribution and the true tail probabilities. Copyright © 2013 Federation of European Biochemical Societies. Published by Elsevier B.V. All rights reserved.
Modeling the probability distribution of peak discharge for infiltrating hillslopes
NASA Astrophysics Data System (ADS)
Baiamonte, Giorgio; Singh, Vijay P.
2017-07-01
Hillslope response plays a fundamental role in the prediction of peak discharge at the basin outlet. The peak discharge for the critical duration of rainfall and its probability distribution are needed for designing urban infrastructure facilities. This study derives the probability distribution, denoted as GABS model, by coupling three models: (1) the Green-Ampt model for computing infiltration, (2) the kinematic wave model for computing discharge hydrograph from the hillslope, and (3) the intensity-duration-frequency (IDF) model for computing design rainfall intensity. The Hortonian mechanism for runoff generation is employed for computing the surface runoff hydrograph. Since the antecedent soil moisture condition (ASMC) significantly affects the rate of infiltration, its effect on the probability distribution of peak discharge is investigated. Application to a watershed in Sicily, Italy, shows that with the increase of probability, the expected effect of ASMC to increase the maximum discharge diminishes. Only for low values of probability, the critical duration of rainfall is influenced by ASMC, whereas its effect on the peak discharge seems to be less for any probability. For a set of parameters, the derived probability distribution of peak discharge seems to be fitted by the gamma distribution well. Finally, an application to a small watershed, with the aim to test the possibility to arrange in advance the rational runoff coefficient tables to be used for the rational method, and a comparison between peak discharges obtained by the GABS model with those measured in an experimental flume for a loamy-sand soil were carried out.
Constructing inverse probability weights for continuous exposures: a comparison of methods.
Naimi, Ashley I; Moodie, Erica E M; Auger, Nathalie; Kaufman, Jay S
2014-03-01
Inverse probability-weighted marginal structural models with binary exposures are common in epidemiology. Constructing inverse probability weights for a continuous exposure can be complicated by the presence of outliers, and the need to identify a parametric form for the exposure and account for nonconstant exposure variance. We explored the performance of various methods to construct inverse probability weights for continuous exposures using Monte Carlo simulation. We generated two continuous exposures and binary outcomes using data sampled from a large empirical cohort. The first exposure followed a normal distribution with homoscedastic variance. The second exposure followed a contaminated Poisson distribution, with heteroscedastic variance equal to the conditional mean. We assessed six methods to construct inverse probability weights using: a normal distribution, a normal distribution with heteroscedastic variance, a truncated normal distribution with heteroscedastic variance, a gamma distribution, a t distribution (1, 3, and 5 degrees of freedom), and a quantile binning approach (based on 10, 15, and 20 exposure categories). We estimated the marginal odds ratio for a single-unit increase in each simulated exposure in a regression model weighted by the inverse probability weights constructed using each approach, and then computed the bias and mean squared error for each method. For the homoscedastic exposure, the standard normal, gamma, and quantile binning approaches performed best. For the heteroscedastic exposure, the quantile binning, gamma, and heteroscedastic normal approaches performed best. Our results suggest that the quantile binning approach is a simple and versatile way to construct inverse probability weights for continuous exposures.
Estimating neuronal connectivity from axonal and dendritic density fields
van Pelt, Jaap; van Ooyen, Arjen
2013-01-01
Neurons innervate space by extending axonal and dendritic arborizations. When axons and dendrites come in close proximity of each other, synapses between neurons can be formed. Neurons vary greatly in their morphologies and synaptic connections with other neurons. The size and shape of the arborizations determine the way neurons innervate space. A neuron may therefore be characterized by the spatial distribution of its axonal and dendritic “mass.” A population mean “mass” density field of a particular neuron type can be obtained by averaging over the individual variations in neuron geometries. Connectivity in terms of candidate synaptic contacts between neurons can be determined directly on the basis of their arborizations but also indirectly on the basis of their density fields. To decide when a candidate synapse can be formed, we previously developed a criterion defining that axonal and dendritic line pieces should cross in 3D and have an orthogonal distance less than a threshold value. In this paper, we developed new methodology for applying this criterion to density fields. We show that estimates of the number of contacts between neuron pairs calculated from their density fields are fully consistent with the number of contacts calculated from the actual arborizations. However, the estimation of the connection probability and the expected number of contacts per connection cannot be calculated directly from density fields, because density fields do not carry anymore the correlative structure in the spatial distribution of synaptic contacts. Alternatively, these two connectivity measures can be estimated from the expected number of contacts by using empirical mapping functions. The neurons used for the validation studies were generated by our neuron simulator NETMORPH. An example is given of the estimation of average connectivity and Euclidean pre- and postsynaptic distance distributions in a network of neurons represented by their population mean density fields. PMID:24324430
Optimizing exoplanet transit searches around low-mass stars with inclination constraints
NASA Astrophysics Data System (ADS)
Herrero, E.; Ribas, I.; Jordi, C.; Guinan, E. F.; Engle, S. G.
2012-01-01
Aims: We investigate a method to increase the efficiency of a targeted exoplanet search with the transit technique by preselecting a subset of candidates from large catalogs of stars. Assuming spin-orbit alignment, this can be achieved by considering stars that have a higher probability to be oriented nearly equator-on (inclination close to 90°). Methods: We used activity-rotation velocity relations for low-mass stars with a convective envelope to study the dependence of the position in the activity-vsini diagram on the stellar axis inclination. We composed a catalog of G-, K-, M-type main-sequence simulated stars using isochrones, an isotropic inclination distribution and empirical relations to obtain their rotation periods and activity indexes. Then the activity-vsini diagram was completed and statistics were applied to trace the areas containing the higher ratio of stars with inclinations above 80°. A similar statistics was applied to stars from real catalogs with log(R'HK) and vsini data to find their probability of being oriented equator-on. Results: We present our method to generate the simulated star catalog and the subsequent statistics to find the highly inclined stars from real catalogs using the activity-vsini diagram. Several catalogs from the literature are analyzed and a subsample of stars with the highest probability of being equator-on is presented. Conclusions: Assuming spin-orbit alignment, the efficiency of an exoplanet transit search in the resulting subsample of probably highly inclined stars is estimated to be two to three times higher than with a general search without preselection. Table 4 is only available at the CDS via anonymous ftp to cdsarc.u-strasbg.fr (130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/537/A147
The Chandra Xbootes Survey - IV: Mid-Infrared and Submillimeter Counterparts
NASA Astrophysics Data System (ADS)
Brown, Arianna; Mitchell-Wynne, Ketron; Cooray, Asantha R.; Nayyeri, Hooshang
2016-06-01
In this work, we use a Bayesian technique to identify mid-IR and submillimeter counterparts for 3,213 X-ray point sources detected in the Chandra XBoötes Survey so as to characterize the relationship between black hole activity and star formation in the XBoötes region. The Chandra XBoötes Survey is a 5-ks X-ray survey of the 9.3 square degree Boötes Field of the NOAO Deep Wide-Field Survey (NDWFS), a survey imaged from the optical to the near-IR. We use a likelihood ratio analysis on Spitzer-IRAC data taken from The Spitzer Deep, Wide-Field Survey (SDWFS) to determine mid-IR counterparts, and a similar method on Herschel-SPIRE sources detected at 250µm from The Herschel Multi-tiered Extragalactic Survey to determine the submillimeter counterparts. The likelihood ratio analysis (LRA) provides the probability that a(n) IRAC or SPIRE point source is the true counterpart to a Chandra source. The analysis is comprised of three parts: the normalized magnitude distributions of counterparts and background sources, and the radial probability distribution of the separation distance between the IRAC or SPIRE source and the Chandra source. Many Chandra sources have multiple prospective counterparts in each band, so additional analysis is performed to determine the identification reliability of the candidates. Identification reliability values lie between 0 and 1, and sources with identification reliability values ≥0.8 are chosen to be the true counterparts. With these results, we will consider the statistical implications of the sample's redshifts, mid-IR and submillimeter luminosities, and star formation rates.
Smolin, John A; Gambetta, Jay M; Smith, Graeme
2012-02-17
We provide an efficient method for computing the maximum-likelihood mixed quantum state (with density matrix ρ) given a set of measurement outcomes in a complete orthonormal operator basis subject to Gaussian noise. Our method works by first changing basis yielding a candidate density matrix μ which may have nonphysical (negative) eigenvalues, and then finding the nearest physical state under the 2-norm. Our algorithm takes at worst O(d(4)) for the basis change plus O(d(3)) for finding ρ where d is the dimension of the quantum state. In the special case where the measurement basis is strings of Pauli operators, the basis change takes only O(d(3)) as well. The workhorse of the algorithm is a new linear-time method for finding the closest probability distribution (in Euclidean distance) to a set of real numbers summing to one.
Use of the negative binomial-truncated Poisson distribution in thunderstorm prediction
NASA Technical Reports Server (NTRS)
Cohen, A. C.
1971-01-01
A probability model is presented for the distribution of thunderstorms over a small area given that thunderstorm events (1 or more thunderstorms) are occurring over a larger area. The model incorporates the negative binomial and truncated Poisson distributions. Probability tables for Cape Kennedy for spring, summer, and fall months and seasons are presented. The computer program used to compute these probabilities is appended.
Comparison of Deterministic and Probabilistic Radial Distribution Systems Load Flow
NASA Astrophysics Data System (ADS)
Gupta, Atma Ram; Kumar, Ashwani
2017-12-01
Distribution system network today is facing the challenge of meeting increased load demands from the industrial, commercial and residential sectors. The pattern of load is highly dependent on consumer behavior and temporal factors such as season of the year, day of the week or time of the day. For deterministic radial distribution load flow studies load is taken as constant. But, load varies continually with a high degree of uncertainty. So, there is a need to model probable realistic load. Monte-Carlo Simulation is used to model the probable realistic load by generating random values of active and reactive power load from the mean and standard deviation of the load and for solving a Deterministic Radial Load Flow with these values. The probabilistic solution is reconstructed from deterministic data obtained for each simulation. The main contribution of the work is: Finding impact of probable realistic ZIP load modeling on balanced radial distribution load flow. Finding impact of probable realistic ZIP load modeling on unbalanced radial distribution load flow. Compare the voltage profile and losses with probable realistic ZIP load modeling for balanced and unbalanced radial distribution load flow.
NASA Astrophysics Data System (ADS)
Lozovatsky, I.; Fernando, H. J. S.; Planella-Morato, J.; Liu, Zhiyu; Lee, J.-H.; Jinadasa, S. U. P.
2017-10-01
The probability distribution of turbulent kinetic energy dissipation rate in stratified ocean usually deviates from the classic lognormal distribution that has been formulated for and often observed in unstratified homogeneous layers of atmospheric and oceanic turbulence. Our measurements of vertical profiles of micro-scale shear, collected in the East China Sea, northern Bay of Bengal, to the south and east of Sri Lanka, and in the Gulf Stream region, show that the probability distributions of the dissipation rate ɛ˜r in the pycnoclines (r ˜ 1.4 m is the averaging scale) can be successfully modeled by the Burr (type XII) probability distribution. In weakly stratified boundary layers, lognormal distribution of ɛ˜r is preferable, although the Burr is an acceptable alternative. The skewness Skɛ and the kurtosis Kɛ of the dissipation rate appear to be well correlated in a wide range of Skɛ and Kɛ variability.
Kepler False Positive Rate & Occurrence of Earth-size and Larger Planets
NASA Astrophysics Data System (ADS)
Fressin, Francois; Torres, G.; Charbonneau, D.; Kepler Team
2013-01-01
We model the Kepler exoplanet survey targets and their background stars to estimate the occurrence of astrophysical configurations which could mimic an exoplanetary transit. Using real noise level estimates, we compute the number and the characteristics of detectable eclipsing pairs involving stars or planets. We select the fraction of those that would pass the Kepler candidate vetting procedure, including the modeling of the centroid shift of their position on the Kepler camera. By comparing their distribution with that of the Kepler Object Interests from the first 6 quarters of Kepler data, we quantify the false positive rate of Kepler, as a function of candidate planet size and period. Most importantly, this approach allows quantifying and characterizing the distribution of planets, with no assumption of any prior, as the remaining population of the Kepler candidate list minus the simulated population of alternate astrophysical causes. We study the actual detection recovery rate for Kepler that allows reproducing both the KOI size and period distribution as well as their SNR distribution. We estimate the occurrence of planets down to Earth-size, and study if their frequency is correlated with their host star spectral type. This work is supported by the Spitzer General Observer Proposal #80117 - Validating the First Habitable-Zone Planet Candidates Identified by the NASA Kepler Mission, and by the Kepler Participating Scientist Contract led by David Charbonneau, to confirm the planetary nature of candidates identified by the Kepler mission
Milagro Observations of Potential TeV Emitters
NASA Technical Reports Server (NTRS)
Abdo, A. A.; Abeysekara, A. U.; Allen, B. T.; Aune, T.; Barber, A. S.; Berley, D.; Braun, J.; Chen, C.; Christopher, G. E.; DeYoung, T.;
2014-01-01
This paper reports the results from three targeted searches of Milagro TeV sky maps: two extragalactic point source lists and one pulsar source list. The first extragalactic candidate list consists of 709 candidates selected from the Fermi-LAT 2FGL catalog. The second extragalactic candidate list contains 31 candidates selected from the TeVCat source catalog that have been detected by imaging atmospheric Cherenkov telescopes (IACTs). In both extragalactic candidate lists Mkn 421 was the only source detected by Milagro. This paper presents the Milagro TeV flux for Mkn 421 and flux limits for the brighter Fermi- LAT extragalactic sources and for all TeVCat candidates. The pulsar list extends a previously published Milagro targeted search for Galactic sources. With the 32 new gamma-ray pulsars identified in 2FGL, the number of pulsars that are studied by both Fermi-LAT and Milagro is increased to 52. In this sample, we find that the probability of Milagro detecting a TeV emission coincident with a pulsar increases with the GeV flux observed by the Fermi-LAT in the energy range from 0.1 GeV to 100 GeV.
1978-03-01
for the risk of rupture for a unidirectionally laminat - ed composite subjected to pure bending. (5D This equation can be simplified further by use of...C EVALUATION OF THE THREE PARAMETER WEIBULL DISTRIBUTION FUNCTION FOR PREDICTING FRACTURE PROBABILITY IN COMPOSITE MATERIALS. THESIS / AFIT/GAE...EVALUATION OF THE THREE PARAMETER WE1BULL DISTRIBUTION FUNCTION FOR PREDICTING FRACTURE PROBABILITY IN COMPOSITE MATERIALS THESIS Presented
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gagné, Jonathan; Lafrenière, David; Doyon, René
2015-01-10
We present the BANYAN All-Sky Survey (BASS) catalog, consisting of 228 new late-type (M4-L6) candidate members of nearby young moving groups (YMGs) with an expected false-positive rate of ∼13%. This sample includes 79 new candidate young brown dwarfs and 22 planetary-mass objects. These candidates were identified through the first systematic all-sky survey for late-type low-mass stars and brown dwarfs in YMGs. We cross-matched the Two Micron All Sky Survey and AllWISE catalogs outside of the galactic plane to build a sample of 98,970 potential ≥M5 dwarfs in the solar neighborhood and calculated their proper motions with typical precisions of 5-15more » mas yr{sup –1}. We selected highly probable candidate members of several YMGs from this sample using the Bayesian Analysis for Nearby Young AssociatioNs II tool (BANYAN II). We used the most probable statistical distances inferred from BANYAN II to estimate the spectral type and mass of these candidate YMG members. We used this unique sample to show tentative signs of mass segregation in the AB Doradus moving group and the Tucana-Horologium and Columba associations. The BASS sample has already been successful in identifying several new young brown dwarfs in earlier publications, and will be of great interest in studying the initial mass function of YMGs and for the search of exoplanets by direct imaging; the input sample of potential close-by ≥M5 dwarfs will be useful to study the kinematics of low-mass stars and brown dwarfs and search for new proper motion pairs.« less
Bivariate extreme value distributions
NASA Technical Reports Server (NTRS)
Elshamy, M.
1992-01-01
In certain engineering applications, such as those occurring in the analyses of ascent structural loads for the Space Transportation System (STS), some of the load variables have a lower bound of zero. Thus, the need for practical models of bivariate extreme value probability distribution functions with lower limits was identified. We discuss the Gumbel models and present practical forms of bivariate extreme probability distributions of Weibull and Frechet types with two parameters. Bivariate extreme value probability distribution functions can be expressed in terms of the marginal extremel distributions and a 'dependence' function subject to certain analytical conditions. Properties of such bivariate extreme distributions, sums and differences of paired extremals, as well as the corresponding forms of conditional distributions, are discussed. Practical estimation techniques are also given.
Karim, Mohammad Ehsanul; Platt, Robert W
2017-06-15
Correct specification of the inverse probability weighting (IPW) model is necessary for consistent inference from a marginal structural Cox model (MSCM). In practical applications, researchers are typically unaware of the true specification of the weight model. Nonetheless, IPWs are commonly estimated using parametric models, such as the main-effects logistic regression model. In practice, assumptions underlying such models may not hold and data-adaptive statistical learning methods may provide an alternative. Many candidate statistical learning approaches are available in the literature. However, the optimal approach for a given dataset is impossible to predict. Super learner (SL) has been proposed as a tool for selecting an optimal learner from a set of candidates using cross-validation. In this study, we evaluate the usefulness of a SL in estimating IPW in four different MSCM simulation scenarios, in which we varied the specification of the true weight model specification (linear and/or additive). Our simulations show that, in the presence of weight model misspecification, with a rich and diverse set of candidate algorithms, SL can generally offer a better alternative to the commonly used statistical learning approaches in terms of MSE as well as the coverage probabilities of the estimated effect in an MSCM. The findings from the simulation studies guided the application of the MSCM in a multiple sclerosis cohort from British Columbia, Canada (1995-2008), to estimate the impact of beta-interferon treatment in delaying disability progression. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.
LEAP: biomarker inference through learning and evaluating association patterns.
Jiang, Xia; Neapolitan, Richard E
2015-03-01
Single nucleotide polymorphism (SNP) high-dimensional datasets are available from Genome Wide Association Studies (GWAS). Such data provide researchers opportunities to investigate the complex genetic basis of diseases. Much of genetic risk might be due to undiscovered epistatic interactions, which are interactions in which combination of several genes affect disease. Research aimed at discovering interacting SNPs from GWAS datasets proceeded in two directions. First, tools were developed to evaluate candidate interactions. Second, algorithms were developed to search over the space of candidate interactions. Another problem when learning interacting SNPs, which has not received much attention, is evaluating how likely it is that the learned SNPs are associated with the disease. A complete system should provide this information as well. We develop such a system. Our system, called LEAP, includes a new heuristic search algorithm for learning interacting SNPs, and a Bayesian network based algorithm for computing the probability of their association. We evaluated the performance of LEAP using 100 1,000-SNP simulated datasets, each of which contains 15 SNPs involved in interactions. When learning interacting SNPs from these datasets, LEAP outperformed seven others methods. Furthermore, only SNPs involved in interactions were found to be probable. We also used LEAP to analyze real Alzheimer's disease and breast cancer GWAS datasets. We obtained interesting and new results from the Alzheimer's dataset, but limited results from the breast cancer dataset. We conclude that our results support that LEAP is a useful tool for extracting candidate interacting SNPs from high-dimensional datasets and determining their probability. © 2015 The Authors. *Genetic Epidemiology published by Wiley Periodicals, Inc.
Nathenson, Manuel; Clynne, Michael A.; Muffler, L.J. Patrick
2012-01-01
Chronologies for eruptive activity of the Lassen Volcanic Center and for eruptions from the regional mafic vents in the surrounding area of the Lassen segment of the Cascade Range are here used to estimate probabilities of future eruptions. For the regional mafic volcanism, the ages of many vents are known only within broad ranges, and two models are developed that should bracket the actual eruptive ages. These chronologies are used with exponential, Weibull, and mixed-exponential probability distributions to match the data for time intervals between eruptions. For the Lassen Volcanic Center, the probability of an eruption in the next year is 1.4x10-4 for the exponential distribution and 2.3x10-4 for the mixed exponential distribution. For the regional mafic vents, the exponential distribution gives a probability of an eruption in the next year of 6.5x10-4, but the mixed exponential distribution indicates that the current probability, 12,000 years after the last event, could be significantly lower. For the exponential distribution, the highest probability is for an eruption from a regional mafic vent. Data on areas and volumes of lava flows and domes of the Lassen Volcanic Center and of eruptions from the regional mafic vents provide constraints on the probable sizes of future eruptions. Probabilities of lava-flow coverage are similar for the Lassen Volcanic Center and for regional mafic vents, whereas the probable eruptive volumes for the mafic vents are generally smaller. Data have been compiled for large explosive eruptions (>≈ 5 km3 in deposit volume) in the Cascade Range during the past 1.2 m.y. in order to estimate probabilities of eruption. For erupted volumes >≈5 km3, the rate of occurrence since 13.6 ka is much higher than for the entire period, and we use these data to calculate the annual probability of a large eruption at 4.6x10-4. For erupted volumes ≥10 km3, the rate of occurrence has been reasonably constant from 630 ka to the present, giving more confidence in the estimate, and we use those data to calculate the annual probability of a large eruption in the next year at 1.4x10-5.
Burst wait time simulation of CALIBAN reactor at delayed super-critical state
DOE Office of Scientific and Technical Information (OSTI.GOV)
Humbert, P.; Authier, N.; Richard, B.
2012-07-01
In the past, the super prompt critical wait time probability distribution was measured on CALIBAN fast burst reactor [4]. Afterwards, these experiments were simulated with a very good agreement by solving the non-extinction probability equation [5]. Recently, the burst wait time probability distribution has been measured at CEA-Valduc on CALIBAN at different delayed super-critical states [6]. However, in the delayed super-critical case the non-extinction probability does not give access to the wait time distribution. In this case it is necessary to compute the time dependent evolution of the full neutron count number probability distribution. In this paper we present themore » point model deterministic method used to calculate the probability distribution of the wait time before a prescribed count level taking into account prompt neutrons and delayed neutron precursors. This method is based on the solution of the time dependent adjoint Kolmogorov master equations for the number of detections using the generating function methodology [8,9,10] and inverse discrete Fourier transforms. The obtained results are then compared to the measurements and Monte-Carlo calculations based on the algorithm presented in [7]. (authors)« less
Steady state, relaxation and first-passage properties of a run-and-tumble particle in one-dimension
NASA Astrophysics Data System (ADS)
Malakar, Kanaya; Jemseena, V.; Kundu, Anupam; Vijay Kumar, K.; Sabhapandit, Sanjib; Majumdar, Satya N.; Redner, S.; Dhar, Abhishek
2018-04-01
We investigate the motion of a run-and-tumble particle (RTP) in one dimension. We find the exact probability distribution of the particle with and without diffusion on the infinite line, as well as in a finite interval. In the infinite domain, this probability distribution approaches a Gaussian form in the long-time limit, as in the case of a regular Brownian particle. At intermediate times, this distribution exhibits unexpected multi-modal forms. In a finite domain, the probability distribution reaches a steady-state form with peaks at the boundaries, in contrast to a Brownian particle. We also study the relaxation to the steady-state analytically. Finally we compute the survival probability of the RTP in a semi-infinite domain with an absorbing boundary condition at the origin. In the finite interval, we compute the exit probability and the associated exit times. We provide numerical verification of our analytical results.
Fitness Probability Distribution of Bit-Flip Mutation.
Chicano, Francisco; Sutton, Andrew M; Whitley, L Darrell; Alba, Enrique
2015-01-01
Bit-flip mutation is a common mutation operator for evolutionary algorithms applied to optimize functions over binary strings. In this paper, we develop results from the theory of landscapes and Krawtchouk polynomials to exactly compute the probability distribution of fitness values of a binary string undergoing uniform bit-flip mutation. We prove that this probability distribution can be expressed as a polynomial in p, the probability of flipping each bit. We analyze these polynomials and provide closed-form expressions for an easy linear problem (Onemax), and an NP-hard problem, MAX-SAT. We also discuss a connection of the results with runtime analysis.
Ring-Shaped Microlanes and Chemical Barriers as a Platform for Probing Single-Cell Migration.
Schreiber, Christoph; Segerer, Felix J; Wagner, Ernst; Roidl, Andreas; Rädler, Joachim O
2016-05-31
Quantification and discrimination of pharmaceutical and disease-related effects on cell migration requires detailed characterization of single-cell motility. In this context, micropatterned substrates that constrain cells within defined geometries facilitate quantitative readout of locomotion. Here, we study quasi-one-dimensional cell migration in ring-shaped microlanes. We observe bimodal behavior in form of alternating states of directional migration (run state) and reorientation (rest state). Both states show exponential lifetime distributions with characteristic persistence times, which, together with the cell velocity in the run state, provide a set of parameters that succinctly describe cell motion. By introducing PEGylated barriers of different widths into the lane, we extend this description by quantifying the effects of abrupt changes in substrate chemistry on migrating cells. The transit probability decreases exponentially as a function of barrier width, thus specifying a characteristic penetration depth of the leading lamellipodia. Applying this fingerprint-like characterization of cell motion, we compare different cell lines, and demonstrate that the cancer drug candidate salinomycin affects transit probability and resting time, but not run time or run velocity. Hence, the presented assay allows to assess multiple migration-related parameters, permits detailed characterization of cell motility, and has potential applications in cell biology and advanced drug screening.
Hayer, C.-A.; Irwin, E.R.
2008-01-01
We used an information-theoretic approach to examine the variation in detection probabilities for 87 Piedmont and Coastal Plain fishes in relation to instream gravel mining in four Alabama streams of the Mobile River drainage. Biotic and abiotic variables were also included in candidate models. Detection probabilities were heterogeneous across species and varied with habitat type, stream, season, and water quality. Instream gravel mining influenced the variation in detection probabilities for 38% of the species collected, probably because it led to habitat loss and increased sedimentation. Higher detection probabilities were apparent at unmined sites than at mined sites for 78% of the species for which gravel mining was shown to influence detection probabilities, indicating potential negative impacts to these species. Physical and chemical attributes also explained the variation in detection probabilities for many species. These results indicate that anthropogenic impacts can affect detection probabilities for fishes, and such variation should be considered when developing monitoring programs or routine sampling protocols. ?? Copyright by the American Fisheries Society 2008.
ERIC Educational Resources Information Center
Conant, Darcy Lynn
2013-01-01
Stochastic understanding of probability distribution undergirds development of conceptual connections between probability and statistics and supports development of a principled understanding of statistical inference. This study investigated the impact of an instructional course intervention designed to support development of stochastic…
Probability distributions of the electroencephalogram envelope of preterm infants.
Saji, Ryoya; Hirasawa, Kyoko; Ito, Masako; Kusuda, Satoshi; Konishi, Yukuo; Taga, Gentaro
2015-06-01
To determine the stationary characteristics of electroencephalogram (EEG) envelopes for prematurely born (preterm) infants and investigate the intrinsic characteristics of early brain development in preterm infants. Twenty neurologically normal sets of EEGs recorded in infants with a post-conceptional age (PCA) range of 26-44 weeks (mean 37.5 ± 5.0 weeks) were analyzed. Hilbert transform was applied to extract the envelope. We determined the suitable probability distribution of the envelope and performed a statistical analysis. It was found that (i) the probability distributions for preterm EEG envelopes were best fitted by lognormal distributions at 38 weeks PCA or less, and by gamma distributions at 44 weeks PCA; (ii) the scale parameter of the lognormal distribution had positive correlations with PCA as well as a strong negative correlation with the percentage of low-voltage activity; (iii) the shape parameter of the lognormal distribution had significant positive correlations with PCA; (iv) the statistics of mode showed significant linear relationships with PCA, and, therefore, it was considered a useful index in PCA prediction. These statistics, including the scale parameter of the lognormal distribution and the skewness and mode derived from a suitable probability distribution, may be good indexes for estimating stationary nature in developing brain activity in preterm infants. The stationary characteristics, such as discontinuity, asymmetry, and unimodality, of preterm EEGs are well indicated by the statistics estimated from the probability distribution of the preterm EEG envelopes. Copyright © 2014 International Federation of Clinical Neurophysiology. Published by Elsevier Ireland Ltd. All rights reserved.
Flood Frequency Curves - Use of information on the likelihood of extreme floods
NASA Astrophysics Data System (ADS)
Faber, B.
2011-12-01
Investment in the infrastructure that reduces flood risk for flood-prone communities must incorporate information on the magnitude and frequency of flooding in that area. Traditionally, that information has been a probability distribution of annual maximum streamflows developed from the historical gaged record at a stream site. Practice in the United States fits a Log-Pearson type3 distribution to the annual maximum flows of an unimpaired streamflow record, using the method of moments to estimate distribution parameters. The procedure makes the assumptions that annual peak streamflow events are (1) independent, (2) identically distributed, and (3) form a representative sample of the overall probability distribution. Each of these assumptions can be challenged. We rarely have enough data to form a representative sample, and therefore must compute and display the uncertainty in the estimated flood distribution. But, is there a wet/dry cycle that makes precipitation less than independent between successive years? Are the peak flows caused by different types of events from different statistical populations? How does the watershed or climate changing over time (non-stationarity) affect the probability distribution floods? Potential approaches to avoid these assumptions vary from estimating trend and shift and removing them from early data (and so forming a homogeneous data set), to methods that estimate statistical parameters that vary with time. A further issue in estimating a probability distribution of flood magnitude (the flood frequency curve) is whether a purely statistical approach can accurately capture the range and frequency of floods that are of interest. A meteorologically-based analysis produces "probable maximum precipitation" (PMP) and subsequently a "probable maximum flood" (PMF) that attempts to describe an upper bound on flood magnitude in a particular watershed. This analysis can help constrain the upper tail of the probability distribution, well beyond the range of gaged data or even historical or paleo-flood data, which can be very important in risk analyses performed for flood risk management and dam and levee safety studies.
Taming stochastic bifurcations in fractional-order systems via noise and delayed feedback
NASA Astrophysics Data System (ADS)
Sun, Zhongkui; Zhang, Jintian; Yang, Xiaoli; Xu, Wei
2017-08-01
The dynamics in fractional-order systems have been widely studied during the past decade due to the potential applications in new materials and anomalous diffusions, but the investigations have been so far restricted to a fractional-order system without time delay(s). In this paper, we report the first study of random responses of fractional-order system coupled with noise and delayed feedback. Stochastic averaging method has been utilized to determine the stationary probability density functions (PDFs) by means of the principle of minimum mean-square error, based on which stochastic bifurcations could be identified through recognizing the shape of the PDFs. It has been found that by changing the fractional order the shape of the PDFs can switch from unimodal distribution to bimodal one, or from bimodal distribution to unimodal one, thus announcing the onset of stochastic bifurcation. Further, we have demonstrated that by merely modulating the time delay, the feedback strengths, or the noise intensity, the shapes of PDFs can transit between a single peak and a double peak. Therefore, it provides an efficient candidate to control, say, induce or suppress, the stochastic bifurcations in fractional-order systems.
MPI Enhancements in John the Ripper
NASA Astrophysics Data System (ADS)
Sykes, Edward R.; Lin, Michael; Skoczen, Wesley
2010-11-01
John the Ripper (JtR) is an open source software package commonly used by system administrators to enforce password policy. JtR is designed to attack (i.e., crack) passwords encrypted in a wide variety of commonly used formats. While parallel implementations of JtR exist, there are several limitations to them. This research reports on two distinct algorithms that enhance this password cracking tool using the Message Passing Interface. The first algorithm is a novel approach that uses numerous processors to crack one password by using an innovative approach to workload distribution. In this algorithm the candidate password is distributed to all participating processors and the word list is divided based on probability so that each processor has the same likelihood of cracking the password while eliminating overlapping operations. The second algorithm developed in this research involves dividing the passwords within a password file equally amongst available processors while ensuring load-balanced and fault-tolerant behavior. This paper describes John the Ripper, the design of these two algorithms and preliminary results. Given the same amount of time, the original JtR can crack 29 passwords, whereas our algorithms 1 and 2 can crack an additional 35 and 45 passwords respectively.
Discovery of Remote Globular Cluster Satellites of M87
NASA Astrophysics Data System (ADS)
Sparkman, Lea; Guo, Rachel; Toloba, Elisa; Guhathakurta, Puragra; Peng, Eric W.; Ferrarese, Laura; Cote, Patrick; NGVS Collaboration
2016-01-01
We present the discovery of several tens of globular clusters (GCs) in the outer regions of the giant elliptical M87, the brightest galaxy in the Virgo Cluster. These M87 GC satellites were discovered in the course of Keck/DEIMOS spectroscopic follow up of GC candidates that were identified in the Next Generation Virgo cluster Survey (NGVS). Specifically, the primary targets of this Keck spectroscopic campaign were GC satellites of early-type dwarf (dE) galaxies. However, we found that our sample contained a subset of GCs for which M87 is the most likely host. This subset is consistent with having an r^-1 power-law surface density distribution and a radial velocity distribution both centered on M87. The remote M87 GC satellites span the radial range 140 to 900 kpc, out to about a third of the Virgo Cluster's virial radius (for comparison, M87's effective radius is only 8 kpc). These M87 GC satellites are probably former satellites of other Virgo Cluster galaxies that have subsequently been cannibalized by M87.This research was supported by the National Science Foundation and the UC Santa Cruz Science Internship Program.
Net present value approaches for drug discovery.
Svennebring, Andreas M; Wikberg, Jarl Es
2013-12-01
Three dedicated approaches to the calculation of the risk-adjusted net present value (rNPV) in drug discovery projects under different assumptions are suggested. The probability of finding a candidate drug suitable for clinical development and the time to the initiation of the clinical development is assumed to be flexible in contrast to the previously used models. The rNPV of the post-discovery cash flows is calculated as the probability weighted average of the rNPV at each potential time of initiation of clinical development. Practical considerations how to set probability rates, in particular during the initiation and termination of a project is discussed.
Dinov, Ivo D.; Kamino, Scott; Bhakhrani, Bilal; Christou, Nicolas
2014-01-01
Summary Data analysis requires subtle probability reasoning to answer questions like What is the chance of event A occurring, given that event B was observed? This generic question arises in discussions of many intriguing scientific questions such as What is the probability that an adolescent weighs between 120 and 140 pounds given that they are of average height? and What is the probability of (monetary) inflation exceeding 4% and housing price index below 110? To address such problems, learning some applied, theoretical or cross-disciplinary probability concepts is necessary. Teaching such courses can be improved by utilizing modern information technology resources. Students’ understanding of multivariate distributions, conditional probabilities, correlation and causation can be significantly strengthened by employing interactive web-based science educational resources. Independent of the type of a probability course (e.g. majors, minors or service probability course, rigorous measure-theoretic, applied or statistics course) student motivation, learning experiences and knowledge retention may be enhanced by blending modern technological tools within the classical conceptual pedagogical models. We have designed, implemented and disseminated a portable open-source web-application for teaching multivariate distributions, marginal, joint and conditional probabilities using the special case of bivariate Normal distribution. A real adolescent height and weight dataset is used to demonstrate the classroom utilization of the new web-application to address problems of parameter estimation, univariate and multivariate inference. PMID:25419016
Dinov, Ivo D; Kamino, Scott; Bhakhrani, Bilal; Christou, Nicolas
2013-01-01
Data analysis requires subtle probability reasoning to answer questions like What is the chance of event A occurring, given that event B was observed? This generic question arises in discussions of many intriguing scientific questions such as What is the probability that an adolescent weighs between 120 and 140 pounds given that they are of average height? and What is the probability of (monetary) inflation exceeding 4% and housing price index below 110? To address such problems, learning some applied, theoretical or cross-disciplinary probability concepts is necessary. Teaching such courses can be improved by utilizing modern information technology resources. Students' understanding of multivariate distributions, conditional probabilities, correlation and causation can be significantly strengthened by employing interactive web-based science educational resources. Independent of the type of a probability course (e.g. majors, minors or service probability course, rigorous measure-theoretic, applied or statistics course) student motivation, learning experiences and knowledge retention may be enhanced by blending modern technological tools within the classical conceptual pedagogical models. We have designed, implemented and disseminated a portable open-source web-application for teaching multivariate distributions, marginal, joint and conditional probabilities using the special case of bivariate Normal distribution. A real adolescent height and weight dataset is used to demonstrate the classroom utilization of the new web-application to address problems of parameter estimation, univariate and multivariate inference.
Banai, Benjamin; Laustsen, Lasse; Banai, Irena Pavela; Bovan, Kosta
2018-01-01
Previous studies have shown that voters rely on sexually dimorphic traits that signal masculinity and dominance when they choose political leaders. For example, voters exert strong preferences for candidates with lower pitched voices because these candidates are perceived as stronger and more competent. Moreover, experimental studies demonstrate that conservative voters, more than liberals, prefer political candidates with traits that signal dominance, probably because conservatives are more likely to perceive the world as a threatening place and to be more attentive to dangerous and threatening contexts. In light of these findings, this study investigates whether country-level ideology influences the relationship between candidate voice pitch and electoral outcomes of real elections. Specifically, we collected voice pitch data for presidential and prime minister candidates, aggregate national ideology for the countries in which the candidates were nominated, and measures of electoral outcomes for 69 elections held across the world. In line with previous studies, we found that candidates with lower pitched voices received more votes and had greater likelihood of winning the elections. Furthermore, regression analysis revealed an interaction between candidate voice pitch, national ideology, and election type (presidential or parliamentary). That is, having a lower pitched voice was a particularly valuable asset for presidential candidates in conservative and right-leaning countries (in comparison to presidential candidates in liberal and left-leaning countries and parliamentary elections). We discuss the practical implications of these findings, and how they relate to existing research on candidates' voices, voting preferences, and democratic elections in general.
Stylized facts in internal rates of return on stock index and its derivative transactions
NASA Astrophysics Data System (ADS)
Pichl, Lukáš; Kaizoji, Taisei; Yamano, Takuya
2007-08-01
Universal features in stock markets and their derivative markets are studied by means of probability distributions in internal rates of return on buy and sell transaction pairs. Unlike the stylized facts in normalized log returns, the probability distributions for such single asset encounters incorporate the time factor by means of the internal rate of return, defined as the continuous compound interest. Resulting stylized facts are shown in the probability distributions derived from the daily series of TOPIX, S & P 500 and FTSE 100 index close values. The application of the above analysis to minute-tick data of NIKKEI 225 and its futures market, respectively, reveals an interesting difference in the behavior of the two probability distributions, in case a threshold on the minimal duration of the long position is imposed. It is therefore suggested that the probability distributions of the internal rates of return could be used for causality mining between the underlying and derivative stock markets. The highly specific discrete spectrum, which results from noise trader strategies as opposed to the smooth distributions observed for fundamentalist strategies in single encounter transactions may be useful in deducing the type of investment strategy from trading revenues of small portfolio investors.
Probabilistic Reasoning for Robustness in Automated Planning
NASA Technical Reports Server (NTRS)
Schaffer, Steven; Clement, Bradley; Chien, Steve
2007-01-01
A general-purpose computer program for planning the actions of a spacecraft or other complex system has been augmented by incorporating a subprogram that reasons about uncertainties in such continuous variables as times taken to perform tasks and amounts of resources to be consumed. This subprogram computes parametric probability distributions for time and resource variables on the basis of user-supplied models of actions and resources that they consume. The current system accepts bounded Gaussian distributions over action duration and resource use. The distributions are then combined during planning to determine the net probability distribution of each resource at any time point. In addition to a full combinatoric approach, several approximations for arriving at these combined distributions are available, including maximum-likelihood and pessimistic algorithms. Each such probability distribution can then be integrated to obtain a probability that execution of the plan under consideration would violate any constraints on the resource. The key idea is to use these probabilities of conflict to score potential plans and drive a search toward planning low-risk actions. An output plan provides a balance between the user s specified averseness to risk and other measures of optimality.
Mathematical Model to estimate the wind power using four-parameter Burr distribution
NASA Astrophysics Data System (ADS)
Liu, Sanming; Wang, Zhijie; Pan, Zhaoxu
2018-03-01
When the real probability of wind speed in the same position needs to be described, the four-parameter Burr distribution is more suitable than other distributions. This paper introduces its important properties and characteristics. Also, the application of the four-parameter Burr distribution in wind speed prediction is discussed, and the expression of probability distribution of output power of wind turbine is deduced.
SLoWPoKES-II: 100,000 WIDE BINARIES IDENTIFIED IN SDSS WITHOUT PROPER MOTIONS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dhital, Saurav; West, Andrew A.; Schluns, Kyle J.
2015-08-15
We present the Sloan Low-mass Wide Pairs of Kinematically Equivalent Stars (SLoWPoKES)-II catalog of low-mass visual binaries identified from the Sloan Digital Sky Survey (SDSS) by matching photometric distances. The candidate pairs are vetted by comparing the stellar information. The candidate pairs are vetted by comparing the stellar density at their respective Galactic positions to Monte Carlo realizations of a simulated Milky Way. In this way, we are able to identify large numbers of bona fide wide binaries without the need for proper motions. Here, 105,537 visual binaries with angular separations of ∼1–20″ were identified, each with a probability ofmore » chance alignment of ≤5%. This is the largest catalog of bona fide wide binaries to date, and it contains a diversity of systems—in mass, mass ratios, binary separations, metallicity, and evolutionary states—that should facilitate follow-up studies to characterize the properties of M dwarfs and white dwarfs. There is a subtle but definitive suggestion of multiple populations in the physical separation distribution, supporting earlier findings. We suggest that wide binaries are composed of multiple populations, most likely representing different formation modes. There are 141 M7 or later wide binary candidates, representing a seven-fold increase over the number currently known. These binaries are too wide to have been formed via the ejection mechanism. Finally, we found that 6% of spectroscopically confirmed M dwarfs are not included in the SDSS STAR catalog; they are misclassified as extended sources due to the presence of a nearby or partially resolved companion. The SLoWPoKES-II catalog is publicly available to the entire community on the World Wide Web via the Filtergraph data visualization portal.« less
Entropy Methods For Univariate Distributions in Decision Analysis
NASA Astrophysics Data System (ADS)
Abbas, Ali E.
2003-03-01
One of the most important steps in decision analysis practice is the elicitation of the decision-maker's belief about an uncertainty of interest in the form of a representative probability distribution. However, the probability elicitation process is a task that involves many cognitive and motivational biases. Alternatively, the decision-maker may provide other information about the distribution of interest, such as its moments, and the maximum entropy method can be used to obtain a full distribution subject to the given moment constraints. In practice however, decision makers cannot readily provide moments for the distribution, and are much more comfortable providing information about the fractiles of the distribution of interest or bounds on its cumulative probabilities. In this paper we present a graphical method to determine the maximum entropy distribution between upper and lower probability bounds and provide an interpretation for the shape of the maximum entropy distribution subject to fractile constraints, (FMED). We also discuss the problems with the FMED in that it is discontinuous and flat over each fractile interval. We present a heuristic approximation to a distribution if in addition to its fractiles, we also know it is continuous and work through full examples to illustrate the approach.
A time-calibrated phylogeny of southern hemisphere stoneflies: Testing for Gondwanan origins.
McCulloch, Graham A; Wallis, Graham P; Waters, Jonathan M
2016-03-01
For more than two centuries biogeographers have attempted to explain why terrestrial or freshwater lineages have geographic distributions broken by oceans, with these disjunct distributions either attributed to vicariance associated with Gondwanan fragmentation or trans-oceanic dispersal. Stoneflies (order: Plecoptera) are a widespread order of freshwater insects whose poor dispersal ability and intolerance for salt water make them ideal candidates for Gondwanan relicts - taxa whose distribution can be explained by vicariant isolation driven by the breakup of Gondwana. Here we reconstruct the phylogenetic relationships among southern hemisphere stoneflies (5 families; 86 genera) using 2864bp of mitochondrial (COI) and nuclear (18S, H3) DNA, with a calibrated relaxed molecular clock used to estimate the chronology of diversification. Our analysis suggests that largely antitropical stonefly sub-orders, Arctoperlaria (northern hemisphere) and Antarctoperlaria (southern hemisphere), were formed approximately 121Ma (95% prior probability distribution 107-143Ma), which may reflect the vicariant rifting of the supercontinent Pangaea. Subsequently, we infer that a single Arctoperlaria lineage has dispersed into southern hemisphere 76Ma (95% range 65-98Ma). The majority of divergences between South American and Australian stonefly lineages appear to coincide with the opening of Drake Passage around 40Ma, suggesting vicariant isolation of these landmasses may be responsible for these biogeographic disjunctions. In contrast, divergences between New Zealand lineages and their sister taxa appear to post-date vicariant timeframes, implying more recent dispersal events. Copyright © 2015 Elsevier Inc. All rights reserved.
Chen, Yunjie; Roux, Benoît
2014-09-21
Hybrid schemes combining the strength of molecular dynamics (MD) and Metropolis Monte Carlo (MC) offer a promising avenue to improve the sampling efficiency of computer simulations of complex systems. A number of recently proposed hybrid methods consider new configurations generated by driving the system via a non-equilibrium MD (neMD) trajectory, which are subsequently treated as putative candidates for Metropolis MC acceptance or rejection. To obey microscopic detailed balance, it is necessary to alter the momentum of the system at the beginning and/or the end of the neMD trajectory. This strict rule then guarantees that the random walk in configurational space generated by such hybrid neMD-MC algorithm will yield the proper equilibrium Boltzmann distribution. While a number of different constructs are possible, the most commonly used prescription has been to simply reverse the momenta of all the particles at the end of the neMD trajectory ("one-end momentum reversal"). Surprisingly, it is shown here that the choice of momentum reversal prescription can have a considerable effect on the rate of convergence of the hybrid neMD-MC algorithm, with the simple one-end momentum reversal encountering particularly acute problems. In these neMD-MC simulations, different regions of configurational space end up being essentially isolated from one another due to a very small transition rate between regions. In the worst-case scenario, it is almost as if the configurational space does not constitute a single communicating class that can be sampled efficiently by the algorithm, and extremely long neMD-MC simulations are needed to obtain proper equilibrium probability distributions. To address this issue, a novel momentum reversal prescription, symmetrized with respect to both the beginning and the end of the neMD trajectory ("symmetric two-ends momentum reversal"), is introduced. Illustrative simulations demonstrate that the hybrid neMD-MC algorithm robustly yields a correct equilibrium probability distribution with this prescription.
NASA Astrophysics Data System (ADS)
Chen, Yunjie; Roux, Benoît
2014-09-01
Hybrid schemes combining the strength of molecular dynamics (MD) and Metropolis Monte Carlo (MC) offer a promising avenue to improve the sampling efficiency of computer simulations of complex systems. A number of recently proposed hybrid methods consider new configurations generated by driving the system via a non-equilibrium MD (neMD) trajectory, which are subsequently treated as putative candidates for Metropolis MC acceptance or rejection. To obey microscopic detailed balance, it is necessary to alter the momentum of the system at the beginning and/or the end of the neMD trajectory. This strict rule then guarantees that the random walk in configurational space generated by such hybrid neMD-MC algorithm will yield the proper equilibrium Boltzmann distribution. While a number of different constructs are possible, the most commonly used prescription has been to simply reverse the momenta of all the particles at the end of the neMD trajectory ("one-end momentum reversal"). Surprisingly, it is shown here that the choice of momentum reversal prescription can have a considerable effect on the rate of convergence of the hybrid neMD-MC algorithm, with the simple one-end momentum reversal encountering particularly acute problems. In these neMD-MC simulations, different regions of configurational space end up being essentially isolated from one another due to a very small transition rate between regions. In the worst-case scenario, it is almost as if the configurational space does not constitute a single communicating class that can be sampled efficiently by the algorithm, and extremely long neMD-MC simulations are needed to obtain proper equilibrium probability distributions. To address this issue, a novel momentum reversal prescription, symmetrized with respect to both the beginning and the end of the neMD trajectory ("symmetric two-ends momentum reversal"), is introduced. Illustrative simulations demonstrate that the hybrid neMD-MC algorithm robustly yields a correct equilibrium probability distribution with this prescription.
Finley, B L; Scott, P K; Mayhall, D A
1994-08-01
It has recently been suggested that "standard" data distributions for key exposure variables should be developed wherever appropriate for use in probabilistic or "Monte Carlo" exposure analyses. Soil-on-skin adherence estimates represent an ideal candidate for development of a standard data distribution: There are several readily available studies which offer a consistent pattern of reported results, and more importantly, soil adherence to skin is likely to vary little from site-to-site. In this paper, we thoroughly review each of the published soil adherence studies with respect to study design, sampling, and analytical methods, and level of confidence in the reported results. Based on these studies, probability density functions (PDF) of soil adherence values were examined for different age groups and different sampling techniques. The soil adherence PDF developed from adult data was found to resemble closely the soil adherence PDF based on child data in terms of both central tendency (mean = 0.49 and 0.63 mg-soil/cm2-skin, respectively) and 95th percentile values (1.6 and 2.4 mg-soil/cm2-skin, respectively). Accordingly, a single, "standard" PDF is presented based on all data collected for all age groups. This standard PDF is lognormally distributed; the arithmetic mean and standard deviation are 0.52 +/- 0.9 mg-soil/cm2-skin. Since our review of the literature indicates that soil adherence under environmental conditions will be minimally influenced by age, sex, soil type, or particle size, this PDF should be considered applicable to all settings. The 50th and 95th percentile values of the standard PDF (0.25 and 1.7 mg-soil/cm2-skin, respectively) are very similar to recent U.S. EPA estimates of "average" and "upper-bound" soil adherence (0.2 and 1.0 mg-soil/cm2-skin, respectively).
Work probability distribution for a ferromagnet with long-ranged and short-ranged correlations
NASA Astrophysics Data System (ADS)
Bhattacharjee, J. K.; Kirkpatrick, T. R.; Sengers, J. V.
2018-04-01
Work fluctuations and work probability distributions are fundamentally different in systems with short-ranged versus long-ranged correlations. Specifically, in systems with long-ranged correlations the work distribution is extraordinarily broad compared to systems with short-ranged correlations. This difference profoundly affects the possible applicability of fluctuation theorems like the Jarzynski fluctuation theorem. The Heisenberg ferromagnet, well below its Curie temperature, is a system with long-ranged correlations in very low magnetic fields due to the presence of Goldstone modes. As the magnetic field is increased the correlations gradually become short ranged. Hence, such a ferromagnet is an ideal system for elucidating the changes of the work probability distribution as one goes from a domain with long-ranged correlations to a domain with short-ranged correlations by tuning the magnetic field. A quantitative analysis of this crossover behavior of the work probability distribution and the associated fluctuations is presented.
Grigore, Bogdan; Peters, Jaime; Hyde, Christopher; Stein, Ken
2013-11-01
Elicitation is a technique that can be used to obtain probability distribution from experts about unknown quantities. We conducted a methodology review of reports where probability distributions had been elicited from experts to be used in model-based health technology assessments. Databases including MEDLINE, EMBASE and the CRD database were searched from inception to April 2013. Reference lists were checked and citation mapping was also used. Studies describing their approach to the elicitation of probability distributions were included. Data was abstracted on pre-defined aspects of the elicitation technique. Reports were critically appraised on their consideration of the validity, reliability and feasibility of the elicitation exercise. Fourteen articles were included. Across these studies, the most marked features were heterogeneity in elicitation approach and failure to report key aspects of the elicitation method. The most frequently used approaches to elicitation were the histogram technique and the bisection method. Only three papers explicitly considered the validity, reliability and feasibility of the elicitation exercises. Judged by the studies identified in the review, reports of expert elicitation are insufficient in detail and this impacts on the perceived usability of expert-elicited probability distributions. In this context, the wider credibility of elicitation will only be improved by better reporting and greater standardisation of approach. Until then, the advantage of eliciting probability distributions from experts may be lost.
Computer simulation of random variables and vectors with arbitrary probability distribution laws
NASA Technical Reports Server (NTRS)
Bogdan, V. M.
1981-01-01
Assume that there is given an arbitrary n-dimensional probability distribution F. A recursive construction is found for a sequence of functions x sub 1 = f sub 1 (U sub 1, ..., U sub n), ..., x sub n = f sub n (U sub 1, ..., U sub n) such that if U sub 1, ..., U sub n are independent random variables having uniform distribution over the open interval (0,1), then the joint distribution of the variables x sub 1, ..., x sub n coincides with the distribution F. Since uniform independent random variables can be well simulated by means of a computer, this result allows one to simulate arbitrary n-random variables if their joint probability distribution is known.
Nuclear risk analysis of the Ulysses mission
NASA Astrophysics Data System (ADS)
Bartram, Bart W.; Vaughan, Frank R.; Englehart, Richard W., Dr.
1991-01-01
The use of a radioisotope thermoelectric generator fueled with plutonium-238 dioxide on the Space Shuttle-launched Ulysses mission implies some level of risk due to potential accidents. This paper describes the method used to quantify risks in the Ulysses mission Final Safety Analysis Report prepared for the U.S. Department of Energy. The starting point for the analysis described herein is following input of source term probability distributions from the General Electric Company. A Monte Carlo technique is used to develop probability distributions of radiological consequences for a range of accident scenarios thoughout the mission. Factors affecting radiological consequences are identified, the probability distribution of the effect of each factor determined, and the functional relationship among all the factors established. The probability distributions of all the factor effects are then combined using a Monte Carlo technique. The results of the analysis are presented in terms of complementary cumulative distribution functions (CCDF) by mission sub-phase, phase, and the overall mission. The CCDFs show the total probability that consequences (calculated health effects) would be equal to or greater than a given value.
2012-02-29
surface and Swiss roll) and real-world data sets (UCI Machine Learning Repository [12] and USPS digit handwriting data). In our experiments, we use...less than µn ( say µ = 0.8), we can first use screening technique to select µn candidate nodes, and then apply BIPS on them for further selection and...identified from node j to node i. So we can say the probability for the existence of this connection is approximately 82%. Given the probability matrix
Force Density Function Relationships in 2-D Granular Media
NASA Technical Reports Server (NTRS)
Youngquist, Robert C.; Metzger, Philip T.; Kilts, Kelly N.
2004-01-01
An integral transform relationship is developed to convert between two important probability density functions (distributions) used in the study of contact forces in granular physics. Developing this transform has now made it possible to compare and relate various theoretical approaches with one another and with the experimental data despite the fact that one may predict the Cartesian probability density and another the force magnitude probability density. Also, the transforms identify which functional forms are relevant to describe the probability density observed in nature, and so the modified Bessel function of the second kind has been identified as the relevant form for the Cartesian probability density corresponding to exponential forms in the force magnitude distribution. Furthermore, it is shown that this transform pair supplies a sufficient mathematical framework to describe the evolution of the force magnitude distribution under shearing. Apart from the choice of several coefficients, whose evolution of values must be explained in the physics, this framework successfully reproduces the features of the distribution that are taken to be an indicator of jamming and unjamming in a granular packing. Key words. Granular Physics, Probability Density Functions, Fourier Transforms
An evaluation of procedures to estimate monthly precipitation probabilities
NASA Astrophysics Data System (ADS)
Legates, David R.
1991-01-01
Many frequency distributions have been used to evaluate monthly precipitation probabilities. Eight of these distributions (including Pearson type III, extreme value, and transform normal probability density functions) are comparatively examined to determine their ability to represent accurately variations in monthly precipitation totals for global hydroclimatological analyses. Results indicate that a modified version of the Box-Cox transform-normal distribution more adequately describes the 'true' precipitation distribution than does any of the other methods. This assessment was made using a cross-validation procedure for a global network of 253 stations for which at least 100 years of monthly precipitation totals were available.
Path Diversity Improved Opportunistic Routing for Underwater Sensor Networks.
Bai, Weigang; Wang, Haiyan; He, Ke; Zhao, Ruiqin
2018-04-23
The packets carried along a pre-defined route in underwater sensor networks are very vulnerble. Node mobility or intermittent channel availability easily leads to unreachable routing. Opportunistic routing has been proven to be a promising paradigm to design routing protocols for underwater sensor networks. It takes advantage of the broadcast nature of the wireless medium to combat packet losses and selects potential paths on the fly. Finding an appropriate forwarding candidate set is a key issue in opportunistic routing. Many existing solutions ignore the impact of candidates location distribution on packet forwarding. In this paper, a path diversity improved candidate selection strategy is applied in opportunistic routing to improve packet forwarding efficiency. It not only maximizes the packet forwarding advancements but also takes the candidate’s location distribution into account. Based on this strategy, we propose two effective routing protocols: position improved candidates selection (PICS) and position random candidates selection (PRCS). PICS employs two-hop neighbor information to make routing decisions. PRCS only uses one-hop neighbor information. Simulation results show that both PICS and PRCS can significantly improve network performance when compared with the previous solutions, in terms of packet delivery ratio, average energy consumption and end-to-end delay.
q-Gaussian distributions of leverage returns, first stopping times, and default risk valuations
NASA Astrophysics Data System (ADS)
Katz, Yuri A.; Tian, Li
2013-10-01
We study the probability distributions of daily leverage returns of 520 North American industrial companies that survive de-listing during the financial crisis, 2006-2012. We provide evidence that distributions of unbiased leverage returns of all individual firms belong to the class of q-Gaussian distributions with the Tsallis entropic parameter within the interval 1
Greenwood, J. Arthur; Landwehr, J. Maciunas; Matalas, N.C.; Wallis, J.R.
1979-01-01
Distributions whose inverse forms are explicitly defined, such as Tukey's lambda, may present problems in deriving their parameters by more conventional means. Probability weighted moments are introduced and shown to be potentially useful in expressing the parameters of these distributions.
Univariate Probability Distributions
ERIC Educational Resources Information Center
Leemis, Lawrence M.; Luckett, Daniel J.; Powell, Austin G.; Vermeer, Peter E.
2012-01-01
We describe a web-based interactive graphic that can be used as a resource in introductory classes in mathematical statistics. This interactive graphic presents 76 common univariate distributions and gives details on (a) various features of the distribution such as the functional form of the probability density function and cumulative distribution…
A probability space for quantum models
NASA Astrophysics Data System (ADS)
Lemmens, L. F.
2017-06-01
A probability space contains a set of outcomes, a collection of events formed by subsets of the set of outcomes and probabilities defined for all events. A reformulation in terms of propositions allows to use the maximum entropy method to assign the probabilities taking some constraints into account. The construction of a probability space for quantum models is determined by the choice of propositions, choosing the constraints and making the probability assignment by the maximum entropy method. This approach shows, how typical quantum distributions such as Maxwell-Boltzmann, Fermi-Dirac and Bose-Einstein are partly related with well-known classical distributions. The relation between the conditional probability density, given some averages as constraints and the appropriate ensemble is elucidated.
Marui, Tetsuya; Funatogawa, Ikuko; Koishi, Shinko; Yamamoto, Kenji; Matsumoto, Hideo; Hashimoto, Ohiko; Jinde, Seiichiro; Nishida, Hisami; Sugiyama, Toshiro; Kasai, Kiyoto; Watanabe, Keiichiro; Kano, Yukiko; Kato, Nobumasa
2010-05-01
Autism is a severe neurodevelopmental disorder with a complex genetic aetiology. The wingless-type MMTV integration site family member 2 (WNT2) gene has been considered as a candidate gene for autism. We conducted a case-control study and followed up with a transmission disequilibrium test (TDT) analysis to confirm replication of the significant results for the first time. We conducted a case-control study of nine single nucleotide polymorphisms (SNPs) within the WNT2 gene in 170 patients with autism and 214 normal controls in a Japanese population. We then conducted a TDT analysis in 98 autistic families (trios) to replicate the results of the case-control study. In the case-control study, three SNPs (rs3779547, rs4727847 and rs3729629), two major individual haplotypes (A-T-C and G-G-G, consisting of rs3779547, rs4727847, and rs3729629), and global probability values of the haplotype distributions in the same region (global p=0.0091) showed significant associations with autism. Furthermore, all of these significant associations were also observed in the TDT analysis. Our findings provide evidence for a significant association between WNT2 and autism. Considering the important role of the WNT2 gene in brain development, our results therefore indicate that the WNT2 gene is one of the strong candidate genes for autism.
NASA Astrophysics Data System (ADS)
Khanmohammadi, Neda; Rezaie, Hossein; Montaseri, Majid; Behmanesh, Javad
2017-10-01
The reference evapotranspiration (ET0) plays an important role in water management plans in arid or semi-arid countries such as Iran. For this reason, the regional analysis of this parameter is important. But, ET0 process is affected by several meteorological parameters such as wind speed, solar radiation, temperature and relative humidity. Therefore, the effect of distribution type of effective meteorological variables on ET0 distribution was analyzed. For this purpose, the regional probability distribution of the annual ET0 and its effective parameters were selected. Used data in this research was recorded data at 30 synoptic stations of Iran during 1960-2014. Using the probability plot correlation coefficient (PPCC) test and the L-moment method, five common distributions were compared and the best distribution was selected. The results of PPCC test and L-moment diagram indicated that the Pearson type III distribution was the best probability distribution for fitting annual ET0 and its four effective parameters. The results of RMSE showed that the ability of the PPCC test and L-moment method for regional analysis of reference evapotranspiration and its effective parameters was similar. The results also showed that the distribution type of the parameters which affected ET0 values can affect the distribution of reference evapotranspiration.
Probabilistic reasoning in data analysis.
Sirovich, Lawrence
2011-09-20
This Teaching Resource provides lecture notes, slides, and a student assignment for a lecture on probabilistic reasoning in the analysis of biological data. General probabilistic frameworks are introduced, and a number of standard probability distributions are described using simple intuitive ideas. Particular attention is focused on random arrivals that are independent of prior history (Markovian events), with an emphasis on waiting times, Poisson processes, and Poisson probability distributions. The use of these various probability distributions is applied to biomedical problems, including several classic experimental studies.
Multiobjective fuzzy stochastic linear programming problems with inexact probability distribution
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hamadameen, Abdulqader Othman; Zainuddin, Zaitul Marlizawati
This study deals with multiobjective fuzzy stochastic linear programming problems with uncertainty probability distribution which are defined as fuzzy assertions by ambiguous experts. The problem formulation has been presented and the two solutions strategies are; the fuzzy transformation via ranking function and the stochastic transformation when α{sup –}. cut technique and linguistic hedges are used in the uncertainty probability distribution. The development of Sen’s method is employed to find a compromise solution, supported by illustrative numerical example.
Work probability distribution and tossing a biased coin
NASA Astrophysics Data System (ADS)
Saha, Arnab; Bhattacharjee, Jayanta K.; Chakraborty, Sagar
2011-01-01
We show that the rare events present in dissipated work that enters Jarzynski equality, when mapped appropriately to the phenomenon of large deviations found in a biased coin toss, are enough to yield a quantitative work probability distribution for the Jarzynski equality. This allows us to propose a recipe for constructing work probability distribution independent of the details of any relevant system. The underlying framework, developed herein, is expected to be of use in modeling other physical phenomena where rare events play an important role.
NASA Astrophysics Data System (ADS)
Jenuwine, Natalia M.; Mahesh, Sunny N.; Furst, Jacob D.; Raicu, Daniela S.
2018-02-01
Early detection of lung nodules from CT scans is key to improving lung cancer treatment, but poses a significant challenge for radiologists due to the high throughput required of them. Computer-Aided Detection (CADe) systems aim to automatically detect these nodules with computer algorithms, thus improving diagnosis. These systems typically use a candidate selection step, which identifies all objects that resemble nodules, followed by a machine learning classifier which separates true nodules from false positives. We create a CADe system that uses a 3D convolutional neural network (CNN) to detect nodules in CT scans without a candidate selection step. Using data from the LIDC database, we train a 3D CNN to analyze subvolumes from anywhere within a CT scan and output the probability that each subvolume contains a nodule. Once trained, we apply our CNN to detect nodules from entire scans, by systematically dividing the scan into overlapping subvolumes which we input into the CNN to obtain the corresponding probabilities. By enabling our network to process an entire scan, we expect to streamline the detection process while maintaining its effectiveness. Our results imply that with continued training using an iterative training scheme, the one-step approach has the potential to be highly effective.
Numeracy moderates the influence of task-irrelevant affect on probability weighting.
Traczyk, Jakub; Fulawka, Kamil
2016-06-01
Statistical numeracy, defined as the ability to understand and process statistical and probability information, plays a significant role in superior decision making. However, recent research has demonstrated that statistical numeracy goes beyond simple comprehension of numbers and mathematical operations. On the contrary to previous studies that were focused on emotions integral to risky prospects, we hypothesized that highly numerate individuals would exhibit more linear probability weighting because they would be less biased by incidental and decision-irrelevant affect. Participants were instructed to make a series of insurance decisions preceded by negative (i.e., fear-inducing) or neutral stimuli. We found that incidental negative affect increased the curvature of the probability weighting function (PWF). Interestingly, this effect was significant only for less numerate individuals, while probability weighting in more numerate people was not altered by decision-irrelevant affect. We propose two candidate mechanisms for the observed effect. Copyright © 2016 Elsevier B.V. All rights reserved.
NASA Technical Reports Server (NTRS)
Tinto, Massimo; Armstrong, J. W.
1991-01-01
Massive coalescing binary systems are candidate sources of gravitational radiation in the millihertz frequency band accessible to spacecraft Doppler tracking experiments. This paper discusses signal processing and detection probability for waves from coalescing binaries in the regime where the signal frequency increases linearly with time, i.e., 'chirp' signals. Using known noise statistics, thresholds with given false alarm probabilities are established for one- and two-spacecraft experiments. Given the threshold, the detection probability is calculated as a function of gravitational wave amplitude for both one- and two-spacecraft experiments, assuming random polarization states and under various assumptions about wave directions. This allows quantitative statements about the detection efficiency of these experiments and the utility of coincidence experiments. In particular, coincidence probabilities for two-spacecraft experiments are insensitive to the angle between the directions to the two spacecraft, indicating that near-optical experiments can be done without constraints on spacecraft trajectories.
NASA Astrophysics Data System (ADS)
McCleary, R. J.; Hassan, M. A.
2006-12-01
An automated procedure was developed to model spatial fish distributions within small streams in the Foothills of Alberta. Native fish populations and their habitats are susceptible to impacts arising from both industrial forestry and rapid development of petroleum resources in the region. Knowledge of fish distributions and the effects of industrial activities on their habitats is required to help conserve native fish populations. Resource selection function (RSF) models were used to explain presence/absence of fish in small streams. Target species were bull trout, rainbow trout and non-native brook trout. Using GIS, the drainage network was divided into reaches with uniform slope and drainage area and then polygons for each reach were created. Predictor variables described stream size, stream energy, climate and land-use. We identified a set of candidate models and selected the best model using a standard Akaike Information Criteria approach. The best models were validated with two external data sets. Drainage area and basin slope parameters were included in all best models. This finding emphasizes the importance of controlling for the energy dimension at the basin scale in investigations into the effects of land-use on aquatic resources in this transitional landscape between the mountains and plains. The best model for bull trout indicated a relation between the presence of artificial migration barriers in downstream areas and the extirpation of the species from headwater reaches. We produced reach-scale maps by species and summarized this information within all small catchments across the 12,000 km2 study area. These maps had included three categories based on predicted probability of capture for individual reaches. The high probability category had a 78 percent accuracy for correctly predicting both fish present and fish not-present reaches. Basin scale maps highlight specific watersheds likely to support both native bull trout and invasive brook trout, while reach-scale maps indicate specific reaches where interactions between these two species are likely to occur. With regional calibration, this automated modeling and mapping procedure could apply in headwater catchments throughout the Rocky Mountain Foothills and other areas where sporadic waterfalls or other natural migration barriers are not an important feature limiting fish distribution.
Warris, Sven; Boymans, Sander; Muiser, Iwe; Noback, Michiel; Krijnen, Wim; Nap, Jan-Peter
2014-01-13
Small RNAs are important regulators of genome function, yet their prediction in genomes is still a major computational challenge. Statistical analyses of pre-miRNA sequences indicated that their 2D structure tends to have a minimal free energy (MFE) significantly lower than MFE values of equivalently randomized sequences with the same nucleotide composition, in contrast to other classes of non-coding RNA. The computation of many MFEs is, however, too intensive to allow for genome-wide screenings. Using a local grid infrastructure, MFE distributions of random sequences were pre-calculated on a large scale. These distributions follow a normal distribution and can be used to determine the MFE distribution for any given sequence composition by interpolation. It allows on-the-fly calculation of the normal distribution for any candidate sequence composition. The speedup achieved makes genome-wide screening with this characteristic of a pre-miRNA sequence practical. Although this particular property alone will not be able to distinguish miRNAs from other sequences sufficiently discriminative, the MFE-based P-value should be added to the parameters of choice to be included in the selection of potential miRNA candidates for experimental verification.
S-wave attenuation structure beneath the northern Izu-Bonin arc
NASA Astrophysics Data System (ADS)
Takahashi, Tsutomu; Obana, Koichiro; Kodaira, Shuichi
2016-04-01
To understand temperature structure or magma distribution in the crust and uppermost mantle, it is essential to know their attenuation structure. This study estimated the 3-D S-wave attenuation structure in the crust and uppermost mantle at the northern Izu-Bonin arc, taking into account the apparent attenuation due to multiple forward scattering. In the uppermost mantle, two areas of high seismic attenuation (high Q -1) imaged beneath the volcanic front were mostly colocated with low-velocity anomalies. This coincidence suggests that these high- Q -1 areas in low-velocity zones are the most likely candidates for high-temperature regions beneath volcanoes. The distribution of random inhomogeneities indicated the presence of three anomalies beneath the volcanic front: Two were in high- Q -1 areas but the third was in a moderate- Q -1 area, indicating a low correlation between random inhomogeneities and Q -1. All three anomalies of random inhomogeneities were rich in short-wavelength spectra. The most probable interpretation of such spectra is the presence of volcanic rock, which would be related to accumulated magma intrusion during episodes of volcanic activity. Therefore, the different distributions of Q -1 and random inhomogeneities imply that the positions of hot regions in the uppermost mantle beneath this arc have changed temporally; therefore, they may provide important constraints on the evolutionary processes of arc crust and volcanoes.
Hybrid computer technique yields random signal probability distributions
NASA Technical Reports Server (NTRS)
Cameron, W. D.
1965-01-01
Hybrid computer determines the probability distributions of instantaneous and peak amplitudes of random signals. This combined digital and analog computer system reduces the errors and delays of manual data analysis.
NASA Astrophysics Data System (ADS)
Chen, Fan; Huang, Shaoxiong; Ding, Jinjin; Ding, Jinjin; Gao, Bo; Xie, Yuguang; Wang, Xiaoming
2018-01-01
This paper proposes a fast reliability assessing method for distribution grid with distributed renewable energy generation. First, the Weibull distribution and the Beta distribution are used to describe the probability distribution characteristics of wind speed and solar irradiance respectively, and the models of wind farm, solar park and local load are built for reliability assessment. Then based on power system production cost simulation probability discretization and linearization power flow, a optimal power flow objected with minimum cost of conventional power generation is to be resolved. Thus a reliability assessment for distribution grid is implemented fast and accurately. The Loss Of Load Probability (LOLP) and Expected Energy Not Supplied (EENS) are selected as the reliability index, a simulation for IEEE RBTS BUS6 system in MATLAB indicates that the fast reliability assessing method calculates the reliability index much faster with the accuracy ensured when compared with Monte Carlo method.
Stressor Identification in An Agricultural Watershed: Little Floyd River, Iowa (Final)
EPA announced the availability of the final report, Stressor Identification in An Agricultural Watershed: Little Floyd River, Iowa. This assessment demonstrates that, even when there are many candidate causes and uncertainties are substantial, the probable causes of bio...
40 CFR Appendix C to Part 191 - Guidance for Implementation of Subpart B
Code of Federal Regulations, 2014 CFR
2014-07-01
... that the remaining probability distribution of cumulative releases would not be significantly changed... with § 191.13 into a “complementary cumulative distribution function” that indicates the probability of... distribution function for each disposal system considered. The Agency assumes that a disposal system can be...
40 CFR Appendix C to Part 191 - Guidance for Implementation of Subpart B
Code of Federal Regulations, 2011 CFR
2011-07-01
... that the remaining probability distribution of cumulative releases would not be significantly changed... with § 191.13 into a “complementary cumulative distribution function” that indicates the probability of... distribution function for each disposal system considered. The Agency assumes that a disposal system can be...
Viana, Duarte S; Santamaría, Luis; Figuerola, Jordi
2016-02-01
Propagule retention time is a key factor in determining propagule dispersal distance and the shape of "seed shadows". Propagules dispersed by animal vectors are either ingested and retained in the gut until defecation or attached externally to the body until detachment. Retention time is a continuous variable, but it is commonly measured at discrete time points, according to pre-established sampling time-intervals. Although parametric continuous distributions have been widely fitted to these interval-censored data, the performance of different fitting methods has not been evaluated. To investigate the performance of five different fitting methods, we fitted parametric probability distributions to typical discretized retention-time data with known distribution using as data-points either the lower, mid or upper bounds of sampling intervals, as well as the cumulative distribution of observed values (using either maximum likelihood or non-linear least squares for parameter estimation); then compared the estimated and original distributions to assess the accuracy of each method. We also assessed the robustness of these methods to variations in the sampling procedure (sample size and length of sampling time-intervals). Fittings to the cumulative distribution performed better for all types of parametric distributions (lognormal, gamma and Weibull distributions) and were more robust to variations in sample size and sampling time-intervals. These estimated distributions had negligible deviations of up to 0.045 in cumulative probability of retention times (according to the Kolmogorov-Smirnov statistic) in relation to original distributions from which propagule retention time was simulated, supporting the overall accuracy of this fitting method. In contrast, fitting the sampling-interval bounds resulted in greater deviations that ranged from 0.058 to 0.273 in cumulative probability of retention times, which may introduce considerable biases in parameter estimates. We recommend the use of cumulative probability to fit parametric probability distributions to propagule retention time, specifically using maximum likelihood for parameter estimation. Furthermore, the experimental design for an optimal characterization of unimodal propagule retention time should contemplate at least 500 recovered propagules and sampling time-intervals not larger than the time peak of propagule retrieval, except in the tail of the distribution where broader sampling time-intervals may also produce accurate fits.
How Can Histograms Be Useful for Introducing Continuous Probability Distributions?
ERIC Educational Resources Information Center
Derouet, Charlotte; Parzysz, Bernard
2016-01-01
The teaching of probability has changed a great deal since the end of the last century. The development of technologies is indeed part of this evolution. In France, continuous probability distributions began to be studied in 2002 by scientific 12th graders, but this subject was marginal and appeared only as an application of integral calculus.…
ERIC Educational Resources Information Center
Moses, Tim; Oh, Hyeonjoo J.
2009-01-01
Pseudo Bayes probability estimates are weighted averages of raw and modeled probabilities; these estimates have been studied primarily in nonpsychometric contexts. The purpose of this study was to evaluate pseudo Bayes probability estimates as applied to the estimation of psychometric test score distributions and chained equipercentile equating…
Failure probability under parameter uncertainty.
Gerrard, R; Tsanakas, A
2011-05-01
In many problems of risk analysis, failure is equivalent to the event of a random risk factor exceeding a given threshold. Failure probabilities can be controlled if a decisionmaker is able to set the threshold at an appropriate level. This abstract situation applies, for example, to environmental risks with infrastructure controls; to supply chain risks with inventory controls; and to insurance solvency risks with capital controls. However, uncertainty around the distribution of the risk factor implies that parameter error will be present and the measures taken to control failure probabilities may not be effective. We show that parameter uncertainty increases the probability (understood as expected frequency) of failures. For a large class of loss distributions, arising from increasing transformations of location-scale families (including the log-normal, Weibull, and Pareto distributions), the article shows that failure probabilities can be exactly calculated, as they are independent of the true (but unknown) parameters. Hence it is possible to obtain an explicit measure of the effect of parameter uncertainty on failure probability. Failure probability can be controlled in two different ways: (1) by reducing the nominal required failure probability, depending on the size of the available data set, and (2) by modifying of the distribution itself that is used to calculate the risk control. Approach (1) corresponds to a frequentist/regulatory view of probability, while approach (2) is consistent with a Bayesian/personalistic view. We furthermore show that the two approaches are consistent in achieving the required failure probability. Finally, we briefly discuss the effects of data pooling and its systemic risk implications. © 2010 Society for Risk Analysis.
NASA Astrophysics Data System (ADS)
Wilks, Daniel S.
1993-10-01
Performance of 8 three-parameter probability distributions for representing annual extreme and partial duration precipitation data at stations in the northeastern and southeastern United States is investigated. Particular attention is paid to fidelity on the right tail, through use of a bootstrap procedure simulating extrapolation on the right tail beyond the data. It is found that the beta-κ distribution best describes the extreme right tail of annual extreme series, and the beta-P distribution is best for the partial duration data. The conventionally employed two-parameter Gumbel distribution is found to substantially underestimate probabilities associated with the larger precipitation amounts for both annual extreme and partial duration data. Fitting the distributions using left-censored data did not result in improved fits to the right tail.
NASA Astrophysics Data System (ADS)
Ziegler, Carl; Law, Nicholas M.; Morton, Tim; Baranec, Christoph; Riddle, Reed; Atkinson, Dani; Baker, Anna; Roberts, Sarah; Ciardi, David R.
2017-02-01
The Robo-AO Kepler Planetary Candidate Survey is observing every Kepler planet candidate host star with laser adaptive optics imaging to search for blended nearby stars, which may be physically associated companions and/or responsible for transit false positives. In this paper, we present the results of our search for stars nearby 1629 Kepler planet candidate hosts. With survey sensitivity to objects as close as ˜0.″15, and magnitude differences Δm ≤slant 6, we find 223 stars in the vicinity of 206 target KOIs; 209 of these nearby stars have not been previously imaged in high resolution. We measure an overall nearby-star probability for Kepler planet candidates of 12.6 % +/- 0.9 % at separations between 0.″15 and 4.″0. Particularly interesting KOI systems are discussed, including 26 stars with detected companions that host rocky, habitable zone candidates and five new candidate planet-hosting quadruple star systems. We explore the broad correlations between planetary systems and stellar binarity, using the combined data set of Baranec et al. and this paper. Our previous 2σ result of a low detected nearby star fraction of KOIs hosting close-in giant planets is less apparent in this larger data set. We also find a significant correlation between detected nearby star fraction and KOI number, suggesting possible variation between early and late Kepler data releases.
The estimation of tree posterior probabilities using conditional clade probability distributions.
Larget, Bret
2013-07-01
In this article I introduce the idea of conditional independence of separated subtrees as a principle by which to estimate the posterior probability of trees using conditional clade probability distributions rather than simple sample relative frequencies. I describe an algorithm for these calculations and software which implements these ideas. I show that these alternative calculations are very similar to simple sample relative frequencies for high probability trees but are substantially more accurate for relatively low probability trees. The method allows the posterior probability of unsampled trees to be calculated when these trees contain only clades that are in other sampled trees. Furthermore, the method can be used to estimate the total probability of the set of sampled trees which provides a measure of the thoroughness of a posterior sample.
Hunting the parent of the Orphan stream. II. The first high-resolution spectroscopic study
DOE Office of Scientific and Technical Information (OSTI.GOV)
Casey, Andrew R.; Keller, Stefan C.; Da Costa, Gary
2014-03-20
We present the first high-resolution spectroscopic study on the Orphan stream for five stream candidates, observed with the Magellan Inamori Kyocera Echelle spectrograph on the Magellan Clay telescope. The targets were selected from the low-resolution catalog of Casey et al.: three high-probability members, one medium, and one low-probability stream candidate were observed. Our analysis indicates that the low- and medium-probability targets are metal-rich field stars. The remaining three high-probability targets range over ∼1 dex in metallicity, and are chemically distinct compared to the other two targets and all standard stars: low [α/Fe] abundance ratios are observed, and lower limits aremore » ascertained for [Ba/Y], which sit well above the Milky Way trend. These chemical signatures demonstrate that the undiscovered parent system is unequivocally a dwarf spheroidal galaxy, consistent with dynamical constraints inferred from the stream width and arc. As such, we firmly exclude the proposed association between NGC 2419 and the Orphan stream. A wide range in metallicities adds to the similarities between the Orphan stream and Segue 1, although the low [α/Fe] abundance ratios in the Orphan stream are in tension with the high [α/Fe] values observed in Segue 1. Open questions remain before Segue 1 could possibly be claimed as the 'parent' of the Orphan stream. The parent system could well remain undiscovered in the southern sky.« less
NASA Astrophysics Data System (ADS)
Zuo, Weiguang; Liu, Ming; Fan, Tianhui; Wang, Pengtao
2018-06-01
This paper presents the probability distribution of the slamming pressure from an experimental study of regular wave slamming on an elastically supported horizontal deck. The time series of the slamming pressure during the wave impact were first obtained through statistical analyses on experimental data. The exceeding probability distribution of the maximum slamming pressure peak and distribution parameters were analyzed, and the results show that the exceeding probability distribution of the maximum slamming pressure peak accords with the three-parameter Weibull distribution. Furthermore, the range and relationships of the distribution parameters were studied. The sum of the location parameter D and the scale parameter L was approximately equal to 1.0, and the exceeding probability was more than 36.79% when the random peak was equal to the sample average during the wave impact. The variation of the distribution parameters and slamming pressure under different model conditions were comprehensively presented, and the parameter values of the Weibull distribution of wave-slamming pressure peaks were different due to different test models. The parameter values were found to decrease due to the increased stiffness of the elastic support. The damage criterion of the structure model caused by the wave impact was initially discussed, and the structure model was destroyed when the average slamming time was greater than a certain value during the duration of the wave impact. The conclusions of the experimental study were then described.
On the inequivalence of the CH and CHSH inequalities due to finite statistics
NASA Astrophysics Data System (ADS)
Renou, M. O.; Rosset, D.; Martin, A.; Gisin, N.
2017-06-01
Different variants of a Bell inequality, such as CHSH and CH, are known to be equivalent when evaluated on nonsignaling outcome probability distributions. However, in experimental setups, the outcome probability distributions are estimated using a finite number of samples. Therefore the nonsignaling conditions are only approximately satisfied and the robustness of the violation depends on the chosen inequality variant. We explain that phenomenon using the decomposition of the space of outcome probability distributions under the action of the symmetry group of the scenario, and propose a method to optimize the statistical robustness of a Bell inequality. In the process, we describe the finite group composed of relabeling of parties, measurement settings and outcomes, and identify correspondences between the irreducible representations of this group and properties of outcome probability distributions such as normalization, signaling or having uniform marginals.
Ishimori, Naoki; Stylianou, Ioannis M; Korstanje, Ron; Marion, Michael A; Li, Renhua; Donahue, Leah Rae; Rosen, Clifford J; Beamer, Wesley G; Paigen, Beverly; Churchill, Gary A
2008-01-01
Identification of genes that regulate BMD will enhance our understanding of osteoporosis and could provide novel molecular targets for treatment or prevention. We generated a mouse intercross population and carried out a quantitative trait locus (QTL) analysis of 143 female and 124 male F2 progeny from progenitor strains SM/J and NZB/BlNJ using whole body and vertebral areal BMD (aBMD) as measured by DXA. We found that both whole body and vertebral aBMD was affected by two loci on chromosome 9: one with a significant epistatic interaction on distal chromosome 8 and the other with a sex-specific effect. Two additional significant QTLs were identified on chromosome 12, and several suggestive ones were identified on chromosomes 5, 8, 15, and 19. The chromosome 9, 12, and 15 loci have been previously identified in other crosses. SNP-based haplotype analysis of the progenitor strains identified blocks within the QTL region that distinguish the low allele strains from the high allele strains, significantly narrowing the QTL region and reducing the possible candidate genes to 98 for chromosome 9, 31 for chromosome 12, and only 2 for chromosome 15. Trps1 is the most probable candidate gene for the chromosome 15 QTL. The sex-specific effects may help to elucidate the BMD differences between males and females. This study shows the power of statistical modeling to resolve linked QTLs and the use of haplotype analysis in narrowing the list of candidates. PMID:18442308
Confidence as Bayesian Probability: From Neural Origins to Behavior.
Meyniel, Florent; Sigman, Mariano; Mainen, Zachary F
2015-10-07
Research on confidence spreads across several sub-fields of psychology and neuroscience. Here, we explore how a definition of confidence as Bayesian probability can unify these viewpoints. This computational view entails that there are distinct forms in which confidence is represented and used in the brain, including distributional confidence, pertaining to neural representations of probability distributions, and summary confidence, pertaining to scalar summaries of those distributions. Summary confidence is, normatively, derived or "read out" from distributional confidence. Neural implementations of readout will trade off optimality versus flexibility of routing across brain systems, allowing confidence to serve diverse cognitive functions. Copyright © 2015 Elsevier Inc. All rights reserved.
Exact probability distribution functions for Parrondo's games
NASA Astrophysics Data System (ADS)
Zadourian, Rubina; Saakian, David B.; Klümper, Andreas
2016-12-01
We study the discrete time dynamics of Brownian ratchet models and Parrondo's games. Using the Fourier transform, we calculate the exact probability distribution functions for both the capital dependent and history dependent Parrondo's games. In certain cases we find strong oscillations near the maximum of the probability distribution with two limiting distributions for odd and even number of rounds of the game. Indications of such oscillations first appeared in the analysis of real financial data, but now we have found this phenomenon in model systems and a theoretical understanding of the phenomenon. The method of our work can be applied to Brownian ratchets, molecular motors, and portfolio optimization.
Exact probability distribution functions for Parrondo's games.
Zadourian, Rubina; Saakian, David B; Klümper, Andreas
2016-12-01
We study the discrete time dynamics of Brownian ratchet models and Parrondo's games. Using the Fourier transform, we calculate the exact probability distribution functions for both the capital dependent and history dependent Parrondo's games. In certain cases we find strong oscillations near the maximum of the probability distribution with two limiting distributions for odd and even number of rounds of the game. Indications of such oscillations first appeared in the analysis of real financial data, but now we have found this phenomenon in model systems and a theoretical understanding of the phenomenon. The method of our work can be applied to Brownian ratchets, molecular motors, and portfolio optimization.
NASA Astrophysics Data System (ADS)
Lyon, R. J.; Stappers, B. W.; Cooper, S.; Brooke, J. M.; Knowles, J. D.
2016-06-01
Improving survey specifications are causing an exponential rise in pulsar candidate numbers and data volumes. We study the candidate filters used to mitigate these problems during the past fifty years. We find that some existing methods such as applying constraints on the total number of candidates collected per observation, may have detrimental effects on the success of pulsar searches. Those methods immune to such effects are found to be ill-equipped to deal with the problems associated with increasing data volumes and candidate numbers, motivating the development of new approaches. We therefore present a new method designed for on-line operation. It selects promising candidates using a purpose-built tree-based machine learning classifier, the Gaussian Hellinger Very Fast Decision Tree (GH-VFDT), and a new set of features for describing candidates. The features have been chosen so as to I) maximise the separation between candidates arising from noise and those of probable astrophysical origin, and II) be as survey-independent as possible. Using these features our new approach can process millions of candidates in seconds (˜1 million every 15 seconds), with high levels of pulsar recall (90%+). This technique is therefore applicable to the large volumes of data expected to be produced by the Square Kilometre Array (SKA). Use of this approach has assisted in the discovery of 20 new pulsars in data obtained during the LOFAR Tied-Array All-Sky Survey (LOTAAS).
Biasi, G.P.; Weldon, R.J.; Fumal, T.E.; Seitz, G.G.
2002-01-01
We introduce a quantitative approach to paleoearthquake dating and apply it to paleoseismic data from the Wrightwood and Pallett Creek sites on the southern San Andreas fault. We illustrate how stratigraphic ordering, sedimentological, and historical data can be used quantitatively in the process of estimating earthquake ages. Calibrated radiocarbon age distributions are used directly from layer dating through recurrence intervals and recurrence probability estimation. The method does not eliminate subjective judgements in event dating, but it does provide a means of systematically and objectively approaching the dating process. Date distributions for the most recent 14 events at Wrightwood are based on sample and contextual evidence in Fumal et al. (2002) and site context and slip history in Weldon et al. (2002). Pallett Creek event and dating descriptions are from published sources. For the five most recent events at Wrightwood, our results are consistent with previously published estimates, with generally comparable or narrower uncertainties. For Pallett Creek, our earthquake date estimates generally overlap with previous results but typically have broader uncertainties. Some event date estimates are very sensitive to details of data interpretation. The historical earthquake in 1857 ruptured the ground at both sites but is not constrained by radiocarbon data. Radiocarbon ages, peat accumulation rates, and historical constraints at Pallett Creek for event X yield a date estimate in the earliest 1800s and preclude a date in the late 1600s. This event is almost certainly the historical 1812 earthquake, as previously concluded by Sieh et al. (1989). This earthquake also produced ground deformation at Wrightwood. All events at Pallett Creek, except for event T, about A.D. 1360, and possibly event I, about A.D. 960, have corresponding events at Wrightwood with some overlap in age ranges. Event T falls during a period of low sedimentation at Wrightwood when conditions were not favorable for recording earthquake evidence. Previously proposed correlations of Pallett Creek X with Wrightwood W3 in the 1690s and Pallett Creek event V with W5 around 1480 (Fumal et al., 1993) appear unlikely after our dating reevaluation. Apparent internal inconsistencies among event, layer, and dating relationships around events R and V identify them as candidates for further investigation at the site. Conditional probabilities of earthquake recurrence were estimated using Poisson, lognormal, and empirical models. The presence of 12 or 13 events at Wrightwood during the same interval that 10 events are reported at Pallett Creek is reflected in mean recurrence intervals of 105 and 135 years, respectively. Average Poisson model 30-year conditional probabilities are about 20% at Pallett Creek and 25% at Wrightwood. The lognormal model conditional probabilities are somewhat higher, about 25% for Pallett Creek and 34% for Wrightwood. Lognormal variance ??ln estimates of 0.76 and 0.70, respectively, imply only weak time predictability. Conditional probabilities of 29% and 46%, respectively, were estimated for an empirical distribution derived from the data alone. Conditional probability uncertainties are dominated by the brevity of the event series; dating uncertainty contributes only secondarily. Wrightwood and Pallett Creek event chronologies both suggest variations in recurrence interval with time, hinting that some form of recurrence rate modulation may be at work, but formal testing shows that neither series is more ordered than might be produced by a Poisson process.
Why anthropic reasoning cannot predict Lambda.
Starkman, Glenn D; Trotta, Roberto
2006-11-17
We revisit anthropic arguments purporting to explain the measured value of the cosmological constant. We argue that different ways of assigning probabilities to candidate universes lead to totally different anthropic predictions. As an explicit example, we show that weighting different universes by the total number of possible observations leads to an extremely small probability for observing a value of Lambda equal to or greater than what we now measure. We conclude that anthropic reasoning within the framework of probability as frequency is ill-defined and that in the absence of a fundamental motivation for selecting one weighting scheme over another the anthropic principle cannot be used to explain the value of Lambda, nor, likely, any other physical parameters.
What Can Quantum Optics Say about Computational Complexity Theory?
NASA Astrophysics Data System (ADS)
Rahimi-Keshari, Saleh; Lund, Austin P.; Ralph, Timothy C.
2015-02-01
Considering the problem of sampling from the output photon-counting probability distribution of a linear-optical network for input Gaussian states, we obtain results that are of interest from both quantum theory and the computational complexity theory point of view. We derive a general formula for calculating the output probabilities, and by considering input thermal states, we show that the output probabilities are proportional to permanents of positive-semidefinite Hermitian matrices. It is believed that approximating permanents of complex matrices in general is a #P-hard problem. However, we show that these permanents can be approximated with an algorithm in the BPPNP complexity class, as there exists an efficient classical algorithm for sampling from the output probability distribution. We further consider input squeezed-vacuum states and discuss the complexity of sampling from the probability distribution at the output.
VizieR Online Data Catalog: Planck high-z source candidates catalog (PHZ) (Planck+, 2016)
NASA Astrophysics Data System (ADS)
Planck Collaboration; Ade, P. A. R.; Aghanim, N.; Arnaud, M.; Aumont, J.; Baccigalupi, C.; Banday, A. J.; Barreiro, R. B.; Bartolo, N.; Battaner, E.; Benabed, K.; Benoit-Levy, A.; Bernard, J.-P.; Bersanelli, M.; Bielewicz, P.; Bonaldi, A.; Bonavera, L.; Bond, J. R.; Borrill, J.; Bouchet, F. R.; Boulanger, F.; Burigana, C.; Butler, R. C.; Calabrese, E.; Catalano, A.; Chiang, H. C.; Christensen, P. R.; Clements, D. L.; Colombo, L. P. L.; Couchot, F.; Coulais, A.; Crill, B. P.; Curto, A.; Cuttaia, F.; Danese, L.; Davies, R. D.; Davis, R. J.; de Bernardis, P.; De Rosa, A.; de Zotti, G.; Delabrouille, J.; Dickinson, C.; Diego, J. M.; Dole, H.; Dore, O.; Douspis, M.; Ducout, A.; Dupac, X.; Elsner, F.; Ensslin, T. A.; Eriksen, H. K.; Falgarone, E.; Finelli, F.; Flores-Cacho, I.; Frailis, M.; Fraisse, A. A.; Franceschi, E.; Galeotta, S.; Galli, S.; Ganga, K.; Giard, M.; Giraud-Heraud, Y.; Gjerlow, E.; Gonzalez-Nuevo, J.; Gorski, K. M.; Gregorio, A.; Gruppuso, A.; Gudmundsson, J. E.; Hansen, F. K.; Harrison, D. L.; Helou, G.; Hernandez-Monteagudo, C.; Herranz, D.; Hildebrandt, S. R.; Hivon, E.; Hobson, M.; Hornstrup, A.; Hovest, W.; Huffenberger, K. M.; Hurier, G.; Jaffe, A. H.; Jaffe, T. R.; Keihanen, E.; Keskitalo, R.; Kisner, T. S.; Kneissl, R.; Knoche, J.; Kunz, M.; Kurki-Suonio, H.; Lagache, G.; Lamarre, J.-M.; Lasenby, A.; Lattanzi, M.; Lawrence, C. R.; Leonardi, R.; Levrier, F.; Liguori, M.; Lilje, P. B.; Linden-Vornle, M.; Lopez-Caniego, M.; Lubin, P. M.; Macias-Perez, J. F.; Maffei, B.; Maggio, G.; Maino, D.; Mandolesi, N.; Mangilli, A.; Maris, M.; Martin, P. G.; Martinez-Gonzalez, E.; Masi, S.; Matarrese, S.; Melchiorri, A.; Mennella, A.; Migliaccio, M.; Mitra, S.; Miville-Deschenes, M.-A.; Moneti, A.; Montier, L.; Morgante, G.; Mortlock, D.; Munshi, D.; Murphy, J. A.; Nati, F.; Natoli, P.; Nesvadba, N. P. H.; Noviello, F.; Novikov, D.; Novikov, I.; Oxborrow, C. A.; Pagano, L.; Pajot, F.; Paoletti, D.; Partridge, B.; Pasian, F.; Pearson, T. J.; Perdereau, O.; Perotto, L.; Pettorino, V.; Piacentini, F.; Piat, M.; Plaszczynski, S.; Pointecouteau, E.; Polenta, G.; Pratt, G. W.; Prunet, S.; Puget, J.-L.; Rachen, J. P.; Reinecke, M.; Remazeilles, M.; Renault, C.; Renzi, A.; Ristorcelli, I.; Rocha, G.; Rosset, C.; Rossetti, M.; Roudier, G.; Rubino-Martin, J. A.; Rusholme, B.; Sandri, M.; Santos, D.; Savelainen, M.; Savini, G.; Scott, D.; Spencer, L. D.; Stolyarov, V.; Stompor, R.; Sudiwala, R.; Sunyaev, R.; Suur-Uski, A.-S.; Sygnet, J.-F.; Tauber, J. A.; Terenzi, L.; Toffolatti, L.; Tomasi, M.; Tristram, M.; Tucci, M.; Turler, M.; Umana, G.; Valenziano, L.; Valiviita, J.; van Tent, F.; Vielva, P.; Villa, F.; Wade, L. A.; Wandelt, B. D.; Wehus, I. K.; Welikala, N.; Yvon, D.; Zacchei, A.; Zonca, A.
2016-11-01
We present in this work the Planck List of Highredshift Source Candidates (the "PHZ"), which includes 2151 sources distributed over 26% of the sky, with redshifts likely to be greater than 2. (2 data files).
Muffly, Tyler M; Barber, Matthew D; Karafa, Matthew T; Kattan, Michael W; Shniter, Abigail; Jelovsek, J Eric
2012-01-01
The purpose of the study was to develop a model that predicts an individual applicant's probability of successful placement into a surgical subspecialty fellowship program. Candidates who applied to surgical fellowships during a 3-year period were identified in a set of databases that included the electronic application materials. Of the 1281 applicants who were available for analysis, 951 applicants (74%) successfully placed into a colon and rectal surgery, thoracic surgery, vascular surgery, or pediatric surgery fellowship. The optimal final prediction model, which was based on a logistic regression, included 14 variables. This model, with a c statistic of 0.74, allowed for the determination of a useful estimate of the probability of placement for an individual candidate. Of the factors that are available at the time of fellowship application, 14 were used to predict accurately the proportion of applicants who will successfully gain a fellowship position. Copyright © 2012 Association of Program Directors in Surgery. Published by Elsevier Inc. All rights reserved.
Vacuum quantum stress tensor fluctuations: A diagonalization approach
NASA Astrophysics Data System (ADS)
Schiappacasse, Enrico D.; Fewster, Christopher J.; Ford, L. H.
2018-01-01
Large vacuum fluctuations of a quantum stress tensor can be described by the asymptotic behavior of its probability distribution. Here we focus on stress tensor operators which have been averaged with a sampling function in time. The Minkowski vacuum state is not an eigenstate of the time-averaged operator, but can be expanded in terms of its eigenstates. We calculate the probability distribution and the cumulative probability distribution for obtaining a given value in a measurement of the time-averaged operator taken in the vacuum state. In these calculations, we study a specific operator that contributes to the stress-energy tensor of a massless scalar field in Minkowski spacetime, namely, the normal ordered square of the time derivative of the field. We analyze the rate of decrease of the tail of the probability distribution for different temporal sampling functions, such as compactly supported functions and the Lorentzian function. We find that the tails decrease relatively slowly, as exponentials of fractional powers, in agreement with previous work using the moments of the distribution. Our results lend additional support to the conclusion that large vacuum stress tensor fluctuations are more probable than large thermal fluctuations, and may have observable effects.
Measurements of gas hydrate formation probability distributions on a quasi-free water droplet
NASA Astrophysics Data System (ADS)
Maeda, Nobuo
2014-06-01
A High Pressure Automated Lag Time Apparatus (HP-ALTA) can measure gas hydrate formation probability distributions from water in a glass sample cell. In an HP-ALTA gas hydrate formation originates near the edges of the sample cell and gas hydrate films subsequently grow across the water-guest gas interface. It would ideally be desirable to be able to measure gas hydrate formation probability distributions of a single water droplet or mist that is freely levitating in a guest gas, but this is technically challenging. The next best option is to let a water droplet sit on top of a denser, immiscible, inert, and wall-wetting hydrophobic liquid to avoid contact of a water droplet with the solid walls. Here we report the development of a second generation HP-ALTA which can measure gas hydrate formation probability distributions of a water droplet which sits on a perfluorocarbon oil in a container that is coated with 1H,1H,2H,2H-Perfluorodecyltriethoxysilane. It was found that the gas hydrate formation probability distributions of such a quasi-free water droplet were significantly lower than those of water in a glass sample cell.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Xiao Hongyu; Covey, Kevin R.; Lloyd, James P.
2012-09-15
We analyze light curves obtained by the Trans-atlantic Exoplanet Survey (TrES) for a field centered on the L1495 dark cloud in Taurus. The Spitzer Taurus Legacy Survey catalog identifies 179 bona fide Taurus members within the TrES field; 48 of the known Taurus members are detected by TrES, as well as 26 candidate members identified by the Spitzer Legacy team. We quantify the variability of each star in our sample using the ratio of the standard deviation of the original light curve ({sigma}{sub orig.}) to the standard deviation of a light curve that has been smoothed by 9 or 1001more » epochs ({sigma}{sub 9} and {sigma}{sub 1001}, respectively). Known Taurus members typically demonstrate ({sigma}{sub orig.}/{sigma}{sub 9}) < 2.0, and ({sigma}{sub orig.}/{sigma}{sub 1001}) < 5, while field stars reveal ({sigma}{sub orig.}/{sigma}{sub 9}) {approx} 3.0 and ({sigma}{sub orig.}/{sigma}{sub 1001}) {approx} 10, as expected for light curves dominated by unstructured white noise. Of the 74 Taurus members/candidates with TrES light curves, we detect significant variability in 49 sources. Adapting a quantitative metric originally developed to assess the reliability of transit detections, we measure the amount of red and white noise in each light curve and identify 18 known or candidate Taurus members with highly significant period measurements. These appear to be the first periods measured for four of these sources (HD 282276, CX Tau, FP Tau, TrES J042423+265008), and in two other cases, the first non-aliased periods (LkCa 21 and DK Tau AB). For the remainder, the TrES measurements typically agree very well ({delta}P < 1%) with previously reported values. Including periods measured at lower confidence for 15 additional sources, we report periods for 11 objects where no previous periods were found, including 8 confirmed Taurus members. We also identify 10 of the 26 candidate Taurus members that demonstrate variability levels consistent with being bona fide T Tauri stars. A Kolomgorov-Smirnov (K-S) test confirms that these new periods confirm the distinction between the rotation period distributions of stars with and without circumstellar disks, with only a 10% probability of the two populations sharing the same parent period distribution. K-S tests do suggest, however, that the updated Taurus period distribution now more closely resembles those measured in other young star-forming clusters (i.e., NGC 2264, NGC 6530, and the ONC). This improved agreement may reflect the exclusion of long rotation periods which are detected in Taurus at lower significance, and which may be beyond the limits of detectability in more distant star-forming regions.« less
Fragment size distribution in viscous bag breakup of a drop
NASA Astrophysics Data System (ADS)
Kulkarni, Varun; Bulusu, Kartik V.; Plesniak, Michael W.; Sojka, Paul E.
2015-11-01
In this study we examine the drop size distribution resulting from the fragmentation of a single drop in the presence of a continuous air jet. Specifically, we study the effect of Weber number, We, and Ohnesorge number, Oh on the disintegration process. The regime of breakup considered is observed between 12 <= We <= 16 for Oh <= 0.1. Experiments are conducted using phase Doppler anemometry. Both the number and volume fragment size probability distributions are plotted. The volume probability distribution revealed a bi-modal behavior with two distinct peaks: one corresponding to the rim fragments and the other to the bag fragments. This behavior was suppressed in the number probability distribution. Additionally, we employ an in-house particle detection code to isolate the rim fragment size distribution from the total probability distributions. Our experiments showed that the bag fragments are smaller in diameter and larger in number, while the rim fragments are larger in diameter and smaller in number. Furthermore, with increasing We for a given Ohwe observe a large number of small-diameter drops and small number of large-diameter drops. On the other hand, with increasing Oh for a fixed We the opposite is seen.
Nuclear risk analysis of the Ulysses mission
NASA Astrophysics Data System (ADS)
Bartram, Bart W.; Vaughan, Frank R.; Englehart, Richard W.
An account is given of the method used to quantify the risks accruing to the use of a radioisotope thermoelectric generator fueled by Pu-238 dioxide aboard the Space Shuttle-launched Ulysses mission. After using a Monte Carlo technique to develop probability distributions for the radiological consequences of a range of accident scenarios throughout the mission, factors affecting those consequences are identified in conjunction with their probability distributions. The functional relationship among all the factors is then established, and probability distributions for all factor effects are combined by means of a Monte Carlo technique.
Score distributions of gapped multiple sequence alignments down to the low-probability tail
NASA Astrophysics Data System (ADS)
Fieth, Pascal; Hartmann, Alexander K.
2016-08-01
Assessing the significance of alignment scores of optimally aligned DNA or amino acid sequences can be achieved via the knowledge of the score distribution of random sequences. But this requires obtaining the distribution in the biologically relevant high-scoring region, where the probabilities are exponentially small. For gapless local alignments of infinitely long sequences this distribution is known analytically to follow a Gumbel distribution. Distributions for gapped local alignments and global alignments of finite lengths can only be obtained numerically. To obtain result for the small-probability region, specific statistical mechanics-based rare-event algorithms can be applied. In previous studies, this was achieved for pairwise alignments. They showed that, contrary to results from previous simple sampling studies, strong deviations from the Gumbel distribution occur in case of finite sequence lengths. Here we extend the studies to multiple sequence alignments with gaps, which are much more relevant for practical applications in molecular biology. We study the distributions of scores over a large range of the support, reaching probabilities as small as 10-160, for global and local (sum-of-pair scores) multiple alignments. We find that even after suitable rescaling, eliminating the sequence-length dependence, the distributions for multiple alignment differ from the pairwise alignment case. Furthermore, we also show that the previously discussed Gaussian correction to the Gumbel distribution needs to be refined, also for the case of pairwise alignments.
Site occupancy models with heterogeneous detection probabilities
Royle, J. Andrew
2006-01-01
Models for estimating the probability of occurrence of a species in the presence of imperfect detection are important in many ecological disciplines. In these ?site occupancy? models, the possibility of heterogeneity in detection probabilities among sites must be considered because variation in abundance (and other factors) among sampled sites induces variation in detection probability (p). In this article, I develop occurrence probability models that allow for heterogeneous detection probabilities by considering several common classes of mixture distributions for p. For any mixing distribution, the likelihood has the general form of a zero-inflated binomial mixture for which inference based upon integrated likelihood is straightforward. A recent paper by Link (2003, Biometrics 59, 1123?1130) demonstrates that in closed population models used for estimating population size, different classes of mixture distributions are indistinguishable from data, yet can produce very different inferences about population size. I demonstrate that this problem can also arise in models for estimating site occupancy in the presence of heterogeneous detection probabilities. The implications of this are discussed in the context of an application to avian survey data and the development of animal monitoring programs.
Implementation of the Iterative Proportion Fitting Algorithm for Geostatistical Facies Modeling
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li Yupeng, E-mail: yupeng@ualberta.ca; Deutsch, Clayton V.
2012-06-15
In geostatistics, most stochastic algorithm for simulation of categorical variables such as facies or rock types require a conditional probability distribution. The multivariate probability distribution of all the grouped locations including the unsampled location permits calculation of the conditional probability directly based on its definition. In this article, the iterative proportion fitting (IPF) algorithm is implemented to infer this multivariate probability. Using the IPF algorithm, the multivariate probability is obtained by iterative modification to an initial estimated multivariate probability using lower order bivariate probabilities as constraints. The imposed bivariate marginal probabilities are inferred from profiles along drill holes or wells.more » In the IPF process, a sparse matrix is used to calculate the marginal probabilities from the multivariate probability, which makes the iterative fitting more tractable and practical. This algorithm can be extended to higher order marginal probability constraints as used in multiple point statistics. The theoretical framework is developed and illustrated with estimation and simulation example.« less
Modeling highway travel time distribution with conditional probability models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oliveira Neto, Francisco Moraes; Chin, Shih-Miao; Hwang, Ho-Ling
ABSTRACT Under the sponsorship of the Federal Highway Administration's Office of Freight Management and Operations, the American Transportation Research Institute (ATRI) has developed performance measures through the Freight Performance Measures (FPM) initiative. Under this program, travel speed information is derived from data collected using wireless based global positioning systems. These telemetric data systems are subscribed and used by trucking industry as an operations management tool. More than one telemetric operator submits their data dumps to ATRI on a regular basis. Each data transmission contains truck location, its travel time, and a clock time/date stamp. Data from the FPM program providesmore » a unique opportunity for studying the upstream-downstream speed distributions at different locations, as well as different time of the day and day of the week. This research is focused on the stochastic nature of successive link travel speed data on the continental United States Interstates network. Specifically, a method to estimate route probability distributions of travel time is proposed. This method uses the concepts of convolution of probability distributions and bivariate, link-to-link, conditional probability to estimate the expected distributions for the route travel time. Major contribution of this study is the consideration of speed correlation between upstream and downstream contiguous Interstate segments through conditional probability. The established conditional probability distributions, between successive segments, can be used to provide travel time reliability measures. This study also suggests an adaptive method for calculating and updating route travel time distribution as new data or information is added. This methodology can be useful to estimate performance measures as required by the recent Moving Ahead for Progress in the 21st Century Act (MAP 21).« less
Probability of success for phase III after exploratory biomarker analysis in phase II.
Götte, Heiko; Kirchner, Marietta; Sailer, Martin Oliver
2017-05-01
The probability of success or average power describes the potential of a future trial by weighting the power with a probability distribution of the treatment effect. The treatment effect estimate from a previous trial can be used to define such a distribution. During the development of targeted therapies, it is common practice to look for predictive biomarkers. The consequence is that the trial population for phase III is often selected on the basis of the most extreme result from phase II biomarker subgroup analyses. In such a case, there is a tendency to overestimate the treatment effect. We investigate whether the overestimation of the treatment effect estimate from phase II is transformed into a positive bias for the probability of success for phase III. We simulate a phase II/III development program for targeted therapies. This simulation allows to investigate selection probabilities and allows to compare the estimated with the true probability of success. We consider the estimated probability of success with and without subgroup selection. Depending on the true treatment effects, there is a negative bias without selection because of the weighting by the phase II distribution. In comparison, selection increases the estimated probability of success. Thus, selection does not lead to a bias in probability of success if underestimation due to the phase II distribution and overestimation due to selection cancel each other out. We recommend to perform similar simulations in practice to get the necessary information about the risk and chances associated with such subgroup selection designs. Copyright © 2017 John Wiley & Sons, Ltd.
General formulation of long-range degree correlations in complex networks
NASA Astrophysics Data System (ADS)
Fujiki, Yuka; Takaguchi, Taro; Yakubo, Kousuke
2018-06-01
We provide a general framework for analyzing degree correlations between nodes separated by more than one step (i.e., beyond nearest neighbors) in complex networks. One joint and four conditional probability distributions are introduced to fully describe long-range degree correlations with respect to degrees k and k' of two nodes and shortest path length l between them. We present general relations among these probability distributions and clarify the relevance to nearest-neighbor degree correlations. Unlike nearest-neighbor correlations, some of these probability distributions are meaningful only in finite-size networks. Furthermore, as a baseline to determine the existence of intrinsic long-range degree correlations in a network other than inevitable correlations caused by the finite-size effect, the functional forms of these probability distributions for random networks are analytically evaluated within a mean-field approximation. The utility of our argument is demonstrated by applying it to real-world networks.
Stochastic analysis of particle movement over a dune bed
Lee, Baum K.; Jobson, Harvey E.
1977-01-01
Stochastic models are available that can be used to predict the transport and dispersion of bed-material sediment particles in an alluvial channel. These models are based on the proposition that the movement of a single bed-material sediment particle consists of a series of steps of random length separated by rest periods of random duration and, therefore, application of the models requires a knowledge of the probability distributions of the step lengths, the rest periods, the elevation of particle deposition, and the elevation of particle erosion. The procedure was tested by determining distributions from bed profiles formed in a large laboratory flume with a coarse sand as the bed material. The elevation of particle deposition and the elevation of particle erosion can be considered to be identically distributed, and their distribution can be described by either a ' truncated Gaussian ' or a ' triangular ' density function. The conditional probability distribution of the rest period given the elevation of particle deposition closely followed the two-parameter gamma distribution. The conditional probability distribution of the step length given the elevation of particle erosion and the elevation of particle deposition also closely followed the two-parameter gamma density function. For a given flow, the scale and shape parameters describing the gamma probability distributions can be expressed as functions of bed-elevation. (Woodard-USGS)
Wang, Jihan; Yang, Kai
2014-07-01
An efficient operating room needs both little underutilised and overutilised time to achieve optimal cost efficiency. The probabilities of underrun and overrun of lists of cases can be estimated by a well defined duration distribution of the lists. To propose a method of predicting the probabilities of underrun and overrun of lists of cases using Type IV Pearson distribution to support case scheduling. Six years of data were collected. The first 5 years of data were used to fit distributions and estimate parameters. The data from the last year were used as testing data to validate the proposed methods. The percentiles of the duration distribution of lists of cases were calculated by Type IV Pearson distribution and t-distribution. Monte Carlo simulation was conducted to verify the accuracy of percentiles defined by the proposed methods. Operating rooms in John D. Dingell VA Medical Center, United States, from January 2005 to December 2011. Differences between the proportion of lists of cases that were completed within the percentiles of the proposed duration distribution of the lists and the corresponding percentiles. Compared with the t-distribution, the proposed new distribution is 8.31% (0.38) more accurate on average and 14.16% (0.19) more accurate in calculating the probabilities at the 10th and 90th percentiles of the distribution, which is a major concern of operating room schedulers. The absolute deviations between the percentiles defined by Type IV Pearson distribution and those from Monte Carlo simulation varied from 0.20 min (0.01) to 0.43 min (0.03). Operating room schedulers can rely on the most recent 10 cases with the same combination of surgeon and procedure(s) for distribution parameter estimation to plan lists of cases. Values are mean (SEM). The proposed Type IV Pearson distribution is more accurate than t-distribution to estimate the probabilities of underrun and overrun of lists of cases. However, as not all the individual case durations followed log-normal distributions, there was some deviation from the true duration distribution of the lists.
A tool for simulating collision probabilities of animals with marine renewable energy devices.
Schmitt, Pál; Culloch, Ross; Lieber, Lilian; Molander, Sverker; Hammar, Linus; Kregting, Louise
2017-01-01
The mathematical problem of establishing a collision probability distribution is often not trivial. The shape and motion of the animal as well as of the the device must be evaluated in a four-dimensional space (3D motion over time). Earlier work on wind and tidal turbines was limited to a simplified two-dimensional representation, which cannot be applied to many new structures. We present a numerical algorithm to obtain such probability distributions using transient, three-dimensional numerical simulations. The method is demonstrated using a sub-surface tidal kite as an example. Necessary pre- and post-processing of the data created by the model is explained, numerical details and potential issues and limitations in the application of resulting probability distributions are highlighted.
Stationary stability for evolutionary dynamics in finite populations
Harper, Marc; Fryer, Dashiell
2016-08-25
Here, we demonstrate a vast expansion of the theory of evolutionary stability to finite populations with mutation, connecting the theory of the stationary distribution of the Moran process with the Lyapunov theory of evolutionary stability. We define the notion of stationary stability for the Moran process with mutation and generalizations, as well as a generalized notion of evolutionary stability that includes mutation called an incentive stable state (ISS) candidate. For sufficiently large populations, extrema of the stationary distribution are ISS candidates and we give a family of Lyapunov quantities that are locally minimized at the stationary extrema and at ISSmore » candidates. In various examples, including for the Moran andWright–Fisher processes, we show that the local maxima of the stationary distribution capture the traditionally-defined evolutionarily stable states. The classical stability theory of the replicator dynamic is recovered in the large population limit. Finally we include descriptions of possible extensions to populations of variable size and populations evolving on graphs.« less
Lognormal Approximations of Fault Tree Uncertainty Distributions.
El-Shanawany, Ashraf Ben; Ardron, Keith H; Walker, Simon P
2018-01-26
Fault trees are used in reliability modeling to create logical models of fault combinations that can lead to undesirable events. The output of a fault tree analysis (the top event probability) is expressed in terms of the failure probabilities of basic events that are input to the model. Typically, the basic event probabilities are not known exactly, but are modeled as probability distributions: therefore, the top event probability is also represented as an uncertainty distribution. Monte Carlo methods are generally used for evaluating the uncertainty distribution, but such calculations are computationally intensive and do not readily reveal the dominant contributors to the uncertainty. In this article, a closed-form approximation for the fault tree top event uncertainty distribution is developed, which is applicable when the uncertainties in the basic events of the model are lognormally distributed. The results of the approximate method are compared with results from two sampling-based methods: namely, the Monte Carlo method and the Wilks method based on order statistics. It is shown that the closed-form expression can provide a reasonable approximation to results obtained by Monte Carlo sampling, without incurring the computational expense. The Wilks method is found to be a useful means of providing an upper bound for the percentiles of the uncertainty distribution while being computationally inexpensive compared with full Monte Carlo sampling. The lognormal approximation method and Wilks's method appear attractive, practical alternatives for the evaluation of uncertainty in the output of fault trees and similar multilinear models. © 2018 Society for Risk Analysis.
Quantum key distribution without the wavefunction
NASA Astrophysics Data System (ADS)
Niestegge, Gerd
A well-known feature of quantum mechanics is the secure exchange of secret bit strings which can then be used as keys to encrypt messages transmitted over any classical communication channel. It is demonstrated that this quantum key distribution allows a much more general and abstract access than commonly thought. The results include some generalizations of the Hilbert space version of quantum key distribution, but are based upon a general nonclassical extension of conditional probability. A special state-independent conditional probability is identified as origin of the superior security of quantum key distribution; this is a purely algebraic property of the quantum logic and represents the transition probability between the outcomes of two consecutive quantum measurements.
The complexity of divisibility.
Bausch, Johannes; Cubitt, Toby
2016-09-01
We address two sets of long-standing open questions in linear algebra and probability theory, from a computational complexity perspective: stochastic matrix divisibility, and divisibility and decomposability of probability distributions. We prove that finite divisibility of stochastic matrices is an NP-complete problem, and extend this result to nonnegative matrices, and completely-positive trace-preserving maps, i.e. the quantum analogue of stochastic matrices. We further prove a complexity hierarchy for the divisibility and decomposability of probability distributions, showing that finite distribution divisibility is in P, but decomposability is NP-hard. For the former, we give an explicit polynomial-time algorithm. All results on distributions extend to weak-membership formulations, proving that the complexity of these problems is robust to perturbations.
Belcher, Wayne R.; Sweetkind, Donald S.; Elliott, Peggy E.
2002-01-01
The use of geologic information such as lithology and rock properties is important to constrain conceptual and numerical hydrogeologic models. This geologic information is difficult to apply explicitly to numerical modeling and analyses because it tends to be qualitative rather than quantitative. This study uses a compilation of hydraulic-conductivity measurements to derive estimates of the probability distributions for several hydrogeologic units within the Death Valley regional ground-water flow system, a geologically and hydrologically complex region underlain by basin-fill sediments, volcanic, intrusive, sedimentary, and metamorphic rocks. Probability distributions of hydraulic conductivity for general rock types have been studied previously; however, this study provides more detailed definition of hydrogeologic units based on lithostratigraphy, lithology, alteration, and fracturing and compares the probability distributions to the aquifer test data. Results suggest that these probability distributions can be used for studies involving, for example, numerical flow modeling, recharge, evapotranspiration, and rainfall runoff. These probability distributions can be used for such studies involving the hydrogeologic units in the region, as well as for similar rock types elsewhere. Within the study area, fracturing appears to have the greatest influence on the hydraulic conductivity of carbonate bedrock hydrogeologic units. Similar to earlier studies, we find that alteration and welding in the Tertiary volcanic rocks greatly influence hydraulic conductivity. As alteration increases, hydraulic conductivity tends to decrease. Increasing degrees of welding appears to increase hydraulic conductivity because welding increases the brittleness of the volcanic rocks, thus increasing the amount of fracturing.
2014-01-01
Background Small RNAs are important regulators of genome function, yet their prediction in genomes is still a major computational challenge. Statistical analyses of pre-miRNA sequences indicated that their 2D structure tends to have a minimal free energy (MFE) significantly lower than MFE values of equivalently randomized sequences with the same nucleotide composition, in contrast to other classes of non-coding RNA. The computation of many MFEs is, however, too intensive to allow for genome-wide screenings. Results Using a local grid infrastructure, MFE distributions of random sequences were pre-calculated on a large scale. These distributions follow a normal distribution and can be used to determine the MFE distribution for any given sequence composition by interpolation. It allows on-the-fly calculation of the normal distribution for any candidate sequence composition. Conclusion The speedup achieved makes genome-wide screening with this characteristic of a pre-miRNA sequence practical. Although this particular property alone will not be able to distinguish miRNAs from other sequences sufficiently discriminative, the MFE-based P-value should be added to the parameters of choice to be included in the selection of potential miRNA candidates for experimental verification. PMID:24418292
Cockroaches probably cleaned up after dinosaurs.
Vršanský, Peter; van de Kamp, Thomas; Azar, Dany; Prokin, Alexander; Vidlička, L'ubomír; Vagovič, Patrik
2013-01-01
Dinosaurs undoubtedly produced huge quantities of excrements. But who cleaned up after them? Dung beetles and flies with rapid development were rare during most of the Mesozoic. Candidates for these duties are extinct cockroaches (Blattulidae), whose temporal range is associated with herbivorous dinosaurs. An opportunity to test this hypothesis arises from coprolites to some extent extruded from an immature cockroach preserved in the amber of Lebanon, studied using synchrotron X-ray microtomography. 1.06% of their volume is filled by particles of wood with smooth edges, in which size distribution directly supports their external pre-digestion. Because fungal pre-processing can be excluded based on the presence of large particles (combined with small total amount of wood) and absence of damages on wood, the likely source of wood are herbivore feces. Smaller particles were broken down biochemically in the cockroach hind gut, which indicates that the recent lignin-decomposing termite and cockroach endosymbionts might have been transferred to the cockroach gut upon feeding on dinosaur feces.
Cockroaches Probably Cleaned Up after Dinosaurs
Vršanský, Peter; van de Kamp, Thomas; Azar, Dany; Prokin, Alexander; Vidlička, L'ubomír; Vagovič, Patrik
2013-01-01
Dinosaurs undoubtedly produced huge quantities of excrements. But who cleaned up after them? Dung beetles and flies with rapid development were rare during most of the Mesozoic. Candidates for these duties are extinct cockroaches (Blattulidae), whose temporal range is associated with herbivorous dinosaurs. An opportunity to test this hypothesis arises from coprolites to some extent extruded from an immature cockroach preserved in the amber of Lebanon, studied using synchrotron X-ray microtomography. 1.06% of their volume is filled by particles of wood with smooth edges, in which size distribution directly supports their external pre-digestion. Because fungal pre-processing can be excluded based on the presence of large particles (combined with small total amount of wood) and absence of damages on wood, the likely source of wood are herbivore feces. Smaller particles were broken down biochemically in the cockroach hind gut, which indicates that the recent lignin-decomposing termite and cockroach endosymbionts might have been transferred to the cockroach gut upon feeding on dinosaur feces. PMID:24324610
Discriminability limits in spatio-temporal stereo block matching.
Jain, Ankit K; Nguyen, Truong Q
2014-05-01
Disparity estimation is a fundamental task in stereo imaging and is a well-studied problem. Recently, methods have been adapted to the video domain where motion is used as a matching criterion to help disambiguate spatially similar candidates. In this paper, we analyze the validity of the underlying assumptions of spatio-temporal disparity estimation, and determine the extent to which motion aids the matching process. By analyzing the error signal for spatio-temporal block matching under the sum of squared differences criterion and treating motion as a stochastic process, we determine the probability of a false match as a function of image features, motion distribution, image noise, and number of frames in the spatio-temporal patch. This performance quantification provides insight into when spatio-temporal matching is most beneficial in terms of the scene and motion, and can be used as a guide to select parameters for stereo matching algorithms. We validate our results through simulation and experiments on stereo video.
Modelling electro-active polymers with a dispersion-type anisotropy
NASA Astrophysics Data System (ADS)
Hossain, Mokarram; Steinmann, Paul
2018-02-01
We propose a novel constitutive framework for electro-active polymers (EAPs) that can take into account anisotropy with a chain dispersion. To enhance actuation behaviour, particle-filled EAPs become promising candidates nowadays. Recent studies suggest that particle-filled EAPs, which can be cured under an electric field during the manufacturing time, do not necessarily form perfect anisotropic composites, rather they create composites with dispersed chains. Hence in this contribution, an electro-mechanically coupled constitutive model is devised that considers the chain dispersion with a probability distribution function in an integral form. To obtain relevant quantities in discrete form, numerical integration over the unit sphere is utilized. Necessary constitutive equations are derived exploiting the basic laws of thermodynamics that result in a thermodynamically consistent formulation. To demonstrate the performance of the proposed electro-mechanically coupled framework, we analytically solve a non-homogeneous boundary value problem, the extension and inflation of an axisymmetric cylindrical tube under electro-mechanically coupled load. The results capture various electro-mechanical couplings with the formulation proposed for EAP composites.
Development of a statistical oil spill model for risk assessment.
Guo, Weijun
2017-11-01
To gain a better understanding of the impacts from potential risk sources, we developed an oil spill model using probabilistic method, which simulates numerous oil spill trajectories under varying environmental conditions. The statistical results were quantified from hypothetical oil spills under multiple scenarios, including area affected probability, mean oil slick thickness, and duration of water surface exposed to floating oil. The three sub-indices together with marine area vulnerability are merged to compute the composite index, characterizing the spatial distribution of risk degree. Integral of the index can be used to identify the overall risk from an emission source. The developed model has been successfully applied in comparison to and selection of an appropriate oil port construction location adjacent to a marine protected area for Phoca largha in China. The results highlight the importance of selection of candidates before project construction, since that risk estimation from two adjacent potential sources may turn out to be significantly different regarding hydrodynamic conditions and eco-environmental sensitivity. Copyright © 2017. Published by Elsevier Ltd.
Link, William A; Barker, Richard J
2005-03-01
We present a hierarchical extension of the Cormack-Jolly-Seber (CJS) model for open population capture-recapture data. In addition to recaptures of marked animals, we model first captures of animals and losses on capture. The parameter set includes capture probabilities, survival rates, and birth rates. The survival rates and birth rates are treated as a random sample from a bivariate distribution, thus the model explicitly incorporates correlation in these demographic rates. A key feature of the model is that the likelihood function, which includes a CJS model factor, is expressed entirely in terms of identifiable parameters; losses on capture can be factored out of the model. Since the computational complexity of classical likelihood methods is prohibitive, we use Markov chain Monte Carlo in a Bayesian analysis. We describe an efficient candidate-generation scheme for Metropolis-Hastings sampling of CJS models and extensions. The procedure is illustrated using mark-recapture data for the moth Gonodontis bidentata.
Link, William A.; Barker, Richard J.
2005-01-01
We present a hierarchical extension of the Cormack–Jolly–Seber (CJS) model for open population capture–recapture data. In addition to recaptures of marked animals, we model first captures of animals and losses on capture. The parameter set includes capture probabilities, survival rates, and birth rates. The survival rates and birth rates are treated as a random sample from a bivariate distribution, thus the model explicitly incorporates correlation in these demographic rates. A key feature of the model is that the likelihood function, which includes a CJS model factor, is expressed entirely in terms of identifiable parameters; losses on capture can be factored out of the model. Since the computational complexity of classical likelihood methods is prohibitive, we use Markov chain Monte Carlo in a Bayesian analysis. We describe an efficient candidate-generation scheme for Metropolis–Hastings sampling of CJS models and extensions. The procedure is illustrated using mark-recapture data for the moth Gonodontis bidentata.
Bayesian alternative to the ISO-GUM's use of the Welch Satterthwaite formula
NASA Astrophysics Data System (ADS)
Kacker, Raghu N.
2006-02-01
In certain disciplines, uncertainty is traditionally expressed as an interval about an estimate for the value of the measurand. Development of such uncertainty intervals with a stated coverage probability based on the International Organization for Standardization (ISO) Guide to the Expression of Uncertainty in Measurement (GUM) requires a description of the probability distribution for the value of the measurand. The ISO-GUM propagates the estimates and their associated standard uncertainties for various input quantities through a linear approximation of the measurement equation to determine an estimate and its associated standard uncertainty for the value of the measurand. This procedure does not yield a probability distribution for the value of the measurand. The ISO-GUM suggests that under certain conditions motivated by the central limit theorem the distribution for the value of the measurand may be approximated by a scaled-and-shifted t-distribution with effective degrees of freedom obtained from the Welch-Satterthwaite (W-S) formula. The approximate t-distribution may then be used to develop an uncertainty interval with a stated coverage probability for the value of the measurand. We propose an approximate normal distribution based on a Bayesian uncertainty as an alternative to the t-distribution based on the W-S formula. A benefit of the approximate normal distribution based on a Bayesian uncertainty is that it greatly simplifies the expression of uncertainty by eliminating altogether the need for calculating effective degrees of freedom from the W-S formula. In the special case where the measurand is the difference between two means, each evaluated from statistical analyses of independent normally distributed measurements with unknown and possibly unequal variances, the probability distribution for the value of the measurand is known to be a Behrens-Fisher distribution. We compare the performance of the approximate normal distribution based on a Bayesian uncertainty and the approximate t-distribution based on the W-S formula with respect to the Behrens-Fisher distribution. The approximate normal distribution is simpler and better in this case. A thorough investigation of the relative performance of the two approximate distributions would require comparison for a range of measurement equations by numerical methods.
Does Litter Size Variation Affect Models of Terrestrial Carnivore Extinction Risk and Management?
Devenish-Nelson, Eleanor S.; Stephens, Philip A.; Harris, Stephen; Soulsbury, Carl; Richards, Shane A.
2013-01-01
Background Individual variation in both survival and reproduction has the potential to influence extinction risk. Especially for rare or threatened species, reliable population models should adequately incorporate demographic uncertainty. Here, we focus on an important form of demographic stochasticity: variation in litter sizes. We use terrestrial carnivores as an example taxon, as they are frequently threatened or of economic importance. Since data on intraspecific litter size variation are often sparse, it is unclear what probability distribution should be used to describe the pattern of litter size variation for multiparous carnivores. Methodology/Principal Findings We used litter size data on 32 terrestrial carnivore species to test the fit of 12 probability distributions. The influence of these distributions on quasi-extinction probabilities and the probability of successful disease control was then examined for three canid species – the island fox Urocyon littoralis, the red fox Vulpes vulpes, and the African wild dog Lycaon pictus. Best fitting probability distributions differed among the carnivores examined. However, the discretised normal distribution provided the best fit for the majority of species, because variation among litter-sizes was often small. Importantly, however, the outcomes of demographic models were generally robust to the distribution used. Conclusion/Significance These results provide reassurance for those using demographic modelling for the management of less studied carnivores in which litter size variation is estimated using data from species with similar reproductive attributes. PMID:23469140
Does litter size variation affect models of terrestrial carnivore extinction risk and management?
Devenish-Nelson, Eleanor S; Stephens, Philip A; Harris, Stephen; Soulsbury, Carl; Richards, Shane A
2013-01-01
Individual variation in both survival and reproduction has the potential to influence extinction risk. Especially for rare or threatened species, reliable population models should adequately incorporate demographic uncertainty. Here, we focus on an important form of demographic stochasticity: variation in litter sizes. We use terrestrial carnivores as an example taxon, as they are frequently threatened or of economic importance. Since data on intraspecific litter size variation are often sparse, it is unclear what probability distribution should be used to describe the pattern of litter size variation for multiparous carnivores. We used litter size data on 32 terrestrial carnivore species to test the fit of 12 probability distributions. The influence of these distributions on quasi-extinction probabilities and the probability of successful disease control was then examined for three canid species - the island fox Urocyon littoralis, the red fox Vulpes vulpes, and the African wild dog Lycaon pictus. Best fitting probability distributions differed among the carnivores examined. However, the discretised normal distribution provided the best fit for the majority of species, because variation among litter-sizes was often small. Importantly, however, the outcomes of demographic models were generally robust to the distribution used. These results provide reassurance for those using demographic modelling for the management of less studied carnivores in which litter size variation is estimated using data from species with similar reproductive attributes.
Planet Hunters: New Kepler Planet Candidates from Analysis of Quarter 2
NASA Astrophysics Data System (ADS)
Lintott, Chris J.; Schwamb, Megan E.; Barclay, Thomas; Sharzer, Charlie; Fischer, Debra A.; Brewer, John; Giguere, Matthew; Lynn, Stuart; Parrish, Michael; Batalha, Natalie; Bryson, Steve; Jenkins, Jon; Ragozzine, Darin; Rowe, Jason F.; Schwainski, Kevin; Gagliano, Robert; Gilardi, Joe; Jek, Kian J.; Pääkkönen, Jari-Pekka; Smits, Tjapko
2013-06-01
We present new planet candidates identified in NASA Kepler Quarter 2 public release data by volunteers engaged in the Planet Hunters citizen science project. The two candidates presented here survive checks for false positives, including examination of the pixel offset to constrain the possibility of a background eclipsing binary. The orbital periods of the planet candidates are 97.46 days (KIC 4552729) and 284.03 (KIC 10005758) days and the modeled planet radii are 5.3 and 3.8 R ⊕. The latter star has an additional known planet candidate with a radius of 5.05 R ⊕ and a period of 134.49 days, which was detected by the Kepler pipeline. The discovery of these candidates illustrates the value of massively distributed volunteer review of the Kepler database to recover candidates which were otherwise uncataloged. .
Mitra, Rajib; Jordan, Michael I.; Dunbrack, Roland L.
2010-01-01
Distributions of the backbone dihedral angles of proteins have been studied for over 40 years. While many statistical analyses have been presented, only a handful of probability densities are publicly available for use in structure validation and structure prediction methods. The available distributions differ in a number of important ways, which determine their usefulness for various purposes. These include: 1) input data size and criteria for structure inclusion (resolution, R-factor, etc.); 2) filtering of suspect conformations and outliers using B-factors or other features; 3) secondary structure of input data (e.g., whether helix and sheet are included; whether beta turns are included); 4) the method used for determining probability densities ranging from simple histograms to modern nonparametric density estimation; and 5) whether they include nearest neighbor effects on the distribution of conformations in different regions of the Ramachandran map. In this work, Ramachandran probability distributions are presented for residues in protein loops from a high-resolution data set with filtering based on calculated electron densities. Distributions for all 20 amino acids (with cis and trans proline treated separately) have been determined, as well as 420 left-neighbor and 420 right-neighbor dependent distributions. The neighbor-independent and neighbor-dependent probability densities have been accurately estimated using Bayesian nonparametric statistical analysis based on the Dirichlet process. In particular, we used hierarchical Dirichlet process priors, which allow sharing of information between densities for a particular residue type and different neighbor residue types. The resulting distributions are tested in a loop modeling benchmark with the program Rosetta, and are shown to improve protein loop conformation prediction significantly. The distributions are available at http://dunbrack.fccc.edu/hdp. PMID:20442867
Probabilistic analysis of preload in the abutment screw of a dental implant complex.
Guda, Teja; Ross, Thomas A; Lang, Lisa A; Millwater, Harry R
2008-09-01
Screw loosening is a problem for a percentage of implants. A probabilistic analysis to determine the cumulative probability distribution of the preload, the probability of obtaining an optimal preload, and the probabilistic sensitivities identifying important variables is lacking. The purpose of this study was to examine the inherent variability of material properties, surface interactions, and applied torque in an implant system to determine the probability of obtaining desired preload values and to identify the significant variables that affect the preload. Using software programs, an abutment screw was subjected to a tightening torque and the preload was determined from finite element (FE) analysis. The FE model was integrated with probabilistic analysis software. Two probabilistic analysis methods (advanced mean value and Monte Carlo sampling) were applied to determine the cumulative distribution function (CDF) of preload. The coefficient of friction, elastic moduli, Poisson's ratios, and applied torque were modeled as random variables and defined by probability distributions. Separate probability distributions were determined for the coefficient of friction in well-lubricated and dry environments. The probabilistic analyses were performed and the cumulative distribution of preload was determined for each environment. A distinct difference was seen between the preload probability distributions generated in a dry environment (normal distribution, mean (SD): 347 (61.9) N) compared to a well-lubricated environment (normal distribution, mean (SD): 616 (92.2) N). The probability of obtaining a preload value within the target range was approximately 54% for the well-lubricated environment and only 0.02% for the dry environment. The preload is predominately affected by the applied torque and coefficient of friction between the screw threads and implant bore at lower and middle values of the preload CDF, and by the applied torque and the elastic modulus of the abutment screw at high values of the preload CDF. Lubrication at the threaded surfaces between the abutment screw and implant bore affects the preload developed in the implant complex. For the well-lubricated surfaces, only approximately 50% of implants will have preload values within the generally accepted range. This probability can be improved by applying a higher torque than normally recommended or a more closely controlled torque than typically achieved. It is also suggested that materials with higher elastic moduli be used in the manufacture of the abutment screw to achieve a higher preload.
Comparative analysis through probability distributions of a data set
NASA Astrophysics Data System (ADS)
Cristea, Gabriel; Constantinescu, Dan Mihai
2018-02-01
In practice, probability distributions are applied in such diverse fields as risk analysis, reliability engineering, chemical engineering, hydrology, image processing, physics, market research, business and economic research, customer support, medicine, sociology, demography etc. This article highlights important aspects of fitting probability distributions to data and applying the analysis results to make informed decisions. There are a number of statistical methods available which can help us to select the best fitting model. Some of the graphs display both input data and fitted distributions at the same time, as probability density and cumulative distribution. The goodness of fit tests can be used to determine whether a certain distribution is a good fit. The main used idea is to measure the "distance" between the data and the tested distribution, and compare that distance to some threshold values. Calculating the goodness of fit statistics also enables us to order the fitted distributions accordingly to how good they fit to data. This particular feature is very helpful for comparing the fitted models. The paper presents a comparison of most commonly used goodness of fit tests as: Kolmogorov-Smirnov, Anderson-Darling, and Chi-Squared. A large set of data is analyzed and conclusions are drawn by visualizing the data, comparing multiple fitted distributions and selecting the best model. These graphs should be viewed as an addition to the goodness of fit tests.
DISTRIBUTION SYSTEM EVALUATIONS OF THE CONTAMINANT CANDIDATE LIST
The Safe Drinking Water Act (SDWA), as amended in 1996, requires EPA to establish a list of unregulated microbiological and chemical contaminants to aid in priority setting for the Agency's drinking water program. The list is referred to as the Contaminant Candidate List (CCL). A...
Impact of temporal probability in 4D dose calculation for lung tumors.
Rouabhi, Ouided; Ma, Mingyu; Bayouth, John; Xia, Junyi
2015-11-08
The purpose of this study was to evaluate the dosimetric uncertainty in 4D dose calculation using three temporal probability distributions: uniform distribution, sinusoidal distribution, and patient-specific distribution derived from the patient respiratory trace. Temporal probability, defined as the fraction of time a patient spends in each respiratory amplitude, was evaluated in nine lung cancer patients. Four-dimensional computed tomography (4D CT), along with deformable image registration, was used to compute 4D dose incorporating the patient's respiratory motion. First, the dose of each of 10 phase CTs was computed using the same planning parameters as those used in 3D treatment planning based on the breath-hold CT. Next, deformable image registration was used to deform the dose of each phase CT to the breath-hold CT using the deformation map between the phase CT and the breath-hold CT. Finally, the 4D dose was computed by summing the deformed phase doses using their corresponding temporal probabilities. In this study, 4D dose calculated from the patient-specific temporal probability distribution was used as the ground truth. The dosimetric evaluation matrix included: 1) 3D gamma analysis, 2) mean tumor dose (MTD), 3) mean lung dose (MLD), and 4) lung V20. For seven out of nine patients, both uniform and sinusoidal temporal probability dose distributions were found to have an average gamma passing rate > 95% for both the lung and PTV regions. Compared with 4D dose calculated using the patient respiratory trace, doses using uniform and sinusoidal distribution showed a percentage difference on average of -0.1% ± 0.6% and -0.2% ± 0.4% in MTD, -0.2% ± 1.9% and -0.2% ± 1.3% in MLD, 0.09% ± 2.8% and -0.07% ± 1.8% in lung V20, -0.1% ± 2.0% and 0.08% ± 1.34% in lung V10, 0.47% ± 1.8% and 0.19% ± 1.3% in lung V5, respectively. We concluded that four-dimensional dose computed using either a uniform or sinusoidal temporal probability distribution can approximate four-dimensional dose computed using the patient-specific respiratory trace.
Multi-model ensembles for assessment of flood losses and associated uncertainty
NASA Astrophysics Data System (ADS)
Figueiredo, Rui; Schröter, Kai; Weiss-Motz, Alexander; Martina, Mario L. V.; Kreibich, Heidi
2018-05-01
Flood loss modelling is a crucial part of risk assessments. However, it is subject to large uncertainty that is often neglected. Most models available in the literature are deterministic, providing only single point estimates of flood loss, and large disparities tend to exist among them. Adopting any one such model in a risk assessment context is likely to lead to inaccurate loss estimates and sub-optimal decision-making. In this paper, we propose the use of multi-model ensembles to address these issues. This approach, which has been applied successfully in other scientific fields, is based on the combination of different model outputs with the aim of improving the skill and usefulness of predictions. We first propose a model rating framework to support ensemble construction, based on a probability tree of model properties, which establishes relative degrees of belief between candidate models. Using 20 flood loss models in two test cases, we then construct numerous multi-model ensembles, based both on the rating framework and on a stochastic method, differing in terms of participating members, ensemble size and model weights. We evaluate the performance of ensemble means, as well as their probabilistic skill and reliability. Our results demonstrate that well-designed multi-model ensembles represent a pragmatic approach to consistently obtain more accurate flood loss estimates and reliable probability distributions of model uncertainty.
Discriminative Random Field Models for Subsurface Contamination Uncertainty Quantification
NASA Astrophysics Data System (ADS)
Arshadi, M.; Abriola, L. M.; Miller, E. L.; De Paolis Kaluza, C.
2017-12-01
Application of flow and transport simulators for prediction of the release, entrapment, and persistence of dense non-aqueous phase liquids (DNAPLs) and associated contaminant plumes is a computationally intensive process that requires specification of a large number of material properties and hydrologic/chemical parameters. Given its computational burden, this direct simulation approach is particularly ill-suited for quantifying both the expected performance and uncertainty associated with candidate remediation strategies under real field conditions. Prediction uncertainties primarily arise from limited information about contaminant mass distributions, as well as the spatial distribution of subsurface hydrologic properties. Application of direct simulation to quantify uncertainty would, thus, typically require simulating multiphase flow and transport for a large number of permeability and release scenarios to collect statistics associated with remedial effectiveness, a computationally prohibitive process. The primary objective of this work is to develop and demonstrate a methodology that employs measured field data to produce equi-probable stochastic representations of a subsurface source zone that capture the spatial distribution and uncertainty associated with key features that control remediation performance (i.e., permeability and contamination mass). Here we employ probabilistic models known as discriminative random fields (DRFs) to synthesize stochastic realizations of initial mass distributions consistent with known, and typically limited, site characterization data. Using a limited number of full scale simulations as training data, a statistical model is developed for predicting the distribution of contaminant mass (e.g., DNAPL saturation and aqueous concentration) across a heterogeneous domain. Monte-Carlo sampling methods are then employed, in conjunction with the trained statistical model, to generate realizations conditioned on measured borehole data. Performance of the statistical model is illustrated through comparisons of generated realizations with the `true' numerical simulations. Finally, we demonstrate how these realizations can be used to determine statistically optimal locations for further interrogation of the subsurface.
Li, Xukai; Liao, Haofeng; Fan, Chunfen; Hu, Huizhen; Li, Ying; Li, Jing; Yi, Zili; Cai, Xiwen; Peng, Liangcai; Tu, Yuanyuan
2016-01-01
Miscanthus is a leading bioenergy candidate for biofuels, and it thus becomes essential to characterize the desire natural Miscanthus germplasm accessions with high biomass saccharification. In this study, total 171 natural Miscanthus accessions were geographically mapped using public database. According to the equation [P(H/L| East) = P(H/L∩East)/P(East)], the probability (P) parameters were calculated on relationships between geographical distributions of Miscanthus accessions in the East of China, and related factors with high(H) or low(L) values including biomass saccahrification under 1% NaOH and 1% H2SO4 pretreatments, lignocellulose features and climate conditions. Based on the maximum P value, a golden cutting line was generated from 42°25’ N, 108°22’ E to 22°58’ N, 116°28’ E on the original locations of Miscanthus accessions with high P(H|East) values (0.800–0.813), indicating that more than 90% Miscanthus accessions were originally located in the East with high biomass saccharification. Furthermore, the averaged insolation showed high P (H|East) and P(East|H) values at 0.782 and 0.754, whereas other climate factors had low P(East|H) values, suggesting that the averaged insolation is unique factor on Miscanthus distributions for biomass saccharification. In terms of cell wall compositions and wall polymer features, both hemicelluloses level and cellulose crystallinity (CrI) of Miscanthus accessions exhibited relative high P values, suggesting that they should be the major factors accounting for geographic distributions of Miscanthus accessions with high biomass digestibility. PMID:27532636
NASA Astrophysics Data System (ADS)
Yamada, Yuhei; Yamazaki, Yoshihiro
2018-04-01
This study considered a stochastic model for cluster growth in a Markov process with a cluster size dependent additive noise. According to this model, the probability distribution of the cluster size transiently becomes an exponential or a log-normal distribution depending on the initial condition of the growth. In this letter, a master equation is obtained for this model, and derivation of the distributions is discussed.
NASA Astrophysics Data System (ADS)
Sato, Aki-Hiro
2010-12-01
This study considers q-Gaussian distributions and stochastic differential equations with both multiplicative and additive noises. In the M-dimensional case a q-Gaussian distribution can be theoretically derived as a stationary probability distribution of the multiplicative stochastic differential equation with both mutually independent multiplicative and additive noises. By using the proposed stochastic differential equation a method to evaluate a default probability under a given risk buffer is proposed.
Net present value probability distributions from decline curve reserves estimates
DOE Office of Scientific and Technical Information (OSTI.GOV)
Simpson, D.E.; Huffman, C.H.; Thompson, R.S.
1995-12-31
This paper demonstrates how reserves probability distributions can be used to develop net present value (NPV) distributions. NPV probability distributions were developed from the rate and reserves distributions presented in SPE 28333. This real data study used practicing engineer`s evaluations of production histories. Two approaches were examined to quantify portfolio risk. The first approach, the NPV Relative Risk Plot, compares the mean NPV with the NPV relative risk ratio for the portfolio. The relative risk ratio is the NPV standard deviation (a) divided the mean ({mu}) NPV. The second approach, a Risk - Return Plot, is a plot of themore » {mu} discounted cash flow rate of return (DCFROR) versus the {sigma} for the DCFROR distribution. This plot provides a risk-return relationship for comparing various portfolios. These methods may help evaluate property acquisition and divestiture alternatives and assess the relative risk of a suite of wells or fields for bank loans.« less
Optimal random search for a single hidden target.
Snider, Joseph
2011-01-01
A single target is hidden at a location chosen from a predetermined probability distribution. Then, a searcher must find a second probability distribution from which random search points are sampled such that the target is found in the minimum number of trials. Here it will be shown that if the searcher must get very close to the target to find it, then the best search distribution is proportional to the square root of the target distribution regardless of dimension. For a Gaussian target distribution, the optimum search distribution is approximately a Gaussian with a standard deviation that varies inversely with how close the searcher must be to the target to find it. For a network where the searcher randomly samples nodes and looks for the fixed target along edges, the optimum is either to sample a node with probability proportional to the square root of the out-degree plus 1 or not to do so at all.
Optimizing exoplanet transit searches
NASA Astrophysics Data System (ADS)
Herrero, E.; Ribas, I.; Jordi, C.
2013-05-01
Exoplanet searches using the transit technique are nowadays providing a great number of findings. Most exoplanet transit detection programs that are currently underway are focused on large catalogs of stars with no pre-selection. This necessarily makes such surveys quite inefficient, because huge amounts of data are processed for a relatively low transiting planet yield. In this work we investigate a method to increase the efficiency of a targeted exoplanet search with the transit technique by preselecting a subset of candidates from large catalogs of stars. Assuming spin-orbit alignment, this can be done by considering stars that have higher probability to be oriented nearly equator-on (inclination close to 90°). We use activity-rotation velocity relations for low-mass stars to study the dependence of the position in the activity - v sin(i) diagram on the stellar axis inclination. We compose a catalog of G-, K-, M-type main sequence simulated stars using isochrones, an isotropic inclination distribution and empirical relations to obtain their rotation periods and activity indexes. Then the activity-vsini diagram is filled and statistics are applied to trace the areas containing the higher ratio of stars with inclinations above 80°. A similar statistics is applied to stars from real catalogs with log(R'_{HK}) and v sin(i) data to find their probability of being equator-on. We present the method used to generate the simulated star catalog and the subsequent statistics to find the highly inclined stars from real catalogs using the activity-v sin(i) diagram. Several catalogs from the literature are analysed and a subsample of stars with the highest probability of being equator-on is presented. Assuming spin-orbit alignment, the efficiency of an exoplanet transit search in the resulting subsample of probably highly inclined stars is estimated to be two to three times higher than with a global search with no pre-selection.
Neural networks for oil spill detection using TerraSAR-X data
NASA Astrophysics Data System (ADS)
Avezzano, Ruggero G.; Velotto, Domenico; Soccorsi, Matteo; Del Frate, Fabio; Lehner, Susanne
2011-11-01
The increased amount of available Synthetic Aperture Radar (SAR) images involves a growing workload on the operators at analysis centers. In addition, even if the operators go through extensive training to learn manual oil spill detection, they can provide different and subjective responses. Hence, the upgrade and improvements of algorithms for automatic detection that can help in screening the images and prioritizing the alarms are of great benefit. In this paper we present the potentialities of TerraSAR-X (TS-X) data and Neural Network algorithms for oil spills detection. The radar on board satellite TS-X provides X-band images with a resolution of up to 1m. Such resolution can be very effective in the monitoring of coastal areas to prevent sea oil pollution. The network input is a vector containing the values of a set of features characterizing an oil spill candidate. The network output gives the probability for the candidate to be a real oil spill. Candidates with a probability less than 50% are classified as look-alikes. The overall classification performances have been evaluated on a data set of 50 TS-X images containing more than 150 examples of certified oil spills and well-known look-alikes (e.g. low wind areas, wind shadows, biogenic films). The preliminary classification results are satisfactory with an overall detection accuracy above 80%.
Data Interpretation: Using Probability
ERIC Educational Resources Information Center
Drummond, Gordon B.; Vowler, Sarah L.
2011-01-01
Experimental data are analysed statistically to allow researchers to draw conclusions from a limited set of measurements. The hard fact is that researchers can never be certain that measurements from a sample will exactly reflect the properties of the entire group of possible candidates available to be studied (although using a sample is often the…
NASA Astrophysics Data System (ADS)
Mahanti, P.; Robinson, M. S.; Boyd, A. K.
2013-12-01
Craters ~20-km diameter and above significantly shaped the lunar landscape. The statistical nature of the slope distribution on their walls and floors dominate the overall slope distribution statistics for the lunar surface. Slope statistics are inherently useful for characterizing the current topography of the surface, determining accurate photometric and surface scattering properties, and in defining lunar surface trafficability [1-4]. Earlier experimental studies on the statistical nature of lunar surface slopes were restricted either by resolution limits (Apollo era photogrammetric studies) or by model error considerations (photoclinometric and radar scattering studies) where the true nature of slope probability distribution was not discernible at baselines smaller than a kilometer[2,3,5]. Accordingly, historical modeling of lunar surface slopes probability distributions for applications such as in scattering theory development or rover traversability assessment is more general in nature (use of simple statistical models such as the Gaussian distribution[1,2,5,6]). With the advent of high resolution, high precision topographic models of the Moon[7,8], slopes in lunar craters can now be obtained at baselines as low as 6-meters allowing unprecedented multi-scale (multiple baselines) modeling possibilities for slope probability distributions. Topographic analysis (Lunar Reconnaissance Orbiter Camera (LROC) Narrow Angle Camera (NAC) 2-m digital elevation models (DEM)) of ~20-km diameter Copernican lunar craters revealed generally steep slopes on interior walls (30° to 36°, locally exceeding 40°) over 15-meter baselines[9]. In this work, we extend the analysis from a probability distribution modeling point-of-view with NAC DEMs to characterize the slope statistics for the floors and walls for the same ~20-km Copernican lunar craters. The difference in slope standard deviations between the Gaussian approximation and the actual distribution (2-meter sampling) was computed over multiple scales. This slope analysis showed that local slope distributions are non-Gaussian for both crater walls and floors. Over larger baselines (~100 meters), crater wall slope probability distributions do approximate Gaussian distributions better, but have long distribution tails. Crater floor probability distributions however, were always asymmetric (for the baseline scales analyzed) and less affected by baseline scale variations. Accordingly, our results suggest that use of long tailed probability distributions (like Cauchy) and a baseline-dependant multi-scale model can be more effective in describing the slope statistics for lunar topography. Refrences: [1]Moore, H.(1971), JGR,75(11) [2]Marcus, A. H.(1969),JGR,74 (22).[3]R.J. Pike (1970),U.S. Geological Survey Working Paper [4]N. C. Costes, J. E. Farmer and E. B. George (1972),NASA Technical Report TR R-401 [5]M. N. Parker and G. L. Tyler(1973), Radio Science, 8(3),177-184 [6]Alekseev, V. A.et al (1968), Soviet Astronomy, Vol. 11, p.860 [7]Burns et al. (2012) Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci., XXXIX-B4, 483-488.[8]Smith et al. (2010) GRL 37, L18204, DOI: 10.1029/2010GL043751. [9]Wagner R., Robinson, M., Speyerer E., Mahanti, P., LPSC 2013, #2924.
NASA Technical Reports Server (NTRS)
Lanzi, R. James; Vincent, Brett T.
1993-01-01
The relationship between actual and predicted re-entry maximum dynamic pressure is characterized using a probability density function and a cumulative distribution function derived from sounding rocket flight data. This paper explores the properties of this distribution and demonstrates applications of this data with observed sounding rocket re-entry body damage characteristics to assess probabilities of sustaining various levels of heating damage. The results from this paper effectively bridge the gap existing in sounding rocket reentry analysis between the known damage level/flight environment relationships and the predicted flight environment.
Probability and the changing shape of response distributions for orientation.
Anderson, Britt
2014-11-18
Spatial attention and feature-based attention are regarded as two independent mechanisms for biasing the processing of sensory stimuli. Feature attention is held to be a spatially invariant mechanism that advantages a single feature per sensory dimension. In contrast to the prediction of location independence, I found that participants were able to report the orientation of a briefly presented visual grating better for targets defined by high probability conjunctions of features and locations even when orientations and locations were individually uniform. The advantage for high-probability conjunctions was accompanied by changes in the shape of the response distributions. High-probability conjunctions had error distributions that were not normally distributed but demonstrated increased kurtosis. The increase in kurtosis could be explained as a change in the variances of the component tuning functions that comprise a population mixture. By changing the mixture distribution of orientation-tuned neurons, it is possible to change the shape of the discrimination function. This prompts the suggestion that attention may not "increase" the quality of perceptual processing in an absolute sense but rather prioritizes some stimuli over others. This results in an increased number of highly accurate responses to probable targets and, simultaneously, an increase in the number of very inaccurate responses. © 2014 ARVO.
NASA Technical Reports Server (NTRS)
Smith, O. E.
1976-01-01
The techniques are presented to derive several statistical wind models. The techniques are from the properties of the multivariate normal probability function. Assuming that the winds can be considered as bivariate normally distributed, then (1) the wind components and conditional wind components are univariate normally distributed, (2) the wind speed is Rayleigh distributed, (3) the conditional distribution of wind speed given a wind direction is Rayleigh distributed, and (4) the frequency of wind direction can be derived. All of these distributions are derived from the 5-sample parameter of wind for the bivariate normal distribution. By further assuming that the winds at two altitudes are quadravariate normally distributed, then the vector wind shear is bivariate normally distributed and the modulus of the vector wind shear is Rayleigh distributed. The conditional probability of wind component shears given a wind component is normally distributed. Examples of these and other properties of the multivariate normal probability distribution function as applied to Cape Kennedy, Florida, and Vandenberg AFB, California, wind data samples are given. A technique to develop a synthetic vector wind profile model of interest to aerospace vehicle applications is presented.
Joint probabilities and quantum cognition
NASA Astrophysics Data System (ADS)
de Barros, J. Acacio
2012-12-01
In this paper we discuss the existence of joint probability distributions for quantumlike response computations in the brain. We do so by focusing on a contextual neural-oscillator model shown to reproduce the main features of behavioral stimulus-response theory. We then exhibit a simple example of contextual random variables not having a joint probability distribution, and describe how such variables can be obtained from neural oscillators, but not from a quantum observable algebra.
Path Diversity Improved Opportunistic Routing for Underwater Sensor Networks
Wang, Haiyan; He, Ke
2018-01-01
The packets carried along a pre-defined route in underwater sensor networks are very vulnerble. Node mobility or intermittent channel availability easily leads to unreachable routing. Opportunistic routing has been proven to be a promising paradigm to design routing protocols for underwater sensor networks. It takes advantage of the broadcast nature of the wireless medium to combat packet losses and selects potential paths on the fly. Finding an appropriate forwarding candidate set is a key issue in opportunistic routing. Many existing solutions ignore the impact of candidates location distribution on packet forwarding. In this paper, a path diversity improved candidate selection strategy is applied in opportunistic routing to improve packet forwarding efficiency. It not only maximizes the packet forwarding advancements but also takes the candidate’s location distribution into account. Based on this strategy, we propose two effective routing protocols: position improved candidates selection (PICS) and position random candidates selection (PRCS). PICS employs two-hop neighbor information to make routing decisions. PRCS only uses one-hop neighbor information. Simulation results show that both PICS and PRCS can significantly improve network performance when compared with the previous solutions, in terms of packet delivery ratio, average energy consumption and end-to-end delay. PMID:29690621
Zhuang, Jiancang; Ogata, Yosihiko
2006-04-01
The space-time epidemic-type aftershock sequence model is a stochastic branching process in which earthquake activity is classified into background and clustering components and each earthquake triggers other earthquakes independently according to certain rules. This paper gives the probability distributions associated with the largest event in a cluster and their properties for all three cases when the process is subcritical, critical, and supercritical. One of the direct uses of these probability distributions is to evaluate the probability of an earthquake to be a foreshock, and magnitude distributions of foreshocks and nonforeshock earthquakes. To verify these theoretical results, the Japan Meteorological Agency earthquake catalog is analyzed. The proportion of events that have 1 or more larger descendants in total events is found to be as high as about 15%. When the differences between background events and triggered event in the behavior of triggering children are considered, a background event has a probability about 8% to be a foreshock. This probability decreases when the magnitude of the background event increases. These results, obtained from a complicated clustering model, where the characteristics of background events and triggered events are different, are consistent with the results obtained in [Ogata, Geophys. J. Int. 127, 17 (1996)] by using the conventional single-linked cluster declustering method.
Herschel Key Program Heritage: a Far-Infrared Source Catalog for the Magellanic Clouds
NASA Astrophysics Data System (ADS)
Seale, Jonathan P.; Meixner, Margaret; Sewiło, Marta; Babler, Brian; Engelbracht, Charles W.; Gordon, Karl; Hony, Sacha; Misselt, Karl; Montiel, Edward; Okumura, Koryo; Panuzzo, Pasquale; Roman-Duval, Julia; Sauvage, Marc; Boyer, Martha L.; Chen, C.-H. Rosie; Indebetouw, Remy; Matsuura, Mikako; Oliveira, Joana M.; Srinivasan, Sundar; van Loon, Jacco Th.; Whitney, Barbara; Woods, Paul M.
2014-12-01
Observations from the HERschel Inventory of the Agents of Galaxy Evolution (HERITAGE) have been used to identify dusty populations of sources in the Large and Small Magellanic Clouds (LMC and SMC). We conducted the study using the HERITAGE catalogs of point sources available from the Herschel Science Center from both the Photodetector Array Camera and Spectrometer (PACS; 100 and 160 μm) and Spectral and Photometric Imaging Receiver (SPIRE; 250, 350, and 500 μm) cameras. These catalogs are matched to each other to create a Herschel band-merged catalog and then further matched to archival Spitzer IRAC and MIPS catalogs from the Spitzer Surveying the Agents of Galaxy Evolution (SAGE) and SAGE-SMC surveys to create single mid- to far-infrared (far-IR) point source catalogs that span the wavelength range from 3.6 to 500 μm. There are 35,322 unique sources in the LMC and 7503 in the SMC. To be bright in the FIR, a source must be very dusty, and so the sources in the HERITAGE catalogs represent the dustiest populations of sources. The brightest HERITAGE sources are dominated by young stellar objects (YSOs), and the dimmest by background galaxies. We identify the sources most likely to be background galaxies by first considering their morphology (distant galaxies are point-like at the resolution of Herschel) and then comparing the flux distribution to that of the Herschel Astrophysical Terahertz Large Area Survey (ATLAS) survey of galaxies. We find a total of 9745 background galaxy candidates in the LMC HERITAGE images and 5111 in the SMC images, in agreement with the number predicted by extrapolating from the ATLAS flux distribution. The majority of the Magellanic Cloud-residing sources are either very young, embedded forming stars or dusty clumps of the interstellar medium. Using the presence of 24 μm emission as a tracer of star formation, we identify 3518 YSO candidates in the LMC and 663 in the SMC. There are far fewer far-IR bright YSOs in the SMC than the LMC due to both the SMC's smaller size and its lower dust content. The YSO candidate lists may be contaminated at low flux levels by background galaxies, and so we differentiate between sources with a high (“probable”) and moderate (“possible”) likelihood of being a YSO. There are 2493/425 probable YSO candidates in the LMC/SMC. Approximately 73% of the Herschel YSO candidates are newly identified in the LMC, and 35% in the SMC. We further identify a small population of dusty objects in the late stages of stellar evolution including extreme and post-asymptotic giant branch, planetary nebulae, and supernova remnants. These populations are identified by matching the HERITAGE catalogs to lists of previously identified objects in the literature. Approximately half of the LMC sources and one quarter of the SMC sources are too faint to obtain accurate ample FIR photometry and are unclassified.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Seale, Jonathan P.; Meixner, Margaret; Sewiło, Marta
Observations from the HERschel Inventory of the Agents of Galaxy Evolution (HERITAGE) have been used to identify dusty populations of sources in the Large and Small Magellanic Clouds (LMC and SMC). We conducted the study using the HERITAGE catalogs of point sources available from the Herschel Science Center from both the Photodetector Array Camera and Spectrometer (PACS; 100 and 160 μm) and Spectral and Photometric Imaging Receiver (SPIRE; 250, 350, and 500 μm) cameras. These catalogs are matched to each other to create a Herschel band-merged catalog and then further matched to archival Spitzer IRAC and MIPS catalogs from themore » Spitzer Surveying the Agents of Galaxy Evolution (SAGE) and SAGE-SMC surveys to create single mid- to far-infrared (far-IR) point source catalogs that span the wavelength range from 3.6 to 500 μm. There are 35,322 unique sources in the LMC and 7503 in the SMC. To be bright in the FIR, a source must be very dusty, and so the sources in the HERITAGE catalogs represent the dustiest populations of sources. The brightest HERITAGE sources are dominated by young stellar objects (YSOs), and the dimmest by background galaxies. We identify the sources most likely to be background galaxies by first considering their morphology (distant galaxies are point-like at the resolution of Herschel) and then comparing the flux distribution to that of the Herschel Astrophysical Terahertz Large Area Survey (ATLAS) survey of galaxies. We find a total of 9745 background galaxy candidates in the LMC HERITAGE images and 5111 in the SMC images, in agreement with the number predicted by extrapolating from the ATLAS flux distribution. The majority of the Magellanic Cloud-residing sources are either very young, embedded forming stars or dusty clumps of the interstellar medium. Using the presence of 24 μm emission as a tracer of star formation, we identify 3518 YSO candidates in the LMC and 663 in the SMC. There are far fewer far-IR bright YSOs in the SMC than the LMC due to both the SMC's smaller size and its lower dust content. The YSO candidate lists may be contaminated at low flux levels by background galaxies, and so we differentiate between sources with a high (“probable”) and moderate (“possible”) likelihood of being a YSO. There are 2493/425 probable YSO candidates in the LMC/SMC. Approximately 73% of the Herschel YSO candidates are newly identified in the LMC, and 35% in the SMC. We further identify a small population of dusty objects in the late stages of stellar evolution including extreme and post-asymptotic giant branch, planetary nebulae, and supernova remnants. These populations are identified by matching the HERITAGE catalogs to lists of previously identified objects in the literature. Approximately half of the LMC sources and one quarter of the SMC sources are too faint to obtain accurate ample FIR photometry and are unclassified.« less
Efficient automatic OCR word validation using word partial format derivation and language model
NASA Astrophysics Data System (ADS)
Chen, Siyuan; Misra, Dharitri; Thoma, George R.
2010-01-01
In this paper we present an OCR validation module, implemented for the System for Preservation of Electronic Resources (SPER) developed at the U.S. National Library of Medicine.1 The module detects and corrects suspicious words in the OCR output of scanned textual documents through a procedure of deriving partial formats for each suspicious word, retrieving candidate words by partial-match search from lexicons, and comparing the joint probabilities of N-gram and OCR edit transformation corresponding to the candidates. The partial format derivation, based on OCR error analysis, efficiently and accurately generates candidate words from lexicons represented by ternary search trees. In our test case comprising a historic medico-legal document collection, this OCR validation module yielded the correct words with 87% accuracy and reduced the overall OCR word errors by around 60%.
When Can Species Abundance Data Reveal Non-neutrality?
Al Hammal, Omar; Alonso, David; Etienne, Rampal S.; Cornell, Stephen J.
2015-01-01
Species abundance distributions (SAD) are probably ecology’s most well-known empirical pattern, and over the last decades many models have been proposed to explain their shape. There is no consensus over which model is correct, because the degree to which different processes can be discerned from SAD patterns has not yet been rigorously quantified. We present a power calculation to quantify our ability to detect deviations from neutrality using species abundance data. We study non-neutral stochastic community models, and show that the presence of non-neutral processes is detectable if sample size is large enough and/or the amplitude of the effect is strong enough. Our framework can be used for any candidate community model that can be simulated on a computer, and determines both the sampling effort required to distinguish between alternative processes, and a range for the strength of non-neutral processes in communities whose patterns are statistically consistent with neutral theory. We find that even data sets of the scale of the 50 Ha forest plot on Barro Colorado Island, Panama, are unlikely to be large enough to detect deviations from neutrality caused by competitive interactions alone, though the presence of multiple non-neutral processes with contrasting effects on abundance distributions may be detectable. PMID:25793889
Khan, Hafiz; Saxena, Anshul; Perisetti, Abhilash; Rafiq, Aamrin; Gabbidon, Kemesha; Mende, Sarah; Lyuksyutova, Maria; Quesada, Kandi; Blakely, Summre; Torres, Tiffany; Afesse, Mahlet
2016-12-01
Background: Breast cancer is a worldwide public health concern and is the most prevalent type of cancer in women in the United States. This study concerned the best fit of statistical probability models on the basis of survival times for nine state cancer registries: California, Connecticut, Georgia, Hawaii, Iowa, Michigan, New Mexico, Utah, and Washington. Materials and Methods: A probability random sampling method was applied to select and extract records of 2,000 breast cancer patients from the Surveillance Epidemiology and End Results (SEER) database for each of the nine state cancer registries used in this study. EasyFit software was utilized to identify the best probability models by using goodness of fit tests, and to estimate parameters for various statistical probability distributions that fit survival data. Results: Statistical analysis for the summary of statistics is reported for each of the states for the years 1973 to 2012. Kolmogorov-Smirnov, Anderson-Darling, and Chi-squared goodness of fit test values were used for survival data, the highest values of goodness of fit statistics being considered indicative of the best fit survival model for each state. Conclusions: It was found that California, Connecticut, Georgia, Iowa, New Mexico, and Washington followed the Burr probability distribution, while the Dagum probability distribution gave the best fit for Michigan and Utah, and Hawaii followed the Gamma probability distribution. These findings highlight differences between states through selected sociodemographic variables and also demonstrate probability modeling differences in breast cancer survival times. The results of this study can be used to guide healthcare providers and researchers for further investigations into social and environmental factors in order to reduce the occurrence of and mortality due to breast cancer. Creative Commons Attribution License
Bouchard, Kristofer E.; Ganguli, Surya; Brainard, Michael S.
2015-01-01
The majority of distinct sensory and motor events occur as temporally ordered sequences with rich probabilistic structure. Sequences can be characterized by the probability of transitioning from the current state to upcoming states (forward probability), as well as the probability of having transitioned to the current state from previous states (backward probability). Despite the prevalence of probabilistic sequencing of both sensory and motor events, the Hebbian mechanisms that mold synapses to reflect the statistics of experienced probabilistic sequences are not well understood. Here, we show through analytic calculations and numerical simulations that Hebbian plasticity (correlation, covariance, and STDP) with pre-synaptic competition can develop synaptic weights equal to the conditional forward transition probabilities present in the input sequence. In contrast, post-synaptic competition can develop synaptic weights proportional to the conditional backward probabilities of the same input sequence. We demonstrate that to stably reflect the conditional probability of a neuron's inputs and outputs, local Hebbian plasticity requires balance between competitive learning forces that promote synaptic differentiation and homogenizing learning forces that promote synaptic stabilization. The balance between these forces dictates a prior over the distribution of learned synaptic weights, strongly influencing both the rate at which structure emerges and the entropy of the final distribution of synaptic weights. Together, these results demonstrate a simple correspondence between the biophysical organization of neurons, the site of synaptic competition, and the temporal flow of information encoded in synaptic weights by Hebbian plasticity while highlighting the utility of balancing learning forces to accurately encode probability distributions, and prior expectations over such probability distributions. PMID:26257637
Lithium in ROSAT-discovered candidate members in the Alpha Persei cluster
NASA Astrophysics Data System (ADS)
Randich, S.; Martin, E. L.; Garcia Lopez, R. J.; Pallavicini, R.
1998-05-01
We present lithium observations of 23 X-ray selected candidate members of alpha Per, which are part of a larger sample of stars identified through two ROSAT surveys of the cluster. Our observations on one hand allowed us to confirm membership for 18 of the candidates, thus suggesting that a high percentage of the whole X-ray selected candidates are probably cluster members. On the other hand, we had the possibility to significantly enlarge the Li database for this cluster. The distribution of Li abundances for stars in our sample (or `new' members) is in good agreement with that for previously known (or `old') members, although `new' members in the 5000 - 4700 T_eff interval stay on the upper envelope of the Li vs. T_eff diagram. The comparison of the merged `new' + `old' sample with the younger IC 2602 and IC 4665 clusters and with the older Pleiades confirms that stars more massive than the Sun do not undergo any PMS Li destruction, whereas some depletion occurs during the early phases on the ZAMS. We re-addressed the issue of the star-to-star scatter and Li-rotation connection for both alpha Per and the Pleiades; as several previous studies have pointed out, fast rotators, as a group, show higher lithium than slow rotators. At the same time, however, fast rotators exhibit a much narrower dispersion than slow rotators. We demonstrate that this dicothomy is unlikely due to projection effects and suggest that the reason for it could reside in the PMS rotational history and, in particular, in the presence (absence) of a circumstellar disk. As to very cool stars (T_eff < 4500 K), we find that alpha Per members do not seem to have higher lithium than the Pleiades. This result, however, must be confirmed with a larger sample of alpha Per stars before any conclusion can be drawn. Based on observations made with the Isaac Newton telescope, operated on the island of La Palma by the Isaac Newton Group in the Spanish Observatorio del Roque de los Muchachos of the Instituto de Astrof\\'\\i sica de Canarias.
NASA Astrophysics Data System (ADS)
Christen, Alejandra; Escarate, Pedro; Curé, Michel; Rial, Diego F.; Cassetti, Julia
2016-10-01
Aims: Knowing the distribution of stellar rotational velocities is essential for understanding stellar evolution. Because we measure the projected rotational speed v sin I, we need to solve an ill-posed problem given by a Fredholm integral of the first kind to recover the "true" rotational velocity distribution. Methods: After discretization of the Fredholm integral we apply the Tikhonov regularization method to obtain directly the probability distribution function for stellar rotational velocities. We propose a simple and straightforward procedure to determine the Tikhonov parameter. We applied Monte Carlo simulations to prove that the Tikhonov method is a consistent estimator and asymptotically unbiased. Results: This method is applied to a sample of cluster stars. We obtain confidence intervals using a bootstrap method. Our results are in close agreement with those obtained using the Lucy method for recovering the probability density distribution of rotational velocities. Furthermore, Lucy estimation lies inside our confidence interval. Conclusions: Tikhonov regularization is a highly robust method that deconvolves the rotational velocity probability density function from a sample of v sin I data directly without the need for any convergence criteria.
NASA Astrophysics Data System (ADS)
Ksoll, Victor F.; Gouliermis, Dimitrios A.; Klessen, Ralf S.; Grebel, Eva K.; Sabbi, Elena; Anderson, Jay; Lennon, Daniel J.; Cignoni, Michele; de Marchi, Guido; Smith, Linda J.; Tosi, Monica; van der Marel, Roeland P.
2018-05-01
The Hubble Tarantula Treasury Project (HTTP) has provided an unprecedented photometric coverage of the entire star-burst region of 30 Doradus down to the half Solar mass limit. We use the deep stellar catalogue of HTTP to identify all the pre-main-sequence (PMS) stars of the region, i.e., stars that have not started their lives on the main-sequence yet. The photometric distinction of these stars from the more evolved populations is not a trivial task due to several factors that alter their colour-magnitude diagram positions. The identification of PMS stars requires, thus, sophisticated statistical methods. We employ Machine Learning Classification techniques on the HTTP survey of more than 800,000 sources to identify the PMS stellar content of the observed field. Our methodology consists of 1) carefully selecting the most probable low-mass PMS stellar population of the star-forming cluster NGC2070, 2) using this sample to train classification algorithms to build a predictive model for PMS stars, and 3) applying this model in order to identify the most probable PMS content across the entire Tarantula Nebula. We employ Decision Tree, Random Forest and Support Vector Machine classifiers to categorise the stars as PMS and Non-PMS. The Random Forest and Support Vector Machine provided the most accurate models, predicting about 20,000 sources with a candidateship probability higher than 50 percent, and almost 10,000 PMS candidates with a probability higher than 95 percent. This is the richest and most accurate photometric catalogue of extragalactic PMS candidates across the extent of a whole star-forming complex.
Quantitative assessment of building fire risk to life safety.
Guanquan, Chu; Jinhua, Sun
2008-06-01
This article presents a quantitative risk assessment framework for evaluating fire risk to life safety. Fire risk is divided into two parts: probability and corresponding consequence of every fire scenario. The time-dependent event tree technique is used to analyze probable fire scenarios based on the effect of fire protection systems on fire spread and smoke movement. To obtain the variation of occurrence probability with time, Markov chain is combined with a time-dependent event tree for stochastic analysis on the occurrence probability of fire scenarios. To obtain consequences of every fire scenario, some uncertainties are considered in the risk analysis process. When calculating the onset time to untenable conditions, a range of fires are designed based on different fire growth rates, after which uncertainty of onset time to untenable conditions can be characterized by probability distribution. When calculating occupant evacuation time, occupant premovement time is considered as a probability distribution. Consequences of a fire scenario can be evaluated according to probability distribution of evacuation time and onset time of untenable conditions. Then, fire risk to life safety can be evaluated based on occurrence probability and consequences of every fire scenario. To express the risk assessment method in detail, a commercial building is presented as a case study. A discussion compares the assessment result of the case study with fire statistics.
NASA Astrophysics Data System (ADS)
Iskandar, I.
2018-03-01
The exponential distribution is the most widely used reliability analysis. This distribution is very suitable for representing the lengths of life of many cases and is available in a simple statistical form. The characteristic of this distribution is a constant hazard rate. The exponential distribution is the lower rank of the Weibull distributions. In this paper our effort is to introduce the basic notions that constitute an exponential competing risks model in reliability analysis using Bayesian analysis approach and presenting their analytic methods. The cases are limited to the models with independent causes of failure. A non-informative prior distribution is used in our analysis. This model describes the likelihood function and follows with the description of the posterior function and the estimations of the point, interval, hazard function, and reliability. The net probability of failure if only one specific risk is present, crude probability of failure due to a specific risk in the presence of other causes, and partial crude probabilities are also included.
On probability-possibility transformations
NASA Technical Reports Server (NTRS)
Klir, George J.; Parviz, Behzad
1992-01-01
Several probability-possibility transformations are compared in terms of the closeness of preserving second-order properties. The comparison is based on experimental results obtained by computer simulation. Two second-order properties are involved in this study: noninteraction of two distributions and projections of a joint distribution.
Theoretical size distribution of fossil taxa: analysis of a null model.
Reed, William J; Hughes, Barry D
2007-03-22
This article deals with the theoretical size distribution (of number of sub-taxa) of a fossil taxon arising from a simple null model of macroevolution. New species arise through speciations occurring independently and at random at a fixed probability rate, while extinctions either occur independently and at random (background extinctions) or cataclysmically. In addition new genera are assumed to arise through speciations of a very radical nature, again assumed to occur independently and at random at a fixed probability rate. The size distributions of the pioneering genus (following a cataclysm) and of derived genera are determined. Also the distribution of the number of genera is considered along with a comparison of the probability of a monospecific genus with that of a monogeneric family.
Newton/Poisson-Distribution Program
NASA Technical Reports Server (NTRS)
Bowerman, Paul N.; Scheuer, Ernest M.
1990-01-01
NEWTPOIS, one of two computer programs making calculations involving cumulative Poisson distributions. NEWTPOIS (NPO-17715) and CUMPOIS (NPO-17714) used independently of one another. NEWTPOIS determines Poisson parameter for given cumulative probability, from which one obtains percentiles for gamma distributions with integer shape parameters and percentiles for X(sup2) distributions with even degrees of freedom. Used by statisticians and others concerned with probabilities of independent events occurring over specific units of time, area, or volume. Program written in C.
PLANET HUNTERS: NEW KEPLER PLANET CANDIDATES FROM ANALYSIS OF QUARTER 2
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lintott, Chris J.; Schwamb, Megan E.; Schwainski, Kevin, E-mail: cjl@astro.ox.ac.uk
2013-06-15
We present new planet candidates identified in NASA Kepler Quarter 2 public release data by volunteers engaged in the Planet Hunters citizen science project. The two candidates presented here survive checks for false positives, including examination of the pixel offset to constrain the possibility of a background eclipsing binary. The orbital periods of the planet candidates are 97.46 days (KIC 4552729) and 284.03 (KIC 10005758) days and the modeled planet radii are 5.3 and 3.8 R{sub Circled-Plus }. The latter star has an additional known planet candidate with a radius of 5.05 R{sub Circled-Plus} and a period of 134.49 days,more » which was detected by the Kepler pipeline. The discovery of these candidates illustrates the value of massively distributed volunteer review of the Kepler database to recover candidates which were otherwise uncataloged.« less
Probability theory for 3-layer remote sensing radiative transfer model: univariate case.
Ben-David, Avishai; Davidson, Charles E
2012-04-23
A probability model for a 3-layer radiative transfer model (foreground layer, cloud layer, background layer, and an external source at the end of line of sight) has been developed. The 3-layer model is fundamentally important as the primary physical model in passive infrared remote sensing. The probability model is described by the Johnson family of distributions that are used as a fit for theoretically computed moments of the radiative transfer model. From the Johnson family we use the SU distribution that can address a wide range of skewness and kurtosis values (in addition to addressing the first two moments, mean and variance). In the limit, SU can also describe lognormal and normal distributions. With the probability model one can evaluate the potential for detecting a target (vapor cloud layer), the probability of observing thermal contrast, and evaluate performance (receiver operating characteristics curves) in clutter-noise limited scenarios. This is (to our knowledge) the first probability model for the 3-layer remote sensing geometry that treats all parameters as random variables and includes higher-order statistics. © 2012 Optical Society of America
Sampling probability distributions of lesions in mammograms
NASA Astrophysics Data System (ADS)
Looney, P.; Warren, L. M.; Dance, D. R.; Young, K. C.
2015-03-01
One approach to image perception studies in mammography using virtual clinical trials involves the insertion of simulated lesions into normal mammograms. To facilitate this, a method has been developed that allows for sampling of lesion positions across the cranio-caudal and medio-lateral radiographic projections in accordance with measured distributions of real lesion locations. 6825 mammograms from our mammography image database were segmented to find the breast outline. The outlines were averaged and smoothed to produce an average outline for each laterality and radiographic projection. Lesions in 3304 mammograms with malignant findings were mapped on to a standardised breast image corresponding to the average breast outline using piecewise affine transforms. A four dimensional probability distribution function was found from the lesion locations in the cranio-caudal and medio-lateral radiographic projections for calcification and noncalcification lesions. Lesion locations sampled from this probability distribution function were mapped on to individual mammograms using a piecewise affine transform which transforms the average outline to the outline of the breast in the mammogram. The four dimensional probability distribution function was validated by comparing it to the two dimensional distributions found by considering each radiographic projection and laterality independently. The correlation of the location of the lesions sampled from the four dimensional probability distribution function across radiographic projections was shown to match the correlation of the locations of the original mapped lesion locations. The current system has been implemented as a web-service on a server using the Python Django framework. The server performs the sampling, performs the mapping and returns the results in a javascript object notation format.
The Gaia-ESO Survey: double-, triple-, and quadruple-line spectroscopic binary candidates
NASA Astrophysics Data System (ADS)
Merle, T.; Van Eck, S.; Jorissen, A.; Van der Swaelmen, M.; Masseron, T.; Zwitter, T.; Hatzidimitriou, D.; Klutsch, A.; Pourbaix, D.; Blomme, R.; Worley, C. C.; Sacco, G.; Lewis, J.; Abia, C.; Traven, G.; Sordo, R.; Bragaglia, A.; Smiljanic, R.; Pancino, E.; Damiani, F.; Hourihane, A.; Gilmore, G.; Randich, S.; Koposov, S.; Casey, A.; Morbidelli, L.; Franciosini, E.; Magrini, L.; Jofre, P.; Costado, M. T.; Jeffries, R. D.; Bergemann, M.; Lanzafame, A. C.; Bayo, A.; Carraro, G.; Flaccomio, E.; Monaco, L.; Zaggia, S.
2017-12-01
Context. The Gaia-ESO Survey (GES) is a large spectroscopic survey that provides a unique opportunity to study the distribution of spectroscopic multiple systems among different populations of the Galaxy. Aims: Our aim is to detect binarity/multiplicity for stars targeted by the GES from the analysis of the cross-correlation functions (CCFs) of the GES spectra with spectral templates. Methods: We developed a method based on the computation of the CCF successive derivatives to detect multiple peaks and determine their radial velocities, even when the peaks are strongly blended. The parameters of the detection of extrema (DOE) code have been optimized for each GES GIRAFFE and UVES setup to maximize detection. The DOE code therefore allows to automatically detect multiple line spectroscopic binaries (SBn, n ≥ 2). Results: We apply this method on the fourth GES internal data release and detect 354 SBn candidates (342 SB2, 11 SB3, and even one SB4), including only nine SBs known in the literature. This implies that about 98% of these SBn candidates are new because of their faint visual magnitude that can reach V = 19. Visual inspection of the SBn candidate spectra reveals that the most probable candidates have indeed a composite spectrum. Among the SB2 candidates, an orbital solution could be computed for two previously unknown binaries: CNAME 06404608+0949173 (known as V642 Mon) in NGC 2264 and CNAME 19013257-0027338 in Berkeley 81 (Be 81). A detailed analysis of the unique SB4 (four peaks in the CCF) reveals that CNAME 08414659-5303449 (HD 74438) in the open cluster IC 2391 is a physically bound stellar quadruple system. The SB candidates belonging to stellar clusters are reviewed in detail to discard false detections. We suggest that atmospheric parameters should not be used for these system components; SB-specific pipelines should be used instead. Conclusions: Our implementation of an automatic detection of spectroscopic binaries within the GES has allowed the efficient discovery of many new multiple systems. With the detection of the SB1 candidates that will be the subject of a forthcoming paper, the study of the statistical and physical properties of the spectroscopic multiple systems will soon be possible for the entire GES sample. Based on data products from observations made with ESO Telescopes at the La Silla Paranal Observatory under programme ID 188.B-3002. These data products have been processed by the Cambridge Astronomy Survey Unit (CASU) at the Institute of Astronomy, University of Cambridge, and by the FLAMES/UVES reduction team at INAF/Osservatorio Astrofisico di Arcetri. These data have been obtained from the Gaia-ESO Survey Data Archive, prepared and hosted by the Wide Field Astronomy Unit, Institute for Astronomy, University of Edinburgh, which is funded by the UK Science and Technology Facilities Council.
First depressed, then discriminated against?
Baert, Stijn; De Visschere, Sarah; Schoors, Koen; Vandenberghe, Désirée; Omey, Eddy
2016-12-01
Each year a substantial share of the European population suffers from major depression. This mental illness may affect individuals' later life outcomes indirectly by the stigma it inflicts. The present study assesses hiring discrimination based on disclosed depression. To this end, between May 2015 and July 2015, we sent out 288 trios of job applications from fictitious candidates to real vacancies in Belgium. Within each trio, one candidate claimed to have become unemployed only recently, whereas the other two candidates revealed former depression or no reason at all for their unemployment during a full year. Disclosing a year of inactivity due to former depression decreases the probability of getting a job interview invitation by about 34% when compared with candidates who just became unemployed, but the stigma effect of a year of depression is not significantly higher than the stigma effect of a year of unexplained unemployment. In addition, we found that these stigmas of depression and unemployment were driven by our male trios of fictitious candidates. As a consequence, our results are in favour of further research on gender heterogeneity in the stigma of depression and other health impairments. Copyright © 2016 Elsevier Ltd. All rights reserved.
Prediction of Human Disease Genes by Human-Mouse Conserved Coexpression Analysis
Grassi, Elena; Damasco, Christian; Silengo, Lorenzo; Oti, Martin; Provero, Paolo; Di Cunto, Ferdinando
2008-01-01
Background Even in the post-genomic era, the identification of candidate genes within loci associated with human genetic diseases is a very demanding task, because the critical region may typically contain hundreds of positional candidates. Since genes implicated in similar phenotypes tend to share very similar expression profiles, high throughput gene expression data may represent a very important resource to identify the best candidates for sequencing. However, so far, gene coexpression has not been used very successfully to prioritize positional candidates. Methodology/Principal Findings We show that it is possible to reliably identify disease-relevant relationships among genes from massive microarray datasets by concentrating only on genes sharing similar expression profiles in both human and mouse. Moreover, we show systematically that the integration of human-mouse conserved coexpression with a phenotype similarity map allows the efficient identification of disease genes in large genomic regions. Finally, using this approach on 850 OMIM loci characterized by an unknown molecular basis, we propose high-probability candidates for 81 genetic diseases. Conclusion Our results demonstrate that conserved coexpression, even at the human-mouse phylogenetic distance, represents a very strong criterion to predict disease-relevant relationships among human genes. PMID:18369433
30+ New & Known SB2s in the SDSS-III/APOGEE M Dwarf Ancillary Science Project Sample
NASA Astrophysics Data System (ADS)
Skinner, Jacob; Covey, Kevin; Bender, Chad; De Lee, Nathan Michael; Chojnowski, Drew; Troup, Nicholas; Badenes, Carles; Mahadevan, Suvrath; Terrien, Ryan
2018-01-01
Close stellar binaries can drive dynamical interactions that affect the structure and evolution of planetary systems. Binary surveys indicate that the multiplicity fraction and typical orbital separation decrease with primary mass, but correlations with higher order architectural parameters such as the system's mass ratio are less well constrained. We seek to identify and characterize double-lined spectroscopic binaries (SB2s) among the 1350 M dwarf ancillary science targets with APOGEE spectra in the SDSS-III Data Release 13. We quantitatively measure the degree of asymmetry in the APOGEE pipeline cross-correlation functions (CCFs), and use those metrics to identify a sample of 44 high-likelihood candidate SB2s. Extracting radial velocities (RVs) for both binary components from the CCF, we then measure mass ratios for 31 SB2s; we also use Bayesian techniques to fit orbits for 4 systems with 8 or more distinct APOGEE observations. The (incomplete) mass ratio distribution of this sample rises quickly towards unity. Two-sided Kolmogorov-Smirnov (K-S) tests find probabilities of 13.8% and 14.2% that the M dwarf mass ratio distribution is consistent with those measured by Pourbaix et al. (2004) and Fernandez et al. (2017), respectively. The samples analyzed by Pourbaix et al. and Fernandez et al. are dominated by higher-mass solar type stars; this suggests that the mass ratio distribution of close binaries is not strongly dependent on primary mass.
How to model a negligible probability under the WTO sanitary and phytosanitary agreement?
Powell, Mark R
2013-06-01
Since the 1997 EC--Hormones decision, World Trade Organization (WTO) Dispute Settlement Panels have wrestled with the question of what constitutes a negligible risk under the Sanitary and Phytosanitary Agreement. More recently, the 2010 WTO Australia--Apples Panel focused considerable attention on the appropriate quantitative model for a negligible probability in a risk assessment. The 2006 Australian Import Risk Analysis for Apples from New Zealand translated narrative probability statements into quantitative ranges. The uncertainty about a "negligible" probability was characterized as a uniform distribution with a minimum value of zero and a maximum value of 10(-6) . The Australia - Apples Panel found that the use of this distribution would tend to overestimate the likelihood of "negligible" events and indicated that a triangular distribution with a most probable value of zero and a maximum value of 10⁻⁶ would correct the bias. The Panel observed that the midpoint of the uniform distribution is 5 × 10⁻⁷ but did not consider that the triangular distribution has an expected value of 3.3 × 10⁻⁷. Therefore, if this triangular distribution is the appropriate correction, the magnitude of the bias found by the Panel appears modest. The Panel's detailed critique of the Australian risk assessment, and the conclusions of the WTO Appellate Body about the materiality of flaws found by the Panel, may have important implications for the standard of review for risk assessments under the WTO SPS Agreement. © 2012 Society for Risk Analysis.
Fourier Method for Calculating Fission Chain Neutron Multiplicity Distributions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chambers, David H.; Chandrasekaran, Hema; Walston, Sean E.
Here, a new way of utilizing the fast Fourier transform is developed to compute the probability distribution for a fission chain to create n neutrons. We then extend this technique to compute the probability distributions for detecting n neutrons. Lastly, our technique can be used for fission chains initiated by either a single neutron inducing a fission or by the spontaneous fission of another isotope.
Fourier Method for Calculating Fission Chain Neutron Multiplicity Distributions
Chambers, David H.; Chandrasekaran, Hema; Walston, Sean E.
2017-03-27
Here, a new way of utilizing the fast Fourier transform is developed to compute the probability distribution for a fission chain to create n neutrons. We then extend this technique to compute the probability distributions for detecting n neutrons. Lastly, our technique can be used for fission chains initiated by either a single neutron inducing a fission or by the spontaneous fission of another isotope.
ERIC Educational Resources Information Center
Duerdoth, Ian
2009-01-01
The subject of uncertainties (sometimes called errors) is traditionally taught (to first-year science undergraduates) towards the end of a course on statistics that defines probability as the limit of many trials, and discusses probability distribution functions and the Gaussian distribution. We show how to introduce students to the concepts of…
Robust approaches to quantification of margin and uncertainty for sparse data
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hund, Lauren; Schroeder, Benjamin B.; Rumsey, Kelin
Characterizing the tails of probability distributions plays a key role in quantification of margins and uncertainties (QMU), where the goal is characterization of low probability, high consequence events based on continuous measures of performance. When data are collected using physical experimentation, probability distributions are typically fit using statistical methods based on the collected data, and these parametric distributional assumptions are often used to extrapolate about the extreme tail behavior of the underlying probability distribution. In this project, we character- ize the risk associated with such tail extrapolation. Specifically, we conducted a scaling study to demonstrate the large magnitude of themore » risk; then, we developed new methods for communicat- ing risk associated with tail extrapolation from unvalidated statistical models; lastly, we proposed a Bayesian data-integration framework to mitigate tail extrapolation risk through integrating ad- ditional information. We conclude that decision-making using QMU is a complex process that cannot be achieved using statistical analyses alone.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, Hang, E-mail: hangchen@mit.edu; Thill, Peter; Cao, Jianshu
In biochemical systems, intrinsic noise may drive the system switch from one stable state to another. We investigate how kinetic switching between stable states in a bistable network is influenced by dynamic disorder, i.e., fluctuations in the rate coefficients. Using the geometric minimum action method, we first investigate the optimal transition paths and the corresponding minimum actions based on a genetic toggle switch model in which reaction coefficients draw from a discrete probability distribution. For the continuous probability distribution of the rate coefficient, we then consider two models of dynamic disorder in which reaction coefficients undergo different stochastic processes withmore » the same stationary distribution. In one, the kinetic parameters follow a discrete Markov process and in the other they follow continuous Langevin dynamics. We find that regulation of the parameters modulating the dynamic disorder, as has been demonstrated to occur through allosteric control in bistable networks in the immune system, can be crucial in shaping the statistics of optimal transition paths, transition probabilities, and the stationary probability distribution of the network.« less
NASA Astrophysics Data System (ADS)
Jenkins, Colleen; Jordan, Jay; Carlson, Jeff
2007-02-01
This paper presents parameter estimation techniques useful for detecting background changes in a video sequence with extreme foreground activity. A specific application of interest is automated detection of the covert placement of threats (e.g., a briefcase bomb) inside crowded public facilities. We propose that a histogram of pixel intensity acquired from a fixed mounted camera over time for a series of images will be a mixture of two Gaussian functions: the foreground probability distribution function and background probability distribution function. We will use Pearson's Method of Moments to separate the two probability distribution functions. The background function can then be "remembered" and changes in the background can be detected. Subsequent comparisons of background estimates are used to detect changes. Changes are flagged to alert security forces to the presence and location of potential threats. Results are presented that indicate the significant potential for robust parameter estimation techniques as applied to video surveillance.
Effects of habitat fragmentation on passerine birds breeding in Intermountain shrubsteppe
Knick, S.T.; Rotenberry, J.T.
2002-01-01
Habitat fragmentation and loss strongly influence the distribution and abundance of passerine birds breeding in Intermountain shrubsteppe. Wildfires, human activities, and change in vegetation communities often are synergistic in these systems and can result in radical conversion from shrubland to grasslands dominated by exotic annuals at large temporal and spatial scales from which recovery to native conditions is unlikely. As a result, populations of 5 of the 12 species in our review of Intermountain shrubsteppe birds are undergoing significant declines; 5 species are listed as at-risk or as candidates for protection in at least one state. The process by which fragmentation affects bird distributions in these habitats remains unknown because most research has emphasized the detection of population trends and patterns of habitat associations at relatively large spatial scales. Our research indicates that the distribution of shrubland-obligate species, such as Brewer's Sparrows (Spizella breweri), Sage Sparrows (Amphispiza belli), and Sage Thrashers (Oreoscoptes montanus), was highly sensitive to fragmentation of shrublands at spatial scales larger than individual home ranges. In contrast, the underlying mechanisms for both habitat change and bird population dynamics may operate independently of habitat boundaries. We propose alternative, but not necessarily exclusive, mechanisms to explain the relationship between habitat fragmentation and bird distribution and abundance. Fragmentation might influence productivity through differences in breeding density, nesting success, or predation. However, local and landscape variables were not significant determinants either of success, number fledged, or probability of predation or parasitism (although our tests had relatively low statistical power). Alternatively, relative absence of natal philopatry and redistribution by individuals among habitats following fledging or post-migration could account for the pattern of distribution and abundance. Thus, boundary dynamics may be important in determining the distribution of shrubland-obligate species but insignificant relative to the mechanisms causing the pattern of habitat and bird distribution. Because of the dichotomy in responses, Intermountain shrubsteppe systems present a unique challenge in understanding how landscape composition, configuration, and change influence bird population dynamics.
A Method for the Evaluation of Thousands of Automated 3D Stem Cell Segmentations
Bajcsy, Peter; Simon, Mylene; Florczyk, Stephen; Simon, Carl G.; Juba, Derek; Brady, Mary
2016-01-01
There is no segmentation method that performs perfectly with any data set in comparison to human segmentation. Evaluation procedures for segmentation algorithms become critical for their selection. The problems associated with segmentation performance evaluations and visual verification of segmentation results are exaggerated when dealing with thousands of 3D image volumes because of the amount of computation and manual inputs needed. We address the problem of evaluating 3D segmentation performance when segmentation is applied to thousands of confocal microscopy images (z-stacks). Our approach is to incorporate experimental imaging and geometrical criteria, and map them into computationally efficient segmentation algorithms that can be applied to a very large number of z-stacks. This is an alternative approach to considering existing segmentation methods and evaluating most state-of-the-art algorithms. We designed a methodology for 3D segmentation performance characterization that consists of design, evaluation and verification steps. The characterization integrates manual inputs from projected surrogate “ground truth” of statistically representative samples and from visual inspection into the evaluation. The novelty of the methodology lies in (1) designing candidate segmentation algorithms by mapping imaging and geometrical criteria into algorithmic steps, and constructing plausible segmentation algorithms with respect to the order of algorithmic steps and their parameters, (2) evaluating segmentation accuracy using samples drawn from probability distribution estimates of candidate segmentations, and (3) minimizing human labor needed to create surrogate “truth” by approximating z-stack segmentations with 2D contours from three orthogonal z-stack projections and by developing visual verification tools. We demonstrate the methodology by applying it to a dataset of 1253 mesenchymal stem cells. The cells reside on 10 different types of biomaterial scaffolds, and are stained for actin and nucleus yielding 128 460 image frames (on average 125 cells/scaffold × 10 scaffold types × 2 stains × 51 frames/cell). After constructing and evaluating six candidates of 3D segmentation algorithms, the most accurate 3D segmentation algorithm achieved an average precision of 0.82 and an accuracy of 0.84 as measured by the Dice similarity index where values greater than 0.7 indicate a good spatial overlap. A probability of segmentation success was 0.85 based on visual verification, and a computation time was 42.3 h to process all z-stacks. While the most accurate segmentation technique was 4.2 times slower than the second most accurate algorithm, it consumed on average 9.65 times less memory per z-stack segmentation. PMID:26268699
p-adic stochastic hidden variable model
NASA Astrophysics Data System (ADS)
Khrennikov, Andrew
1998-03-01
We propose stochastic hidden variables model in which hidden variables have a p-adic probability distribution ρ(λ) and at the same time conditional probabilistic distributions P(U,λ), U=A,A',B,B', are ordinary probabilities defined on the basis of the Kolmogorov measure-theoretical axiomatics. A frequency definition of p-adic probability is quite similar to the ordinary frequency definition of probability. p-adic frequency probability is defined as the limit of relative frequencies νn but in the p-adic metric. We study a model with p-adic stochastics on the level of the hidden variables description. But, of course, responses of macroapparatuses have to be described by ordinary stochastics. Thus our model describes a mixture of p-adic stochastics of the microworld and ordinary stochastics of macroapparatuses. In this model probabilities for physical observables are the ordinary probabilities. At the same time Bell's inequality is violated.
Stationary properties of maximum-entropy random walks.
Dixit, Purushottam D
2015-10-01
Maximum-entropy (ME) inference of state probabilities using state-dependent constraints is popular in the study of complex systems. In stochastic systems, how state space topology and path-dependent constraints affect ME-inferred state probabilities remains unknown. To that end, we derive the transition probabilities and the stationary distribution of a maximum path entropy Markov process subject to state- and path-dependent constraints. A main finding is that the stationary distribution over states differs significantly from the Boltzmann distribution and reflects a competition between path multiplicity and imposed constraints. We illustrate our results with particle diffusion on a two-dimensional landscape. Connections with the path integral approach to diffusion are discussed.
A New M Dwarf Debris Disk Candidate in a Young Moving Group Discovered with Disk Detective
NASA Technical Reports Server (NTRS)
Silverberg, Steven M.; Kuchner, Marc J.; Wisniewski, John P.; Gagne, Jonathan; Bans, Alissa S.; Bhattacharjee, Shambo; Currie, Thayne R.; Debes, John R.; Biggs, Joseph R; Bosch, Milton
2016-01-01
We used the Disk Detective citizen science project and the BANYAN II Bayesian analysis tool to identify a new candidate member of a nearby young association with infrared excess. WISE J080822.18-644357.3, an M5.5-type debris disk system with significant excess at both 12 and 22 microns, is a likely member (approx.90% BANYAN II probability) of the approx.45 Myr old Carina association. Since this would be the oldest M dwarf debris disk detected in a moving group, this discovery could be an important constraint on our understanding of M dwarf debris disk evolution.
NASA Astrophysics Data System (ADS)
Mandal, S.; Choudhury, B. U.
2015-07-01
Sagar Island, setting on the continental shelf of Bay of Bengal, is one of the most vulnerable deltas to the occurrence of extreme rainfall-driven climatic hazards. Information on probability of occurrence of maximum daily rainfall will be useful in devising risk management for sustaining rainfed agrarian economy vis-a-vis food and livelihood security. Using six probability distribution models and long-term (1982-2010) daily rainfall data, we studied the probability of occurrence of annual, seasonal and monthly maximum daily rainfall (MDR) in the island. To select the best fit distribution models for annual, seasonal and monthly time series based on maximum rank with minimum value of test statistics, three statistical goodness of fit tests, viz. Kolmogorove-Smirnov test (K-S), Anderson Darling test ( A 2 ) and Chi-Square test ( X 2) were employed. The fourth probability distribution was identified from the highest overall score obtained from the three goodness of fit tests. Results revealed that normal probability distribution was best fitted for annual, post-monsoon and summer seasons MDR, while Lognormal, Weibull and Pearson 5 were best fitted for pre-monsoon, monsoon and winter seasons, respectively. The estimated annual MDR were 50, 69, 86, 106 and 114 mm for return periods of 2, 5, 10, 20 and 25 years, respectively. The probability of getting an annual MDR of >50, >100, >150, >200 and >250 mm were estimated as 99, 85, 40, 12 and 03 % level of exceedance, respectively. The monsoon, summer and winter seasons exhibited comparatively higher probabilities (78 to 85 %) for MDR of >100 mm and moderate probabilities (37 to 46 %) for >150 mm. For different recurrence intervals, the percent probability of MDR varied widely across intra- and inter-annual periods. In the island, rainfall anomaly can pose a climatic threat to the sustainability of agricultural production and thus needs adequate adaptation and mitigation measures.
RXTE PCA Detection of a New Outburst of XTE J1728-295 (probably IGR J17285-2922)
NASA Astrophysics Data System (ADS)
Markwardt, Craig B.; Swank, Jean H.
2010-08-01
We report the detection of a new outburst of a source designated XTE J1728-295 in the RXTE PCA scans, which is probably the same as IGR J17285-2922. This source was first detected in August-October 2003 with PCA scans of the galactic center region, and is speculated to be a black hole candidate (Barlow et al. 2005, A&A, 437, L27). In PCA scans on 2010-08-28 near 09:35 UTC, the source rose to a flux of 6.5 mCrab (2-10 keV).
Probability of Detection (POD) as a statistical model for the validation of qualitative methods.
Wehling, Paul; LaBudde, Robert A; Brunelle, Sharon L; Nelson, Maria T
2011-01-01
A statistical model is presented for use in validation of qualitative methods. This model, termed Probability of Detection (POD), harmonizes the statistical concepts and parameters between quantitative and qualitative method validation. POD characterizes method response with respect to concentration as a continuous variable. The POD model provides a tool for graphical representation of response curves for qualitative methods. In addition, the model allows comparisons between candidate and reference methods, and provides calculations of repeatability, reproducibility, and laboratory effects from collaborative study data. Single laboratory study and collaborative study examples are given.
Two Universality Properties Associated with the Monkey Model of Zipf's Law
NASA Astrophysics Data System (ADS)
Perline, Richard; Perline, Ron
2016-03-01
The distribution of word probabilities in the monkey model of Zipf's law is associated with two universality properties: (1) the power law exponent converges strongly to $-1$ as the alphabet size increases and the letter probabilities are specified as the spacings from a random division of the unit interval for any distribution with a bounded density function on $[0,1]$; and (2), on a logarithmic scale the version of the model with a finite word length cutoff and unequal letter probabilities is approximately normally distributed in the part of the distribution away from the tails. The first property is proved using a remarkably general limit theorem for the logarithm of sample spacings from Shao and Hahn, and the second property follows from Anscombe's central limit theorem for a random number of i.i.d. random variables. The finite word length model leads to a hybrid Zipf-lognormal mixture distribution closely related to work in other areas.
Computer routines for probability distributions, random numbers, and related functions
Kirby, W.
1983-01-01
Use of previously coded and tested subroutines simplifies and speeds up program development and testing. This report presents routines that can be used to calculate various probability distributions and other functions of importance in statistical hydrology. The routines are designed as general-purpose Fortran subroutines and functions to be called from user-written main progress. The probability distributions provided include the beta, chi-square, gamma, Gaussian (normal), Pearson Type III (tables and approximation), and Weibull. Also provided are the distributions of the Grubbs-Beck outlier test, Kolmogorov 's and Smirnov 's D, Student 's t, noncentral t (approximate), and Snedecor F. Other mathematical functions include the Bessel function, I sub o, gamma and log-gamma functions, error functions, and exponential integral. Auxiliary services include sorting and printer-plotting. Random number generators for uniform and normal numbers are provided and may be used with some of the above routines to generate numbers from other distributions. (USGS)
Computer routines for probability distributions, random numbers, and related functions
Kirby, W.H.
1980-01-01
Use of previously codes and tested subroutines simplifies and speeds up program development and testing. This report presents routines that can be used to calculate various probability distributions and other functions of importance in statistical hydrology. The routines are designed as general-purpose Fortran subroutines and functions to be called from user-written main programs. The probability distributions provided include the beta, chisquare, gamma, Gaussian (normal), Pearson Type III (tables and approximation), and Weibull. Also provided are the distributions of the Grubbs-Beck outlier test, Kolmogorov 's and Smirnov 's D, Student 's t, noncentral t (approximate), and Snedecor F tests. Other mathematical functions include the Bessel function I (subzero), gamma and log-gamma functions, error functions and exponential integral. Auxiliary services include sorting and printer plotting. Random number generators for uniform and normal numbers are provided and may be used with some of the above routines to generate numbers from other distributions. (USGS)
Universal laws of human society's income distribution
NASA Astrophysics Data System (ADS)
Tao, Yong
2015-10-01
General equilibrium equations in economics play the same role with many-body Newtonian equations in physics. Accordingly, each solution of the general equilibrium equations can be regarded as a possible microstate of the economic system. Since Arrow's Impossibility Theorem and Rawls' principle of social fairness will provide a powerful support for the hypothesis of equal probability, then the principle of maximum entropy is available in a just and equilibrium economy so that an income distribution will occur spontaneously (with the largest probability). Remarkably, some scholars have observed such an income distribution in some democratic countries, e.g. USA. This result implies that the hypothesis of equal probability may be only suitable for some "fair" systems (economic or physical systems). From this meaning, the non-equilibrium systems may be "unfair" so that the hypothesis of equal probability is unavailable.
Polynomial chaos representation of databases on manifolds
DOE Office of Scientific and Technical Information (OSTI.GOV)
Soize, C., E-mail: christian.soize@univ-paris-est.fr; Ghanem, R., E-mail: ghanem@usc.edu
2017-04-15
Characterizing the polynomial chaos expansion (PCE) of a vector-valued random variable with probability distribution concentrated on a manifold is a relevant problem in data-driven settings. The probability distribution of such random vectors is multimodal in general, leading to potentially very slow convergence of the PCE. In this paper, we build on a recent development for estimating and sampling from probabilities concentrated on a diffusion manifold. The proposed methodology constructs a PCE of the random vector together with an associated generator that samples from the target probability distribution which is estimated from data concentrated in the neighborhood of the manifold. Themore » method is robust and remains efficient for high dimension and large datasets. The resulting polynomial chaos construction on manifolds permits the adaptation of many uncertainty quantification and statistical tools to emerging questions motivated by data-driven queries.« less
Gravitational lensing, time delay, and gamma-ray bursts
NASA Technical Reports Server (NTRS)
Mao, Shude
1992-01-01
The probability distributions of time delay in gravitational lensing by point masses and isolated galaxies (modeled as singular isothermal spheres) are studied. For point lenses (all with the same mass) the probability distribution is broad, and with a peak at delta(t) of about 50 S; for singular isothermal spheres, the probability distribution is a rapidly decreasing function with increasing time delay, with a median delta(t) equals about 1/h month, and its behavior depends sensitively on the luminosity function of galaxies. The present simplified calculation is particularly relevant to the gamma-ray bursts if they are of cosmological origin. The frequency of 'recurrent' bursts due to gravitational lensing by galaxies is probably between 0.05 and 0.4 percent. Gravitational lensing can be used as a test of the cosmological origin of gamma-ray bursts.
It's a Buyer's Market for Colleges Hiring Junior Faculty
ERIC Educational Resources Information Center
Medina, Brenda
2012-01-01
Colleges looking to hire junior faculty members are enjoying something of a buyer's market these days. With job offers scarce in many fields, small, teaching-intensive colleges and regionally oriented state universities have been able to snag recent graduates of the nation's top programs, the kinds of candidates who would have probably gone to…
Confirmation of AT 2017eaw, a Probable Supernova in NGC 6946
NASA Astrophysics Data System (ADS)
Dong, Subo; Stanek, K. Z.
2017-05-01
Supernova candidate AT 2017eaw was discovered by Patrick Wiggins at 12.8 mag on 2017-05-14 05:42:43. We confirmed AT 2017eaw with images taken by Las Cumbres Observatory (LCO)'s 1m telescope at McDonald observatory on UT 2017-05-14 10:30:31.
Dai, Huanping; Micheyl, Christophe
2015-05-01
Proportion correct (Pc) is a fundamental measure of task performance in psychophysics. The maximum Pc score that can be achieved by an optimal (maximum-likelihood) observer in a given task is of both theoretical and practical importance, because it sets an upper limit on human performance. Within the framework of signal detection theory, analytical solutions for computing the maximum Pc score have been established for several common experimental paradigms under the assumption of Gaussian additive internal noise. However, as the scope of applications of psychophysical signal detection theory expands, the need is growing for psychophysicists to compute maximum Pc scores for situations involving non-Gaussian (internal or stimulus-induced) noise. In this article, we provide a general formula for computing the maximum Pc in various psychophysical experimental paradigms for arbitrary probability distributions of sensory activity. Moreover, easy-to-use MATLAB code implementing the formula is provided. Practical applications of the formula are illustrated, and its accuracy is evaluated, for two paradigms and two types of probability distributions (uniform and Gaussian). The results demonstrate that Pc scores computed using the formula remain accurate even for continuous probability distributions, as long as the conversion from continuous probability density functions to discrete probability mass functions is supported by a sufficiently high sampling resolution. We hope that the exposition in this article, and the freely available MATLAB code, facilitates calculations of maximum performance for a wider range of experimental situations, as well as explorations of the impact of different assumptions concerning internal-noise distributions on maximum performance in psychophysical experiments.
De Backer, A; Martinez, G T; Rosenauer, A; Van Aert, S
2013-11-01
In the present paper, a statistical model-based method to count the number of atoms of monotype crystalline nanostructures from high resolution high-angle annular dark-field (HAADF) scanning transmission electron microscopy (STEM) images is discussed in detail together with a thorough study on the possibilities and inherent limitations. In order to count the number of atoms, it is assumed that the total scattered intensity scales with the number of atoms per atom column. These intensities are quantitatively determined using model-based statistical parameter estimation theory. The distribution describing the probability that intensity values are generated by atomic columns containing a specific number of atoms is inferred on the basis of the experimental scattered intensities. Finally, the number of atoms per atom column is quantified using this estimated probability distribution. The number of atom columns available in the observed STEM image, the number of components in the estimated probability distribution, the width of the components of the probability distribution, and the typical shape of a criterion to assess the number of components in the probability distribution directly affect the accuracy and precision with which the number of atoms in a particular atom column can be estimated. It is shown that single atom sensitivity is feasible taking the latter aspects into consideration. © 2013 Elsevier B.V. All rights reserved.
Theoretical size distribution of fossil taxa: analysis of a null model
Reed, William J; Hughes, Barry D
2007-01-01
Background This article deals with the theoretical size distribution (of number of sub-taxa) of a fossil taxon arising from a simple null model of macroevolution. Model New species arise through speciations occurring independently and at random at a fixed probability rate, while extinctions either occur independently and at random (background extinctions) or cataclysmically. In addition new genera are assumed to arise through speciations of a very radical nature, again assumed to occur independently and at random at a fixed probability rate. Conclusion The size distributions of the pioneering genus (following a cataclysm) and of derived genera are determined. Also the distribution of the number of genera is considered along with a comparison of the probability of a monospecific genus with that of a monogeneric family. PMID:17376249
V/V(max) test applied to SMM gamma-ray bursts
NASA Technical Reports Server (NTRS)
Matz, S. M.; Higdon, J. C.; Share, G. H.; Messina, D. C.; Iadicicco, A.
1992-01-01
We have applied the V/V(max) test to candidate gamma-ray bursts detected by the Gamma-Ray Spectrometer (GRS) aboard the SMM satellite to examine quantitatively the uniformity of the burst source population. For a sample of 132 candidate bursts identified in the GRS data by an automated search using a single uniform trigger criterion we find average V/V(max) = 0.40 +/- 0.025. This value is significantly different from 0.5, the average for a uniform distribution in space of the parent population of burst sources; however, the shape of the observed distribution of V/V(max) is unusual and our result conflicts with previous measurements. For these reasons we can currently draw no firm conclusion about the distribution of burst sources.
Maximum entropy approach to statistical inference for an ocean acoustic waveguide.
Knobles, D P; Sagers, J D; Koch, R A
2012-02-01
A conditional probability distribution suitable for estimating the statistical properties of ocean seabed parameter values inferred from acoustic measurements is derived from a maximum entropy principle. The specification of the expectation value for an error function constrains the maximization of an entropy functional. This constraint determines the sensitivity factor (β) to the error function of the resulting probability distribution, which is a canonical form that provides a conservative estimate of the uncertainty of the parameter values. From the conditional distribution, marginal distributions for individual parameters can be determined from integration over the other parameters. The approach is an alternative to obtaining the posterior probability distribution without an intermediary determination of the likelihood function followed by an application of Bayes' rule. In this paper the expectation value that specifies the constraint is determined from the values of the error function for the model solutions obtained from a sparse number of data samples. The method is applied to ocean acoustic measurements taken on the New Jersey continental shelf. The marginal probability distribution for the values of the sound speed ratio at the surface of the seabed and the source levels of a towed source are examined for different geoacoustic model representations. © 2012 Acoustical Society of America
Probability distributions of continuous measurement results for conditioned quantum evolution
NASA Astrophysics Data System (ADS)
Franquet, A.; Nazarov, Yuli V.
2017-02-01
We address the statistics of continuous weak linear measurement on a few-state quantum system that is subject to a conditioned quantum evolution. For a conditioned evolution, both the initial and final states of the system are fixed: the latter is achieved by the postselection in the end of the evolution. The statistics may drastically differ from the nonconditioned case, and the interference between initial and final states can be observed in the probability distributions of measurement outcomes as well as in the average values exceeding the conventional range of nonconditioned averages. We develop a proper formalism to compute the distributions of measurement outcomes, and evaluate and discuss the distributions in experimentally relevant setups. We demonstrate the manifestations of the interference between initial and final states in various regimes. We consider analytically simple examples of nontrivial probability distributions. We reveal peaks (or dips) at half-quantized values of the measurement outputs. We discuss in detail the case of zero overlap between initial and final states demonstrating anomalously big average outputs and sudden jump in time-integrated output. We present and discuss the numerical evaluation of the probability distribution aiming at extending the analytical results and describing a realistic experimental situation of a qubit in the regime of resonant fluorescence.
Murphy, S.; Scala, A.; Herrero, A.; Lorito, S.; Festa, G.; Trasatti, E.; Tonini, R.; Romano, F.; Molinari, I.; Nielsen, S.
2016-01-01
The 2011 Tohoku earthquake produced an unexpected large amount of shallow slip greatly contributing to the ensuing tsunami. How frequent are such events? How can they be efficiently modelled for tsunami hazard? Stochastic slip models, which can be computed rapidly, are used to explore the natural slip variability; however, they generally do not deal specifically with shallow slip features. We study the systematic depth-dependence of slip along a thrust fault with a number of 2D dynamic simulations using stochastic shear stress distributions and a geometry based on the cross section of the Tohoku fault. We obtain a probability density for the slip distribution, which varies both with depth, earthquake size and whether the rupture breaks the surface. We propose a method to modify stochastic slip distributions according to this dynamically-derived probability distribution. This method may be efficiently applied to produce large numbers of heterogeneous slip distributions for probabilistic tsunami hazard analysis. Using numerous M9 earthquake scenarios, we demonstrate that incorporating the dynamically-derived probability distribution does enhance the conditional probability of exceedance of maximum estimated tsunami wave heights along the Japanese coast. This technique for integrating dynamic features in stochastic models can be extended to any subduction zone and faulting style. PMID:27725733
Using Bayesian Networks for Candidate Generation in Consistency-based Diagnosis
NASA Technical Reports Server (NTRS)
Narasimhan, Sriram; Mengshoel, Ole
2008-01-01
Consistency-based diagnosis relies heavily on the assumption that discrepancies between model predictions and sensor observations can be detected accurately. When sources of uncertainty like sensor noise and model abstraction exist robust schemes have to be designed to make a binary decision on whether predictions are consistent with observations. This risks the occurrence of false alarms and missed alarms when an erroneous decision is made. Moreover when multiple sensors (with differing sensing properties) are available the degree of match between predictions and observations can be used to guide the search for fault candidates. In this paper we propose a novel approach to handle this problem using Bayesian networks. In the consistency- based diagnosis formulation, automatically generated Bayesian networks are used to encode a probabilistic measure of fit between predictions and observations. A Bayesian network inference algorithm is used to compute most probable fault candidates.
Investigating the dynamical history of the interstellar object 'Oumuamua
NASA Astrophysics Data System (ADS)
Dybczyński, Piotr A.; Królikowska, Małgorzata
2018-02-01
Here we try to find the origin of 1I/2017 U1 'Oumuamua, the first interstellar object recorded inside the solar system. To this aim, we searched for close encounters between 'Oumuamua and all nearby stars with known kinematic data during their past motion. We had checked over 200 thousand stars and found just a handful of candidates. If we limit our investigation to within a 60 pc sphere surrounding the Sun, then the most probable candidate for the 'Oumuamua parent stellar habitat is the star UCAC4 535-065571. However GJ 876 is also a favourable candidate. However, the origin of 'Oumuamua from a much more distant source is still an open question. Additionally, we found that the quality of the original orbit of 'Oumuamua is accurate enough for such a study and that none of the checked stars had perturbed its motion significantly. All numerical results of this research are available in the appendix.
A discussion on the origin of quantum probabilities
DOE Office of Scientific and Technical Information (OSTI.GOV)
Holik, Federico, E-mail: olentiev2@gmail.com; Departamento de Matemática - Ciclo Básico Común, Universidad de Buenos Aires - Pabellón III, Ciudad Universitaria, Buenos Aires; Sáenz, Manuel
We study the origin of quantum probabilities as arising from non-Boolean propositional-operational structures. We apply the method developed by Cox to non distributive lattices and develop an alternative formulation of non-Kolmogorovian probability measures for quantum mechanics. By generalizing the method presented in previous works, we outline a general framework for the deduction of probabilities in general propositional structures represented by lattices (including the non-distributive case). -- Highlights: •Several recent works use a derivation similar to that of R.T. Cox to obtain quantum probabilities. •We apply Cox’s method to the lattice of subspaces of the Hilbert space. •We obtain a derivationmore » of quantum probabilities which includes mixed states. •The method presented in this work is susceptible to generalization. •It includes quantum mechanics and classical mechanics as particular cases.« less
Takemura, Kazuhisa; Murakami, Hajime
2016-01-01
A probability weighting function (w(p)) is considered to be a nonlinear function of probability (p) in behavioral decision theory. This study proposes a psychophysical model of probability weighting functions derived from a hyperbolic time discounting model and a geometric distribution. The aim of the study is to show probability weighting functions from the point of view of waiting time for a decision maker. Since the expected value of a geometrically distributed random variable X is 1/p, we formulized the probability weighting function of the expected value model for hyperbolic time discounting as w(p) = (1 - k log p)(-1). Moreover, the probability weighting function is derived from Loewenstein and Prelec's (1992) generalized hyperbolic time discounting model. The latter model is proved to be equivalent to the hyperbolic-logarithmic weighting function considered by Prelec (1998) and Luce (2001). In this study, we derive a model from the generalized hyperbolic time discounting model assuming Fechner's (1860) psychophysical law of time and a geometric distribution of trials. In addition, we develop median models of hyperbolic time discounting and generalized hyperbolic time discounting. To illustrate the fitness of each model, a psychological experiment was conducted to assess the probability weighting and value functions at the level of the individual participant. The participants were 50 university students. The results of individual analysis indicated that the expected value model of generalized hyperbolic discounting fitted better than previous probability weighting decision-making models. The theoretical implications of this finding are discussed.
Hybrid Approaches and Industrial Applications of Pattern Recognition,
1980-10-01
emphasized that the probability distribution in (9) is correct only under the assumption that P( wIx ) is known exactly. In practice this assumption will...sufficient precision. The alternative would be to take the probability distribution of estimates of P( wix ) into account in the analysis. However, from the
Generalized Success-Breeds-Success Principle Leading to Time-Dependent Informetric Distributions.
ERIC Educational Resources Information Center
Egghe, Leo; Rousseau, Ronald
1995-01-01
Reformulates the success-breeds-success (SBS) principle in informetrics in order to generate a general theory of source-item relationships. Topics include a time-dependent probability, a new model for the expected probability that is compared with the SBS principle with exact combinatorial calculations, classical frequency distributions, and…
Measurement of the top-quark mass with dilepton events selected using neuroevolution at CDF.
Aaltonen, T; Adelman, J; Akimoto, T; Albrow, M G; Alvarez González, B; Amerio, S; Amidei, D; Anastassov, A; Annovi, A; Antos, J; Apollinari, G; Apresyan, A; Arisawa, T; Artikov, A; Ashmanskas, W; Attal, A; Aurisano, A; Azfar, F; Azzurri, P; Badgett, W; Barbaro-Galtieri, A; Barnes, V E; Barnett, B A; Bartsch, V; Bauer, G; Beauchemin, P-H; Bedeschi, F; Bednar, P; Beecher, D; Behari, S; Bellettini, G; Bellinger, J; Benjamin, D; Beretvas, A; Beringer, J; Bhatti, A; Binkley, M; Bisello, D; Bizjak, I; Blair, R E; Blocker, C; Blumenfeld, B; Bocci, A; Bodek, A; Boisvert, V; Bolla, G; Bortoletto, D; Boudreau, J; Boveia, A; Brau, B; Bridgeman, A; Brigliadori, L; Bromberg, C; Brubaker, E; Budagov, J; Budd, H S; Budd, S; Burkett, K; Busetto, G; Bussey, P; Buzatu, A; Byrum, K L; Cabrera, S; Calancha, C; Campanelli, M; Campbell, M; Canelli, F; Canepa, A; Carlsmith, D; Carosi, R; Carrillo, S; Carron, S; Casal, B; Casarsa, M; Castro, A; Catastini, P; Cauz, D; Cavaliere, V; Cavalli-Sforza, M; Cerri, A; Cerrito, L; Chang, S H; Chen, Y C; Chertok, M; Chiarelli, G; Chlachidze, G; Chlebana, F; Cho, K; Chokheli, D; Chou, J P; Choudalakis, G; Chuang, S H; Chung, K; Chung, W H; Chung, Y S; Ciobanu, C I; Ciocci, M A; Clark, A; Clark, D; Compostella, G; Convery, M E; Conway, J; Copic, K; Cordelli, M; Cortiana, G; Cox, D J; Crescioli, F; Cuenca Almenar, C; Cuevas, J; Culbertson, R; Cully, J C; Dagenhart, D; Datta, M; Davies, T; de Barbaro, P; De Cecco, S; Deisher, A; De Lorenzo, G; Dell'orso, M; Deluca, C; Demortier, L; Deng, J; Deninno, M; Derwent, P F; di Giovanni, G P; Dionisi, C; Di Ruzza, B; Dittmann, J R; D'Onofrio, M; Donati, S; Dong, P; Donini, J; Dorigo, T; Dube, S; Efron, J; Elagin, A; Erbacher, R; Errede, D; Errede, S; Eusebi, R; Fang, H C; Farrington, S; Fedorko, W T; Feild, R G; Feindt, M; Fernandez, J P; Ferrazza, C; Field, R; Flanagan, G; Forrest, R; Franklin, M; Freeman, J C; Furic, I; Gallinaro, M; Galyardt, J; Garberson, F; Garcia, J E; Garfinkel, A F; Genser, K; Gerberich, H; Gerdes, D; Gessler, A; Giagu, S; Giakoumopoulou, V; Giannetti, P; Gibson, K; Gimmell, J L; Ginsburg, C M; Giokaris, N; Giordani, M; Giromini, P; Giunta, M; Giurgiu, G; Glagolev, V; Glenzinski, D; Gold, M; Goldschmidt, N; Golossanov, A; Gomez, G; Gomez-Ceballos, G; Goncharov, M; González, O; Gorelov, I; Goshaw, A T; Goulianos, K; Gresele, A; Grinstein, S; Grosso-Pilcher, C; Grundler, U; Guimaraes da Costa, J; Gunay-Unalan, Z; Haber, C; Hahn, K; Hahn, S R; Halkiadakis, E; Han, B-Y; Han, J Y; Handler, R; Happacher, F; Hara, K; Hare, D; Hare, M; Harper, S; Harr, R F; Harris, R M; Hartz, M; Hatakeyama, K; Hauser, J; Hays, C; Heck, M; Heijboer, A; Heinemann, B; Heinrich, J; Henderson, C; Herndon, M; Heuser, J; Hewamanage, S; Hidas, D; Hill, C S; Hirschbuehl, D; Hocker, A; Hou, S; Houlden, M; Hsu, S-C; Huffman, B T; Hughes, R E; Husemann, U; Huston, J; Incandela, J; Introzzi, G; Iori, M; Ivanov, A; James, E; Jayatilaka, B; Jeon, E J; Jha, M K; Jindariani, S; Johnson, W; Jones, M; Joo, K K; Jun, S Y; Jung, J E; Junk, T R; Kamon, T; Kar, D; Karchin, P E; Kato, Y; Kephart, R; Keung, J; Khotilovich, V; Kilminster, B; Kim, D H; Kim, H S; Kim, J E; Kim, M J; Kim, S B; Kim, S H; Kim, Y K; Kimura, N; Kirsch, L; Klimenko, S; Knuteson, B; Ko, B R; Koay, S A; Kondo, K; Kong, D J; Konigsberg, J; Korytov, A; Kotwal, A V; Kreps, M; Kroll, J; Krop, D; Krumnack, N; Kruse, M; Krutelyov, V; Kubo, T; Kuhr, T; Kulkarni, N P; Kurata, M; Kusakabe, Y; Kwang, S; Laasanen, A T; Lami, S; Lammel, S; Lancaster, M; Lander, R L; Lannon, K; Lath, A; Latino, G; Lazzizzera, I; Lecompte, T; Lee, E; Lee, S W; Leone, S; Lewis, J D; Lin, C S; Linacre, J; Lindgren, M; Lipeles, E; Lister, A; Litvintsev, D O; Liu, C; Liu, T; Lockyer, N S; Loginov, A; Loreti, M; Lovas, L; Lu, R-S; Lucchesi, D; Lueck, J; Luci, C; Lujan, P; Lukens, P; Lungu, G; Lyons, L; Lys, J; Lysak, R; Lytken, E; Mack, P; Macqueen, D; Madrak, R; Maeshima, K; Makhoul, K; Maki, T; Maksimovic, P; Malde, S; Malik, S; Manca, G; Manousakis-Katsikakis, A; Margaroli, F; Marino, C; Marino, C P; Martin, A; Martin, V; Martínez, M; Martínez-Ballarín, R; Maruyama, T; Mastrandrea, P; Masubuchi, T; Mattson, M E; Mazzanti, P; McFarland, K S; McIntyre, P; McNulty, R; Mehta, A; Mehtala, P; Menzione, A; Merkel, P; Mesropian, C; Miao, T; Miladinovic, N; Miller, R; Mills, C; Milnik, M; Mitra, A; Mitselmakher, G; Miyake, H; Moggi, N; Moon, C S; Moore, R; Morello, M J; Morlok, J; Movilla Fernandez, P; Mülmenstädt, J; Mukherjee, A; Muller, Th; Mumford, R; Murat, P; Mussini, M; Nachtman, J; Nagai, Y; Nagano, A; Naganoma, J; Nakamura, K; Nakano, I; Napier, A; Necula, V; Neu, C; Neubauer, M S; Nielsen, J; Nodulman, L; Norman, M; Norniella, O; Nurse, E; Oakes, L; Oh, S H; Oh, Y D; Oksuzian, I; Okusawa, T; Orava, R; Osterberg, K; Pagan Griso, S; Pagliarone, C; Palencia, E; Papadimitriou, V; Papaikonomou, A; Paramonov, A A; Parks, B; Pashapour, S; Patrick, J; Pauletta, G; Paulini, M; Paus, C; Pellett, D E; Penzo, A; Phillips, T J; Piacentino, G; Pianori, E; Pinera, L; Pitts, K; Plager, C; Pondrom, L; Poukhov, O; Pounder, N; Prakoshyn, F; Pronko, A; Proudfoot, J; Ptohos, F; Pueschel, E; Punzi, G; Pursley, J; Rademacker, J; Rahaman, A; Ramakrishnan, V; Ranjan, N; Redondo, I; Reisert, B; Rekovic, V; Renton, P; Rescigno, M; Richter, S; Rimondi, F; Ristori, L; Robson, A; Rodrigo, T; Rodriguez, T; Rogers, E; Rolli, S; Roser, R; Rossi, M; Rossin, R; Roy, P; Ruiz, A; Russ, J; Rusu, V; Saarikko, H; Safonov, A; Sakumoto, W K; Saltó, O; Santi, L; Sarkar, S; Sartori, L; Sato, K; Savoy-Navarro, A; Scheidle, T; Schlabach, P; Schmidt, A; Schmidt, E E; Schmidt, M A; Schmidt, M P; Schmitt, M; Schwarz, T; Scodellaro, L; Scott, A L; Scribano, A; Scuri, F; Sedov, A; Seidel, S; Seiya, Y; Semenov, A; Sexton-Kennedy, L; Sfyrla, A; Shalhout, S Z; Shears, T; Shekhar, R; Shepard, P F; Sherman, D; Shimojima, M; Shiraishi, S; Shochet, M; Shon, Y; Shreyber, I; Sidoti, A; Sinervo, P; Sisakyan, A; Slaughter, A J; Slaunwhite, J; Sliwa, K; Smith, J R; Snider, F D; Snihur, R; Soha, A; Somalwar, S; Sorin, V; Spalding, J; Spreitzer, T; Squillacioti, P; Stanitzki, M; St Denis, R; Stelzer, B; Stelzer-Chilton, O; Stentz, D; Strologas, J; Stuart, D; Suh, J S; Sukhanov, A; Suslov, I; Suzuki, T; Taffard, A; Takashima, R; Takeuchi, Y; Tanaka, R; Tecchio, M; Teng, P K; Terashi, K; Thom, J; Thompson, A S; Thompson, G A; Thomson, E; Tipton, P; Tiwari, V; Tkaczyk, S; Toback, D; Tokar, S; Tollefson, K; Tomura, T; Tonelli, D; Torre, S; Torretta, D; Totaro, P; Tourneur, S; Tu, Y; Turini, N; Ukegawa, F; Vallecorsa, S; van Remortel, N; Varganov, A; Vataga, E; Vázquez, F; Velev, G; Vellidis, C; Veszpremi, V; Vidal, M; Vidal, R; Vila, I; Vilar, R; Vine, T; Vogel, M; Volobouev, I; Volpi, G; Würthwein, F; Wagner, P; Wagner, R G; Wagner, R L; Wagner-Kuhr, J; Wagner, W; Wakisaka, T; Wallny, R; Wang, S M; Warburton, A; Waters, D; Weinberger, M; Wester, W C; Whitehouse, B; Whiteson, D; Whiteson, S; Wicklund, A B; Wicklund, E; Williams, G; Williams, H H; Wilson, P; Winer, B L; Wittich, P; Wolbers, S; Wolfe, C; Wright, T; Wu, X; Wynne, S M; Xie, S; Yagil, A; Yamamoto, K; Yamaoka, J; Yang, U K; Yang, Y C; Yao, W M; Yeh, G P; Yoh, J; Yorita, K; Yoshida, T; Yu, G B; Yu, I; Yu, S S; Yun, J C; Zanello, L; Zanetti, A; Zaw, I; Zhang, X; Zheng, Y; Zucchelli, S
2009-04-17
We report a measurement of the top-quark mass M_{t} in the dilepton decay channel tt[over ] --> bl;{'+} nu_{l};{'}b[over ]l;{-}nu[over ]_{l}. Events are selected with a neural network which has been directly optimized for statistical precision in top-quark mass using neuroevolution, a technique modeled on biological evolution. The top-quark mass is extracted from per-event probability densities that are formed by the convolution of leading order matrix elements and detector resolution functions. The joint probability is the product of the probability densities from 344 candidate events in 2.0 fb;{-1} of pp[over ] collisions collected with the CDF II detector, yielding a measurement of M_{t} = 171.2 +/- 2.7(stat) +/- 2.9(syst) GeV / c;{2}.
ERIC Educational Resources Information Center
Sayeski, Kristin L.; Earle, Gentry A.; Eslinger, R. Paige; Whitenton, Jessy N.
2017-01-01
Matching phonemes (speech sounds) to graphemes (letters and letter combinations) is an important aspect of decoding (translating print to speech) and encoding (translating speech to print). Yet, many teacher candidates do not receive explicit training in phoneme-grapheme correspondence. Difficulty with accurate phoneme production and/or lack of…
29 CFR 452.80 - Bona fide candidates.
Code of Federal Regulations, 2010 CFR
2010-07-01
... nominated in order to be a bona fide candidate entitled to exercise the rights mentioned in §§ 452.67 and... able to take advantage of the distribution rights even before the convention meets and thus attempt to.... United Mine Workers, 71 LRRM 2606, 60 L.C. 10,204 (D.D.C. 1969). ...
The beta distribution: A statistical model for world cloud cover
NASA Technical Reports Server (NTRS)
Falls, L. W.
1973-01-01
Much work has been performed in developing empirical global cloud cover models. This investigation was made to determine an underlying theoretical statistical distribution to represent worldwide cloud cover. The beta distribution with probability density function is given to represent the variability of this random variable. It is shown that the beta distribution possesses the versatile statistical characteristics necessary to assume the wide variety of shapes exhibited by cloud cover. A total of 160 representative empirical cloud cover distributions were investigated and the conclusion was reached that this study provides sufficient statical evidence to accept the beta probability distribution as the underlying model for world cloud cover.
Core compressor exit stage study, volume 6
NASA Technical Reports Server (NTRS)
Wisler, D. C.
1981-01-01
Rear stage blading designs that have lower losses in their endwall boundary layer regions were studied. A baseline Stage A was designed as a low-speed model of stage 7 of a 10-stage compressor. Candidate rotors and stators were designed which have the potential of reducing endwall losses relative to the baseline. Rotor B uses a type of meanline in the tip region that unloads the leading edge and loads the trailing edge relative to the baseline rotor A designs. Rotor C incorporates a more skewed (hub strong) radial distribution of total pressure and smoother distribution of static pressure on the rotor tip than those of rotor B. Candidate stator B embodies twist gradients in the endwall region. Stator C embodies airfoil sections near the endwalls that have reduced trailing edge loading relative to stator A. The baseline and candidate bladings were tested using four identical stages to produce a true multistage environment. Single-stage tests were also conducted. The test data were analyzed and performances were compared. Several of the candidate configurations showed a performance improvement relative to the baseline.
May, Eric F; Lim, Vincent W; Metaxas, Peter J; Du, Jianwei; Stanwix, Paul L; Rowland, Darren; Johns, Michael L; Haandrikman, Gert; Crosby, Daniel; Aman, Zachary M
2018-03-13
Gas hydrate formation is a stochastic phenomenon of considerable significance for any risk-based approach to flow assurance in the oil and gas industry. In principle, well-established results from nucleation theory offer the prospect of predictive models for hydrate formation probability in industrial production systems. In practice, however, heuristics are relied on when estimating formation risk for a given flowline subcooling or when quantifying kinetic hydrate inhibitor (KHI) performance. Here, we present statistically significant measurements of formation probability distributions for natural gas hydrate systems under shear, which are quantitatively compared with theoretical predictions. Distributions with over 100 points were generated using low-mass, Peltier-cooled pressure cells, cycled in temperature between 40 and -5 °C at up to 2 K·min -1 and analyzed with robust algorithms that automatically identify hydrate formation and initial growth rates from dynamic pressure data. The application of shear had a significant influence on the measured distributions: at 700 rpm mass-transfer limitations were minimal, as demonstrated by the kinetic growth rates observed. The formation probability distributions measured at this shear rate had mean subcoolings consistent with theoretical predictions and steel-hydrate-water contact angles of 14-26°. However, the experimental distributions were substantially wider than predicted, suggesting that phenomena acting on macroscopic length scales are responsible for much of the observed stochastic formation. Performance tests of a KHI provided new insights into how such chemicals can reduce the risk of hydrate blockage in flowlines. Our data demonstrate that the KHI not only reduces the probability of formation (by both shifting and sharpening the distribution) but also reduces hydrate growth rates by a factor of 2.
NASA Astrophysics Data System (ADS)
Piani, L.; Yurimoto, H.; Remusat, L.; Gonzales, A.; Marty, B.
2017-12-01
Chondrite meteorites are fragments of rocks coming from small bodies of the asteroid belt and constitute witnesses of the volatile-rich reservoirs present in the inner protoplanetary disk. Among these meteorites, carbonaceous chondrites contain the largest quantity of water and organic matter and are one of the most probable candidates for the delivery of water and molecular origin of life to Earth. Organic matter in carbonaceous chondrites is intimately mixed with hydrated minerals challenging its in situ characterization and the determination of its H-isotope composition (Le Guillou et al., GCA 131, 2014). Organic matter occurs as soluble components (in water or organic solvents) and an insoluble macromolecule. The insoluble organic matter (IOM) is efficiently isolated after acid leaching of the chondrite minerals. IOM has thus been investigated by a large set of analytical techniques allowing its structural organization, chemical composition and isotopic composition to be determined at several scales (e.g. Derenne and Robert, MAPS 45, 2010). In the soluble counterpart (SOM), targeted studies have shown large ranges of D/H ratios in the different classes of soluble organic compounds (i.e. carboxylic acids, ketones and aldehydes, amino-acids etc.) (Remusat, Planetary Mineralogy 15, 2015 and references therein). This D/H distribution indicates a complex and probably multiple-stage synthesis of this organic compounds occurring at different stages of the disk evolution. Nevertheless, inventories of the known C-bearing species in carbonaceous chondrites (carbonates, SOM and IOM) show that about 40-50 % of the carbon is hidden within the matrix (Alexander et al., MAPS 50, 2015). In this study, we perform in situ hydrogen isotope analyses at the micrometer scale by secondary ion mass spectrometry to investigate the distribution of organic matter in primitive chondrites without the use of any chemical treatment. Correlated analyses of the D/H and C/H ratios allow us to decipher the H contribution of water-bearing minerals and to estimate the hydrogen isotopic composition of water in chondrites (Piani et al., submitted). Comparison of spot analyses and isotope images obtained in situ and on isolated IOM gives clues on the nature of the organic components of carbonaceous asteroid rocks.
Quantum work in the Bohmian framework
NASA Astrophysics Data System (ADS)
Sampaio, R.; Suomela, S.; Ala-Nissila, T.; Anders, J.; Philbin, T. G.
2018-01-01
At nonzero temperature classical systems exhibit statistical fluctuations of thermodynamic quantities arising from the variation of the system's initial conditions and its interaction with the environment. The fluctuating work, for example, is characterized by the ensemble of system trajectories in phase space and, by including the probabilities for various trajectories to occur, a work distribution can be constructed. However, without phase-space trajectories, the task of constructing a work probability distribution in the quantum regime has proven elusive. Here we use quantum trajectories in phase space and define fluctuating work as power integrated along the trajectories, in complete analogy to classical statistical physics. The resulting work probability distribution is valid for any quantum evolution, including cases with coherences in the energy basis. We demonstrate the quantum work probability distribution and its properties with an exactly solvable example of a driven quantum harmonic oscillator. An important feature of the work distribution is its dependence on the initial statistical mixture of pure states, which is reflected in higher moments of the work. The proposed approach introduces a fundamentally different perspective on quantum thermodynamics, allowing full thermodynamic characterization of the dynamics of quantum systems, including the measurement process.
Applying the log-normal distribution to target detection
NASA Astrophysics Data System (ADS)
Holst, Gerald C.
1992-09-01
Holst and Pickard experimentally determined that MRT responses tend to follow a log-normal distribution. The log normal distribution appeared reasonable because nearly all visual psychological data is plotted on a logarithmic scale. It has the additional advantage that it is bounded to positive values; an important consideration since probability of detection is often plotted in linear coordinates. Review of published data suggests that the log-normal distribution may have universal applicability. Specifically, the log-normal distribution obtained from MRT tests appears to fit the target transfer function and the probability of detection of rectangular targets.
Tsunami Size Distributions at Far-Field Locations from Aggregated Earthquake Sources
NASA Astrophysics Data System (ADS)
Geist, E. L.; Parsons, T.
2015-12-01
The distribution of tsunami amplitudes at far-field tide gauge stations is explained by aggregating the probability of tsunamis derived from individual subduction zones and scaled by their seismic moment. The observed tsunami amplitude distributions of both continental (e.g., San Francisco) and island (e.g., Hilo) stations distant from subduction zones are examined. Although the observed probability distributions nominally follow a Pareto (power-law) distribution, there are significant deviations. Some stations exhibit varying degrees of tapering of the distribution at high amplitudes and, in the case of the Hilo station, there is a prominent break in slope on log-log probability plots. There are also differences in the slopes of the observed distributions among stations that can be significant. To explain these differences we first estimate seismic moment distributions of observed earthquakes for major subduction zones. Second, regression models are developed that relate the tsunami amplitude at a station to seismic moment at a subduction zone, correcting for epicentral distance. The seismic moment distribution is then transformed to a site-specific tsunami amplitude distribution using the regression model. Finally, a mixture distribution is developed, aggregating the transformed tsunami distributions from all relevant subduction zones. This mixture distribution is compared to the observed distribution to assess the performance of the method described above. This method allows us to estimate the largest tsunami that can be expected in a given time period at a station.
A New Way to Confirm Planet Candidates
NASA Astrophysics Data System (ADS)
Kohler, Susanna
2016-05-01
What was the big deal behind the Kepler news conference yesterday? Its not just that the number of confirmed planets found by Kepler has more than doubled (though thats certainly exciting news!). Whats especially interesting is the way in which these new planets were confirmed.Number of planet discoveries by year since 1995, including previous non-Kepler discoveries (blue), previous Kepler discoveries (light blue) and the newly validated Kepler planets (orange). [NASA Ames/W. Stenzel; Princeton University/T. Morton]No Need for Follow-UpBefore Kepler, the way we confirmed planet candidates was with follow-up observations. The candidate could be validated either by directly imaging (which is rare) or obtaining a large number radial-velocity measurements of the wobble of the planets host star due to the planets orbit. But once Kepler started producing planet candidates, these approaches to validation became less feasible. A lot of Kepler candidates are small and orbit faint stars, making follow-up observations difficult or impossible.This problem is what inspired the development of whats known as probabilistic validation, an analysis technique that involves assessing the likelihood that the candidates signal is caused by various false-positive scenarios. Using this technique allows astronomers to estimate the likelihood of a candidate signal being a true planet detection; if that likelihood is high enough, the planet candidate can be confirmed without the need for follow-up observations.A breakdown of the catalog of Kepler Objects of Interest. Just over half had previously been identified as false positives or confirmed as candidates. 1284 are newly validated, and another 455 have FPP of1090%. [Morton et al. 2016]Probabilistic validation has been used in the past to confirm individual planet candidates in Kepler data, but now Timothy Morton (Princeton University) and collaborators have taken this to a new level: they developed the first code thats designed to do fully automated batch processing of a large number of candidates.In a recently published study the results of which were announced yesterday the teamapplied their code to the entire catalog of 7,470 Kepler objects of interest.New Planets and False PositivesThe teams code was able to successfully evaluate the total false-positive probability (FPP) for 7,056 of the objects of interest. Of these, 428 objects previously identified as candidates were found to have FPP of more than 90%, suggesting that they are most likely false positives.Periods and radii of candidate and confirmed planets in the Kepler Objects of Interest catalog. Blue circles have previously been identified as confirmed planets. Candidates (orange) are shaded by false positive probability; more transparent means more likely to be a false positive. [Morton et al. 2016]In contrast, 1,935 candidates were found to have FPP of less than 1%, and were therefore declared validated planets. Of these confirmations, 1,284 were previously unconfirmed, more than doubling Keplers previous catalog of 1,041 confirmed planets. Morton and collaborators believe that 9 of these newly confirmed planets may fall within the habitable zone of their host stars.While the announcement of 1,284 newly confirmed planets is huge, the analysis presented in this study is the real news. The code used is publicly available and can be applied to any transiting exoplanet candidate. This means that this analysis technique can be used to find batches of exoplanets in data from the extended Kepler mission (K2) or from the future TESS and PLATO transit missions.CitationTimothy D. Morton et al 2016 ApJ 822 86. doi:10.3847/0004-637X/822/2/86
Asquith, William H.; Kiang, Julie E.; Cohn, Timothy A.
2017-07-17
The U.S. Geological Survey (USGS), in cooperation with the U.S. Nuclear Regulatory Commission, has investigated statistical methods for probabilistic flood hazard assessment to provide guidance on very low annual exceedance probability (AEP) estimation of peak-streamflow frequency and the quantification of corresponding uncertainties using streamgage-specific data. The term “very low AEP” implies exceptionally rare events defined as those having AEPs less than about 0.001 (or 1 × 10–3 in scientific notation or for brevity 10–3). Such low AEPs are of great interest to those involved with peak-streamflow frequency analyses for critical infrastructure, such as nuclear power plants. Flood frequency analyses at streamgages are most commonly based on annual instantaneous peak streamflow data and a probability distribution fit to these data. The fitted distribution provides a means to extrapolate to very low AEPs. Within the United States, the Pearson type III probability distribution, when fit to the base-10 logarithms of streamflow, is widely used, but other distribution choices exist. The USGS-PeakFQ software, implementing the Pearson type III within the Federal agency guidelines of Bulletin 17B (method of moments) and updates to the expected moments algorithm (EMA), was specially adapted for an “Extended Output” user option to provide estimates at selected AEPs from 10–3 to 10–6. Parameter estimation methods, in addition to product moments and EMA, include L-moments, maximum likelihood, and maximum product of spacings (maximum spacing estimation). This study comprehensively investigates multiple distributions and parameter estimation methods for two USGS streamgages (01400500 Raritan River at Manville, New Jersey, and 01638500 Potomac River at Point of Rocks, Maryland). The results of this study specifically involve the four methods for parameter estimation and up to nine probability distributions, including the generalized extreme value, generalized log-normal, generalized Pareto, and Weibull. Uncertainties in streamflow estimates for corresponding AEP are depicted and quantified as two primary forms: quantile (aleatoric [random sampling] uncertainty) and distribution-choice (epistemic [model] uncertainty). Sampling uncertainties of a given distribution are relatively straightforward to compute from analytical or Monte Carlo-based approaches. Distribution-choice uncertainty stems from choices of potentially applicable probability distributions for which divergence among the choices increases as AEP decreases. Conventional goodness-of-fit statistics, such as Cramér-von Mises, and L-moment ratio diagrams are demonstrated in order to hone distribution choice. The results generally show that distribution choice uncertainty is larger than sampling uncertainty for very low AEP values.
Polynomial probability distribution estimation using the method of moments
Mattsson, Lars; Rydén, Jesper
2017-01-01
We suggest a procedure for estimating Nth degree polynomial approximations to unknown (or known) probability density functions (PDFs) based on N statistical moments from each distribution. The procedure is based on the method of moments and is setup algorithmically to aid applicability and to ensure rigor in use. In order to show applicability, polynomial PDF approximations are obtained for the distribution families Normal, Log-Normal, Weibull as well as for a bimodal Weibull distribution and a data set of anonymized household electricity use. The results are compared with results for traditional PDF series expansion methods of Gram–Charlier type. It is concluded that this procedure is a comparatively simple procedure that could be used when traditional distribution families are not applicable or when polynomial expansions of probability distributions might be considered useful approximations. In particular this approach is practical for calculating convolutions of distributions, since such operations become integrals of polynomial expressions. Finally, in order to show an advanced applicability of the method, it is shown to be useful for approximating solutions to the Smoluchowski equation. PMID:28394949
Polynomial probability distribution estimation using the method of moments.
Munkhammar, Joakim; Mattsson, Lars; Rydén, Jesper
2017-01-01
We suggest a procedure for estimating Nth degree polynomial approximations to unknown (or known) probability density functions (PDFs) based on N statistical moments from each distribution. The procedure is based on the method of moments and is setup algorithmically to aid applicability and to ensure rigor in use. In order to show applicability, polynomial PDF approximations are obtained for the distribution families Normal, Log-Normal, Weibull as well as for a bimodal Weibull distribution and a data set of anonymized household electricity use. The results are compared with results for traditional PDF series expansion methods of Gram-Charlier type. It is concluded that this procedure is a comparatively simple procedure that could be used when traditional distribution families are not applicable or when polynomial expansions of probability distributions might be considered useful approximations. In particular this approach is practical for calculating convolutions of distributions, since such operations become integrals of polynomial expressions. Finally, in order to show an advanced applicability of the method, it is shown to be useful for approximating solutions to the Smoluchowski equation.
Long-term Spectral Evolution of Tidal Disruption Candidates Selected by Strong Coronal Lines
NASA Astrophysics Data System (ADS)
Yang, Chen-Wei; Wang, Ting-Gui; Ferland, Gary; Yuan, Weimin; Zhou, Hong-Yan; Jiang, Peng
2013-09-01
We present results of follow-up optical spectroscopic Multi-Mirror Telescope (MMT) observations of seven rare, extreme coronal line-emitting galaxies reported by Wang et al. Large variations in coronal lines are found in four objects, making them strong candidates for tidal disruption events (TDEs). For the four TDE candidates, all the coronal lines with ionization states higher than [Fe VII] disappear within 5-9 yr. The [Fe VII] line faded by a factor of about five in one object (J0952+2143) within 4 yr, whereas the line emerged in another two objects that previously did not show the line. A strong increment in the [O III] flux is observed, shifting the line ratios toward the loci of active galactic nuclei on the BPT diagram. Surprisingly, we detect a non-canonical [O III] λ5007/[O III] λ4959 ratio of ~= 2 in two objects, indicating a large column density of O2 + and thus probably optically thick gas. This result also requires a very large ionization parameter and a relatively soft ionizing spectral energy distribution (e.g., a blackbody with T < 5 × 104 K). Our observations can be explained as the echoing of a strong ultraviolet to soft X-ray flare caused by TDEs on molecular clouds in the inner parsecs of the galactic nuclei. Reanalyzing the Sloan Digital Sky Survey spectra reveals double-peaked or strongly blue-shouldered broad lines in three of the objects, which disappeared in the MMT spectra of two objects and faded by a factor of 10 in 8 yr in the remaining object with a decrease in both the line width and centroid offset. We interpret these broad lines as arising from decelerating biconical outflows. Our results demonstrate that the signatures of echoing can persist for as long as 10 yr and can be used to probe the gas environment in quiescent galactic nuclei.
NASA Astrophysics Data System (ADS)
Cajiao Vélez, F.; Kamiński, J. Z.; Krajewska, K.
2018-04-01
High-energy photoionization driven by short and circularly-polarized laser pulses is studied in the framework of the relativistic strong-field approximation. The saddle-point analysis of the integrals defining the probability amplitude is used to determine the general properties of the probability distributions. Additionally, an approximate solution to the saddle-point equation is derived. This leads to the concept of the three-dimensional spiral of life in momentum space, around which the ionization probability distribution is maximum. We demonstrate that such spiral is also obtained from a classical treatment.
Wu, Kun-Feng; Donnell, Eric T; Aguero-Valverde, Jonathan
2014-06-01
To approach the goal of "Toward Zero Deaths," there is a need to develop an analysis paradigm to better understand the effects of a countermeasure on reducing the number of severe crashes. One of the goals in traffic safety research is to search for an effective treatment to reduce fatal and major injury crashes, referred to as severe crashes. To achieve this goal, the selection of promising countermeasures is of utmost importance, and relies on the effectiveness of candidate countermeasures in reducing severe crashes. Although it is important to precisely evaluate the effectiveness of candidate countermeasures in reducing the number of severe crashes at a site, the current state-of-the-practice often leads to biased estimates. While there have been a few advanced statistical models developed to mitigate the problem in practice, these models are computationally difficult to estimate because severe crashes are dispersed spatially and temporally, and cannot be integrated into the Highway Safety Manual framework, which develops a series of safety performance functions and crash modification factors to predict the number of crashes. Crash severity outcomes are generally integrated into the Highway Safety Manual using deterministic distributions rather than statistical models. Accounting for the variability in crash severity as a function geometric design, traffic flow, and other roadway and roadside features is afforded by estimating statistical models. Therefore, there is a need to develop a new analysis paradigm to resolve the limitations in the current Highway Safety Manual methods. We propose an approach which decomposes the severe crash frequency into a function of the change in the total number of crashes and the probability of a crash becoming a severe crash before and after a countermeasure is implemented. We tested this approach by evaluating the effectiveness of shoulder rumble strips on reducing the number of severe crashes. A total of 310 segments that have had shoulder rumble strips installed during 2002-2009 are included in the analysis. It was found that shoulder rumble strips reduce the total number of crashes, but have no statistically significant effect on reducing the probability of a severe crash outcome. Copyright © 2014 Elsevier Ltd. All rights reserved.
Schmidt, Benedikt R
2003-08-01
The evidence for amphibian population declines is based on count data that were not adjusted for detection probabilities. Such data are not reliable even when collected using standard methods. The formula C = Np (where C is a count, N the true parameter value, and p is a detection probability) relates count data to demography, population size, or distributions. With unadjusted count data, one assumes a linear relationship between C and N and that p is constant. These assumptions are unlikely to be met in studies of amphibian populations. Amphibian population data should be based on methods that account for detection probabilities.
Bayesian data analysis tools for atomic physics
NASA Astrophysics Data System (ADS)
Trassinelli, Martino
2017-10-01
We present an introduction to some concepts of Bayesian data analysis in the context of atomic physics. Starting from basic rules of probability, we present the Bayes' theorem and its applications. In particular we discuss about how to calculate simple and joint probability distributions and the Bayesian evidence, a model dependent quantity that allows to assign probabilities to different hypotheses from the analysis of a same data set. To give some practical examples, these methods are applied to two concrete cases. In the first example, the presence or not of a satellite line in an atomic spectrum is investigated. In the second example, we determine the most probable model among a set of possible profiles from the analysis of a statistically poor spectrum. We show also how to calculate the probability distribution of the main spectral component without having to determine uniquely the spectrum modeling. For these two studies, we implement the program Nested_fit to calculate the different probability distributions and other related quantities. Nested_fit is a Fortran90/Python code developed during the last years for analysis of atomic spectra. As indicated by the name, it is based on the nested algorithm, which is presented in details together with the program itself.
NASA Astrophysics Data System (ADS)
Batac, Rene C.; Paguirigan, Antonino A., Jr.; Tarun, Anjali B.; Longjas, Anthony G.
2017-04-01
We propose a cellular automata model for earthquake occurrences patterned after the sandpile model of self-organized criticality (SOC). By incorporating a single parameter describing the probability to target the most susceptible site, the model successfully reproduces the statistical signatures of seismicity. The energy distributions closely follow power-law probability density functions (PDFs) with a scaling exponent of around -1. 6, consistent with the expectations of the Gutenberg-Richter (GR) law, for a wide range of the targeted triggering probability values. Additionally, for targeted triggering probabilities within the range 0.004-0.007, we observe spatiotemporal distributions that show bimodal behavior, which is not observed previously for the original sandpile. For this critical range of values for the probability, model statistics show remarkable comparison with long-period empirical data from earthquakes from different seismogenic regions. The proposed model has key advantages, the foremost of which is the fact that it simultaneously captures the energy, space, and time statistics of earthquakes by just introducing a single parameter, while introducing minimal parameters in the simple rules of the sandpile. We believe that the critical targeting probability parameterizes the memory that is inherently present in earthquake-generating regions.
NASA Astrophysics Data System (ADS)
Safarzadeh, Mohammadtaher; Ji, Alexander P.; Dooley, Gregory A.; Frebel, Anna; Scannapieco, Evan; Gómez, Facundo A.; O'Shea, Brian W.
2018-06-01
The smallest satellites of the Milky Way ceased forming stars during the epoch of reionization and thus provide archaeological access to galaxy formation at z > 6. Numerical studies of these ultrafaint dwarf galaxies (UFDs) require expensive cosmological simulations with high mass resolution that are carried out down to z = 0. However, if we are able to statistically identify UFD host progenitors at high redshifts with relatively high probabilities, we can avoid this high computational cost. To find such candidates, we analyse the merger trees of Milky Way type haloes from the high-resolution Caterpillar suite of dark matter only simulations. Satellite UFD hosts at z = 0 are identified based on four different abundance matching (AM) techniques. All the haloes at high redshifts are traced forward in time in order to compute the probability of surviving as satellite UFDs today. Our results show that selecting potential UFD progenitors based solely on their mass at z = 12 (8) results in a 10 per cent (20 per cent) chance of obtaining a surviving UFD at z = 0 in three of the AM techniques we adopted. We find that the progenitors of surviving satellite UFDs have lower virial ratios (η), and are preferentially located at large distances from the main MW progenitor, while they show no correlation with concentration parameter. Haloes with favorable locations and virial ratios are ≈3 times more likely to survive as satellite UFD candidates at z = 0.
An Ultraviolet-Excess Optical Candidate for the Luminous Globular Cluster X-Ray Source in NGC 1851
NASA Technical Reports Server (NTRS)
Deutsch, Eric W.; Anderson, Scott F.; Margon, Bruce; Downes, Ronald A.
1996-01-01
The intense, bursting X-ray source in the globular cluster NGC 1851 was one of the first cluster sources discovered, but has remained optically unidentified for 25 years. We report here on results from Hubble Space Telescope WFPC2 multicolor images in NGC 1851. Our high spatial resolution images resolve approximately 200 objects in the 3 minute radius Einstein X-ray error circle, 40 times as many as in previous ground-based work. A color-magnitude diagram of the cluster clearly reveals a markedly UV-excess object with B approximately 21, (U - B) approximately -0.9, only 2 minutes from the X-ray position. The UV-excess candidate is 0.12 minutes distant from a second, unremarkable star that is 0.5 mag brighter in B; thus ground-based studies of this field are probably impractical. Three other UV-excess objects are also present among the approximately 16,000 objects in the surveyed region of the cluster, leaving an approximately 5% probability that a UV-excess object has fallen in the X-ray error circle by chance. No variability of the candidate is seen in these data, although a more complete study is required. If this object is in fact the counterpart of the X-ray source, previous inferences that some globular cluster X-ray sources are optically subluminous with respect to low-mass X-ray binaries in the field are now strengthened.
Characterising RNA secondary structure space using information entropy
2013-01-01
Comparative methods for RNA secondary structure prediction use evolutionary information from RNA alignments to increase prediction accuracy. The model is often described in terms of stochastic context-free grammars (SCFGs), which generate a probability distribution over secondary structures. It is, however, unclear how this probability distribution changes as a function of the input alignment. As prediction programs typically only return a single secondary structure, better characterisation of the underlying probability space of RNA secondary structures is of great interest. In this work, we show how to efficiently compute the information entropy of the probability distribution over RNA secondary structures produced for RNA alignments by a phylo-SCFG, and implement it for the PPfold model. We also discuss interpretations and applications of this quantity, including how it can clarify reasons for low prediction reliability scores. PPfold and its source code are available from http://birc.au.dk/software/ppfold/. PMID:23368905
Nielsen, Bjørn G; Jensen, Morten Ø; Bohr, Henrik G
2003-01-01
The structure of enkephalin, a small neuropeptide with five amino acids, has been simulated on computers using molecular dynamics. Such simulations exhibit a few stable conformations, which also have been identified experimentally. The simulations provide the possibility to perform cluster analysis in the space defined by potentially pharmacophoric measures such as dihedral angles, side-chain orientation, etc. By analyzing the statistics of the resulting clusters, the probability distribution of the side-chain conformations may be determined. These probabilities allow us to predict the selectivity of [Leu]enkephalin and [Met]enkephalin to the known mu- and delta-type opiate receptors to which they bind as agonists. Other plausible consequences of these probability distributions are discussed in relation to the way in which they may influence the dynamics of the synapse. Copyright 2003 Wiley Periodicals, Inc. Biopolymers (Pept Sci) 71: 577-592, 2003
Exact probability distribution function for the volatility of cumulative production
NASA Astrophysics Data System (ADS)
Zadourian, Rubina; Klümper, Andreas
2018-04-01
In this paper we study the volatility and its probability distribution function for the cumulative production based on the experience curve hypothesis. This work presents a generalization of the study of volatility in Lafond et al. (2017), which addressed the effects of normally distributed noise in the production process. Due to its wide applicability in industrial and technological activities we present here the mathematical foundation for an arbitrary distribution function of the process, which we expect will pave the future research on forecasting of the production process.
Habitable-zone super-Earth candidate in a six-planet system around the K2.5V star HD 40307
NASA Astrophysics Data System (ADS)
Tuomi, M.; Anglada-Escudé, G.; Gerlach, E.; Jones, H. R. A.; Reiners, A.; Rivera, E. J.; Vogt, S. S.; Butler, R. P.
2013-01-01
Context. The K2.5 dwarf HD 40307 has been reported to host three super-Earths. The system lacks massive planets and is therefore a potential candidate for having additional low-mass planetary companions. Aims: We re-derive Doppler measurements from public HARPS spectra of HD 40307 to confirm the significance of the reported signals using independent data analysis methods. We also investigate these measurements for additional low-amplitude signals. Methods: We used Bayesian analysis of our radial velocities to estimate the probability densities of different model parameters. We also estimated the relative probabilities of models with differing numbers of Keplerian signals and verified their significance using periodogram analyses. We investigated the relation of the detected signals with the chromospheric emission of the star. As previously reported for other objects, we found that radial velocity signals correlated with the S-index are strongly wavelength dependent. Results: We identify two additional clear signals with periods of 34 and 51 days, both corresponding to planet candidates with minimum masses a few times that of the Earth. An additional sixth candidate is initially found at a period of 320 days. However, this signal correlates strongly with the chromospheric emission from the star and is also strongly wavelength dependent. When analysing the red half of the spectra only, the five putative planetary signals are recovered together with a very significant periodicity at about 200 days. This signal has a similar amplitude as the other new signals reported in the current work and corresponds to a planet candidate with Msini ~ 7 M⊕ (HD 40307 g). Conclusions: We show that Doppler measurements can be filtered for activity-induced signals if enough photons and a sufficient wavelength interval are available. If the signal corresponding to HD 40307 g is a genuine Doppler signal of planetary origin, this candidate planet might be capable of supporting liquid water on its surface according to the current definition of the liquid water habitable zone around a star and is not likely to suffer from tidal locking. Also, at an angular separation of ~46 mas, HD 40307 g would be a primary target for a future space-based direct-imaging mission. Appendix A is available in electronic form at http://www.aanda.org
Gladysz, Szymon; Yaitskova, Natalia; Christou, Julian C
2010-11-01
This paper is an introduction to the problem of modeling the probability density function of adaptive-optics speckle. We show that with the modified Rician distribution one cannot describe the statistics of light on axis. A dual solution is proposed: the modified Rician distribution for off-axis speckle and gamma-based distribution for the core of the point spread function. From these two distributions we derive optimal statistical discriminators between real sources and quasi-static speckles. In the second part of the paper the morphological difference between the two probability density functions is used to constrain a one-dimensional, "blind," iterative deconvolution at the position of an exoplanet. Separation of the probability density functions of signal and speckle yields accurate differential photometry in our simulations of the SPHERE planet finder instrument.
Wang, S Q; Zhang, H Y; Li, Z L
2016-10-01
Understanding spatio-temporal distribution of pest in orchards can provide important information that could be used to design monitoring schemes and establish better means for pest control. In this study, the spatial and temporal distribution of Bactrocera minax (Enderlein) (Diptera: Tephritidae) was assessed, and activity trends were evaluated by using probability kriging. Adults of B. minax were captured in two successive occurrences in a small-scale citrus orchard by using food bait traps, which were placed both inside and outside the orchard. The weekly spatial distribution of B. minax within the orchard and adjacent woods was examined using semivariogram parameters. The edge concentration was discovered during the most weeks in adult occurrence, and the population of the adults aggregated with high probability within a less-than-100-m-wide band on both of the sides of the orchard and the woods. The sequential probability kriged maps showed that the adults were estimated in the marginal zone with higher probability, especially in the early and peak stages. The feeding, ovipositing, and mating behaviors of B. minax are possible explanations for these spatio-temporal patterns. Therefore, spatial arrangement and distance to the forest edge of traps or spraying spot should be considered to enhance pest control on B. minax in small-scale orchards.
"A Richness Study of 14 Distant X-Ray Clusters from the 160 Square Degree Survey"
NASA Technical Reports Server (NTRS)
Jones, Christine; West, Donald (Technical Monitor)
2001-01-01
We have measured the surface density of galaxies toward 14 X-ray-selected cluster candidates at redshifts z(sub i) 0.46, and we show that they are associated with rich galaxy concentrations. These clusters, having X-ray luminosities of Lx(0.5-2 keV) approx. (0.5 - 2.6) x 10(exp 44) ergs/ sec are among the most distant and luminous in our 160 deg(exp 2) ROSAT Position Sensitive Proportional Counter cluster survey. We find that the clusters range between Abell richness classes 0 and 2 and have a most probable richness class of 1. We compare the richness distribution of our distant clusters to those for three samples of nearby clusters with similar X-ray luminosities. We find that the nearby and distant samples have similar richness distributions, which shows that clusters have apparently not evolved substantially in richness since redshift z=0.5. There is, however, a marginal tendency for the distant clusters to be slightly poorer than nearby clusters, although deeper multicolor data for a large sample would be required to confirm this trend. We compare the distribution of distant X-ray clusters in the L(sub X)-richness plane to the distribution of optically selected clusters from the Palomar Distant Cluster Survey. The optically selected clusters appear overly rich for their X-ray luminosities, when compared to X-ray-selected clusters. Apparently, X-ray and optical surveys do not necessarily sample identical mass concentrations at large redshifts. This may indicate the existence of a population of optically rich clusters with anomalously low X-ray emission, More likely, however, it reflects the tendency for optical surveys to select unvirialized mass concentrations, as might be expected when peering along large-scale filaments.
The Detection of Signals in Impulsive Noise.
1983-06-01
ASSI FICATION/ DOWN GRADING SCHEOUL1E * I1S. DISTRIBUTION STATEMENT (of th0i0 Rhport) Approved for Public Release; Distribucion Unlimited * 17...has a symmetric distribution, sgn(x i) will be -1 with probability 1/2 and +1 with probability 1/2. Considering the sum of observations as 0 binomial
DOE Office of Scientific and Technical Information (OSTI.GOV)
Diwaker, E-mail: diwakerphysics@gmail.com; Chakraborty, Aniruddha
The Smoluchowski equation with a time-dependent sink term is solved exactly. In this method, knowing the probability distribution P(0, s) at the origin, allows deriving the probability distribution P(x, s) at all positions. Exact solutions of the Smoluchowski equation are also provided in different cases where the sink term has linear, constant, inverse, and exponential variation in time.
NASA Astrophysics Data System (ADS)
Salama, Paul
2008-02-01
Multi-photon microscopy has provided biologists with unprecedented opportunities for high resolution imaging deep into tissues. Unfortunately deep tissue multi-photon microscopy images are in general noisy since they are acquired at low photon counts. To aid in the analysis and segmentation of such images it is sometimes necessary to initially enhance the acquired images. One way to enhance an image is to find the maximum a posteriori (MAP) estimate of each pixel comprising an image, which is achieved by finding a constrained least squares estimate of the unknown distribution. In arriving at the distribution it is assumed that the noise is Poisson distributed, the true but unknown pixel values assume a probability mass function over a finite set of non-negative values, and since the observed data also assumes finite values because of low photon counts, the sum of the probabilities of the observed pixel values (obtained from the histogram of the acquired pixel values) is less than one. Experimental results demonstrate that it is possible to closely estimate the unknown probability mass function with these assumptions.
Probability distribution for the Gaussian curvature of the zero level surface of a random function
NASA Astrophysics Data System (ADS)
Hannay, J. H.
2018-04-01
A rather natural construction for a smooth random surface in space is the level surface of value zero, or ‘nodal’ surface f(x,y,z) = 0, of a (real) random function f; the interface between positive and negative regions of the function. A physically significant local attribute at a point of a curved surface is its Gaussian curvature (the product of its principal curvatures) because, when integrated over the surface it gives the Euler characteristic. Here the probability distribution for the Gaussian curvature at a random point on the nodal surface f = 0 is calculated for a statistically homogeneous (‘stationary’) and isotropic zero mean Gaussian random function f. Capitalizing on the isotropy, a ‘fixer’ device for axes supplies the probability distribution directly as a multiple integral. Its evaluation yields an explicit algebraic function with a simple average. Indeed, this average Gaussian curvature has long been known. For a non-zero level surface instead of the nodal one, the probability distribution is not fully tractable, but is supplied as an integral expression.
A mechanism producing power law etc. distributions
NASA Astrophysics Data System (ADS)
Li, Heling; Shen, Hongjun; Yang, Bin
2017-07-01
Power law distribution is playing an increasingly important role in the complex system study. Based on the insolvability of complex systems, the idea of incomplete statistics is utilized and expanded, three different exponential factors are introduced in equations about the normalization condition, statistical average and Shannon entropy, with probability distribution function deduced about exponential function, power function and the product form between power function and exponential function derived from Shannon entropy and maximal entropy principle. So it is shown that maximum entropy principle can totally replace equal probability hypothesis. Owing to the fact that power and probability distribution in the product form between power function and exponential function, which cannot be derived via equal probability hypothesis, can be derived by the aid of maximal entropy principle, it also can be concluded that maximal entropy principle is a basic principle which embodies concepts more extensively and reveals basic principles on motion laws of objects more fundamentally. At the same time, this principle also reveals the intrinsic link between Nature and different objects in human society and principles complied by all.
Steady-state distributions of probability fluxes on complex networks
NASA Astrophysics Data System (ADS)
Chełminiak, Przemysław; Kurzyński, Michał
2017-02-01
We consider a simple model of the Markovian stochastic dynamics on complex networks to examine the statistical properties of the probability fluxes. The additional transition, called hereafter a gate, powered by the external constant force breaks a detailed balance in the network. We argue, using a theoretical approach and numerical simulations, that the stationary distributions of the probability fluxes emergent under such conditions converge to the Gaussian distribution. By virtue of the stationary fluctuation theorem, its standard deviation depends directly on the square root of the mean flux. In turn, the nonlinear relation between the mean flux and the external force, which provides the key result of the present study, allows us to calculate the two parameters that entirely characterize the Gaussian distribution of the probability fluxes both close to as well as far from the equilibrium state. Also, the other effects that modify these parameters, such as the addition of shortcuts to the tree-like network, the extension and configuration of the gate and a change in the network size studied by means of computer simulations are widely discussed in terms of the rigorous theoretical predictions.
Time-dependent landslide probability mapping
Campbell, Russell H.; Bernknopf, Richard L.; ,
1993-01-01
Case studies where time of failure is known for rainfall-triggered debris flows can be used to estimate the parameters of a hazard model in which the probability of failure is a function of time. As an example, a time-dependent function for the conditional probability of a soil slip is estimated from independent variables representing hillside morphology, approximations of material properties, and the duration and rate of rainfall. If probabilities are calculated in a GIS (geomorphic information system ) environment, the spatial distribution of the result for any given hour can be displayed on a map. Although the probability levels in this example are uncalibrated, the method offers a potential for evaluating different physical models and different earth-science variables by comparing the map distribution of predicted probabilities with inventory maps for different areas and different storms. If linked with spatial and temporal socio-economic variables, this method could be used for short-term risk assessment.
Generalized Wishart Mixtures for Unsupervised Classification of PolSAR Data
NASA Astrophysics Data System (ADS)
Li, Lan; Chen, Erxue; Li, Zengyuan
2013-01-01
This paper presents an unsupervised clustering algorithm based upon the expectation maximization (EM) algorithm for finite mixture modelling, using the complex wishart probability density function (PDF) for the probabilities. The mixture model enables to consider heterogeneous thematic classes which could not be better fitted by the unimodal wishart distribution. In order to make it fast and robust to calculate, we use the recently proposed generalized gamma distribution (GΓD) for the single polarization intensity data to make the initial partition. Then we use the wishart probability density function for the corresponding sample covariance matrix to calculate the posterior class probabilities for each pixel. The posterior class probabilities are used for the prior probability estimates of each class and weights for all class parameter updates. The proposed method is evaluated and compared with the wishart H-Alpha-A classification. Preliminary results show that the proposed method has better performance.
Oil spill contamination probability in the southeastern Levantine basin.
Goldman, Ron; Biton, Eli; Brokovich, Eran; Kark, Salit; Levin, Noam
2015-02-15
Recent gas discoveries in the eastern Mediterranean Sea led to multiple operations with substantial economic interest, and with them there is a risk of oil spills and their potential environmental impacts. To examine the potential spatial distribution of this threat, we created seasonal maps of the probability of oil spill pollution reaching an area in the Israeli coastal and exclusive economic zones, given knowledge of its initial sources. We performed simulations of virtual oil spills using realistic atmospheric and oceanic conditions. The resulting maps show dominance of the alongshore northerly current, which causes the high probability areas to be stretched parallel to the coast, increasing contamination probability downstream of source points. The seasonal westerly wind forcing determines how wide the high probability areas are, and may also restrict these to a small coastal region near source points. Seasonal variability in probability distribution, oil state, and pollution time is also discussed. Copyright © 2014 Elsevier Ltd. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, Chao ..; Singh, Vijay P.; Mishra, Ashok K.
2013-02-06
This paper presents an improved brivariate mixed distribution, which is capable of modeling the dependence of daily rainfall from two distinct sources (e.g., rainfall from two stations, two consecutive days, or two instruments such as satellite and rain gauge). The distribution couples an existing framework for building a bivariate mixed distribution, the theory of copulae and a hybrid marginal distribution. Contributions of the improved distribution are twofold. One is the appropriate selection of the bivariate dependence structure from a wider admissible choice (10 candidate copula families). The other is the introduction of a marginal distribution capable of better representing lowmore » to moderate values as well as extremes of daily rainfall. Among several applications of the improved distribution, particularly presented here is its utility for single-site daily rainfall simulation. Rather than simulating rainfall occurrences and amounts separately, the developed generator unifies the two processes by generalizing daily rainfall as a Markov process with autocorrelation described by the improved bivariate mixed distribution. The generator is first tested on a sample station in Texas. Results reveal that the simulated and observed sequences are in good agreement with respect to essential characteristics. Then, extensive simulation experiments are carried out to compare the developed generator with three other alternative models: the conventional two-state Markov chain generator, the transition probability matrix model and the semi-parametric Markov chain model with kernel density estimation for rainfall amounts. Analyses establish that overall the developed generator is capable of reproducing characteristics of historical extreme rainfall events and is apt at extrapolating rare values beyond the upper range of available observed data. Moreover, it automatically captures the persistence of rainfall amounts on consecutive wet days in a relatively natural and easy way. Another interesting observation is that the recognized ‘overdispersion’ problem in daily rainfall simulation ascribes more to the loss of rainfall extremes than the under-representation of first-order persistence. The developed generator appears to be a sound option for daily rainfall simulation, especially in particular hydrologic planning situations when rare rainfall events are of great importance.« less
Bellin, Alberto; Tonina, Daniele
2007-10-30
Available models of solute transport in heterogeneous formations lack in providing complete characterization of the predicted concentration. This is a serious drawback especially in risk analysis where confidence intervals and probability of exceeding threshold values are required. Our contribution to fill this gap of knowledge is a probability distribution model for the local concentration of conservative tracers migrating in heterogeneous aquifers. Our model accounts for dilution, mechanical mixing within the sampling volume and spreading due to formation heterogeneity. It is developed by modeling local concentration dynamics with an Ito Stochastic Differential Equation (SDE) that under the hypothesis of statistical stationarity leads to the Beta probability distribution function (pdf) for the solute concentration. This model shows large flexibility in capturing the smoothing effect of the sampling volume and the associated reduction of the probability of exceeding large concentrations. Furthermore, it is fully characterized by the first two moments of the solute concentration, and these are the same pieces of information required for standard geostatistical techniques employing Normal or Log-Normal distributions. Additionally, we show that in the absence of pore-scale dispersion and for point concentrations the pdf model converges to the binary distribution of [Dagan, G., 1982. Stochastic modeling of groundwater flow by unconditional and conditional probabilities, 2, The solute transport. Water Resour. Res. 18 (4), 835-848.], while it approaches the Normal distribution for sampling volumes much larger than the characteristic scale of the aquifer heterogeneity. Furthermore, we demonstrate that the same model with the spatial moments replacing the statistical moments can be applied to estimate the proportion of the plume volume where solute concentrations are above or below critical thresholds. Application of this model to point and vertically averaged bromide concentrations from the first Cape Cod tracer test and to a set of numerical simulations confirms the above findings and for the first time it shows the superiority of the Beta model to both Normal and Log-Normal models in interpreting field data. Furthermore, we show that assuming a-priori that local concentrations are normally or log-normally distributed may result in a severe underestimate of the probability of exceeding large concentrations.
Driven by Power? Probe Question and Presentation Format Effects on Causal Judgment
ERIC Educational Resources Information Center
Perales, Jose C.; Shanks, David R.
2008-01-01
It has been proposed that causal power (defined as the probability with which a candidate cause would produce an effect in the absence of any other background causes) can be intuitively computed from cause-effect covariation information. Estimation of power is assumed to require a special type of counterfactual probe question, worded to remove…
Rasch Measurement and Item Banking: Theory and Practice.
ERIC Educational Resources Information Center
Nakamura, Yuji
The Rasch Model is an item response theory, one parameter model developed that states that the probability of a correct response on a test is a function of the difficulty of the item and the ability of the candidate. Item banking is useful for language testing. The Rasch Model provides estimates of item difficulties that are meaningful,…
NASA Astrophysics Data System (ADS)
Ahn, Hyunjun; Jung, Younghun; Om, Ju-Seong; Heo, Jun-Haeng
2014-05-01
It is very important to select the probability distribution in Statistical hydrology. Goodness of fit test is a statistical method that selects an appropriate probability model for a given data. The probability plot correlation coefficient (PPCC) test as one of the goodness of fit tests was originally developed for normal distribution. Since then, this test has been widely applied to other probability models. The PPCC test is known as one of the best goodness of fit test because it shows higher rejection powers among them. In this study, we focus on the PPCC tests for the GEV distribution which is widely used in the world. For the GEV model, several plotting position formulas are suggested. However, the PPCC statistics are derived only for the plotting position formulas (Goel and De, In-na and Nguyen, and Kim et al.) in which the skewness coefficient (or shape parameter) are included. And then the regression equations are derived as a function of the shape parameter and sample size for a given significance level. In addition, the rejection powers of these formulas are compared using Monte-Carlo simulation. Keywords: Goodness-of-fit test, Probability plot correlation coefficient test, Plotting position, Monte-Carlo Simulation ACKNOWLEDGEMENTS This research was supported by a grant 'Establishing Active Disaster Management System of Flood Control Structures by using 3D BIM Technique' [NEMA-12-NH-57] from the Natural Hazard Mitigation Research Group, National Emergency Management Agency of Korea.
NASA Astrophysics Data System (ADS)
Hayakawa, Hisashi; Iwahashi, Kiyomi; Tamazawa, Harufumi; Ebihara, Yusuke; Kawamura, Akito Davis; Isobe, Hiroaki; Namiki, Katsuko; Shibata, Kazunari
2017-12-01
We present the results of the surveys on sunspots and auroral candidates in Rikkokushi, Japanese official histories from the early 7th century to 887, to review the solar and auroral activities. In total, we found one sunspot record and 13 auroral candidates in Rikkokushi. We then examine the records of the sunspots and auroral candidates, compare the auroral candidates with the lunar phase to estimate their reliability, and compare the records of the sunspots and auroral candidates with the contemporary total solar irradiance reconstructed from radioisotope data. We also identify the locations of the observational sites to review possible equatorward expansion of the auroral oval. These discussions suggest a major gap in auroral candidates from the late 7th to early 9th centuries, which includes the candidate of the grand minimum reconstructed from the radioisotope data, a similar tendency as the distributions of sunspot records in contemporary China, and a relatively high magnetic latitude of observational sites with a higher potential for observing aurorae more frequently than at present.
Baron, Jean-Claude; Farid, Karim; Dolan, Eamon; Turc, Guillaume; Marrapu, Siva T; O'Brien, Eoin; Aigbirhio, Franklin I; Fryer, Tim D; Menon, David K; Warburton, Elizabeth A; Hong, Young T
2014-01-01
By detecting β-amyloid (Aβ) in the wall of cortical arterioles, amyloid positron emission tomography (PET) imaging might help diagnose cerebral amyloid angiopathy (CAA) in patients with lobar intracerebral hemorrhage (l-ICH). No previous study has directly assessed the diagnostic value of 11C-Pittsburgh compound B (PiB) PET in probable CAA-related l-ICH against healthy controls (HCs). 11C-PiB-PET and magnetic resonance imaging (MRI) including T2* were obtained in 11 nondemented patients fulfilling the Boston criteria for probable CAA-related symptomatic l-ICH (sl-ICH) and 20 HCs without cognitive complaints or impairment. After optimal spatial normalization, cerebral spinal fluid (CSF)-corrected PiB distribution volume ratios (DVRs) were obtained. There was no significant difference in whole cortex or regional DVRs between CAA patients and age-matched HCs. The whole cortex DVR was above the 95% confidence limit in 4/9 HCs and 10/11 CAA patients (sensitivity=91%, specificity=55%). Region/frontal or occipital ratios did not have better discriminative value. Similar but less accurate results were found using visual analysis. In patients with sl-ICH, 11C-PiB-PET has low specificity for CAA due to the frequent occurrence of high 11C-PiB uptake in the healthy elderly reflecting incipient Alzheimer's disease (AD), which might also be present in suspected CAA. However, a negative PiB scan rules out CAA with excellent sensitivity, which has clinical implications for prognostication and selection of candidates for drug trials. PMID:24619277
OPTN/SRTR 2013 Annual Data Report: lung.
Valapour, M; Skeans, M A; Heubner, B M; Smith, J M; Hertz, M I; Edwards, L B; Cherikh, W S; Callahan, E R; Snyder, J J; Israni, A K; Kasiske, B L
2015-01-01
Lungs are allocated to adult and adolescent transplant candidates (aged ⩾ 12 years) on the basis of age, geography, blood type compatibility, and the lung allocation score (LAS), which reflects risk of waitlist mortality and probability of posttransplant survival. In 2013, the most adult candidates, 2394, of any year were added to the list. Overall median waiting time for candidates listed in 2013 was 4.0 months. The preferred procedure remained bilateral lung transplant, representing approximately 70% of lung transplants in 2013. Measures of short-term and longterm survival have plateaued since the implementation of the LAS in 2005. The number of new child candidates (aged 0-11 years) added to the lung transplant waiting list increased to 39 in 2013. A total of 28 lung transplants were performed in child recipients, 3 for ages younger than 1 year, 9 for ages 1 to 5 years, and 16 for ages 6 to 11 years. The diagnosis of pulmonary hypertension was associated with higher survival rates than cystic fibrosis or other diagnosis (pulmonary fibrosis, bronchiolitis obliterans, bronchopulmonary dysplasia). For child candidates, infection was the leading cause of death in year 1 posttransplant and graft failure in years 2 to 5. © Copyright 2015 The American Society of Transplantation and the American Society of Transplant Surgeons.
Live vaccines for human metapneumovirus designed by reverse genetics.
Buchholz, Ursula J; Nagashima, Kunio; Murphy, Brian R; Collins, Peter L
2006-10-01
Human metapneumovirus (HMPV) was first described in 2001 and has quickly become recognized as an important cause of respiratory tract disease worldwide, especially in the pediatric population. A vaccine against HMPV is required to prevent severe disease associated with infection in infancy. The primary strategy is to develop a live-attenuated virus for intranasal immunization, which is particularly well suited against a respiratory virus. Reverse genetics provides a means of developing highly characterized 'designer' attenuated vaccine candidates. To date, several promising vaccine candidates have been developed, each using a different mode of attenuation. One candidate involves deletion of the G glycoprotein, providing attenuation that is probably based on reduced efficiency of attachment. A second candidate involves deletion of the M2-2 protein, which participates in regulating RNA synthesis and whose deletion has the advantageous property of upregulating transcription and increasing antigen synthesis. A third candidate involves replacing the P protein gene of HMPV with its counterpart from the related avian metapneumovirus, thereby introducing attenuation owing to its chimeric nature and host range restriction. Another live vaccine strategy involves using an attenuated parainfluenza virus as a vector to express HMPV protective antigens, providing a bivalent pediatric vaccine. Additional modifications to provide improved vaccines will also be discussed.
DeVore, Matthew S; Gull, Stephen F; Johnson, Carey K
2012-04-05
We describe a method for analysis of single-molecule Förster resonance energy transfer (FRET) burst measurements using classic maximum entropy. Classic maximum entropy determines the Bayesian inference for the joint probability describing the total fluorescence photons and the apparent FRET efficiency. The method was tested with simulated data and then with DNA labeled with fluorescent dyes. The most probable joint distribution can be marginalized to obtain both the overall distribution of fluorescence photons and the apparent FRET efficiency distribution. This method proves to be ideal for determining the distance distribution of FRET-labeled biomolecules, and it successfully predicts the shape of the recovered distributions.
SS 383: A NEW S-TYPE YELLOW SYMBIOTIC STAR?
DOE Office of Scientific and Technical Information (OSTI.GOV)
Baella, N. O.; Pereira, C. B.; Miranda, L. F.
Symbiotic stars are key objects in understanding the formation and evolution of interacting binary systems, and are probably the progenitors of Type Ia supernovae. However, the number of known symbiotic stars is much lower than predicted. We aim to search for new symbiotic stars, with particular emphasis on the S-type yellow symbiotic stars, in order to determine their total population, evolutionary timescales, and physical properties. The Two Micron All Sky Survey (2MASS) (J – H) versus (H – K {sub s}) color-color diagram has been previously used to identify new symbiotic star candidates and show that yellow symbiotics are locatedmore » in a particular region of that diagram. Candidate symbiotic stars are selected on the basis of their locus in the 2MASS (J – H) versus (H – K {sub s}) diagram and the presence of Hα line emission in the Stephenson and Sanduleak Hα survey. This diagram separates S-type yellow symbiotic stars from the rest of the S-type symbiotic stars, allowing us to select candidate yellow symbiotics. To establish the true nature of the candidates, intermediate-resolution spectroscopy is obtained. We have identified the Hα emission line source SS 383 as an S-type yellow symbiotic candidate by its position in the 2MASS color-color diagram. The optical spectrum of SS 383 shows Balmer, He I, He II, and [O III] emission lines, in combination with TiO absorption bands that confirm its symbiotic nature. The derived electron density (≅10{sup 8-9} cm{sup –3}), He I emission line intensity ratios, and position in the [O III] λ5007/Hβ versus [O III] λ4363/Hγ diagram indicate that SS 383 is an S-type symbiotic star, with a probable spectral type of K7-M0 deduced for its cool component based on TiO indices. The spectral type and the position of SS 383 (corrected for reddening) in the 2MASS color-color diagram strongly suggest that SS 383 is an S-type yellow symbiotic. Our result points out that the 2MASS color-color diagram is a powerful tool in identifying new S-type yellow symbiotics.« less
Improved first-order uncertainty method for water-quality modeling
Melching, C.S.; Anmangandla, S.
1992-01-01
Uncertainties are unavoidable in water-quality modeling and subsequent management decisions. Monte Carlo simulation and first-order uncertainty analysis (involving linearization at central values of the uncertain variables) have been frequently used to estimate probability distributions for water-quality model output due to their simplicity. Each method has its drawbacks: Monte Carlo simulation's is mainly computational time; and first-order analysis are mainly questions of accuracy and representativeness, especially for nonlinear systems and extreme conditions. An improved (advanced) first-order method is presented, where the linearization point varies to match the output level whose exceedance probability is sought. The advanced first-order method is tested on the Streeter-Phelps equation to estimate the probability distribution of critical dissolved-oxygen deficit and critical dissolved oxygen using two hypothetical examples from the literature. The advanced first-order method provides a close approximation of the exceedance probability for the Streeter-Phelps model output estimated by Monte Carlo simulation using less computer time - by two orders of magnitude - regardless of the probability distributions assumed for the uncertain model parameters.
Measures for a multidimensional multiverse
NASA Astrophysics Data System (ADS)
Chung, Hyeyoun
2015-04-01
We explore the phenomenological implications of generalizing the causal patch and fat geodesic measures to a multidimensional multiverse, where the vacua can have differing numbers of large dimensions. We consider a simple model in which the vacua are nucleated from a D -dimensional parent spacetime through dynamical compactification of the extra dimensions, and compute the geometric contribution to the probability distribution of observations within the multiverse for each measure. We then study how the shape of this probability distribution depends on the time scales for the existence of observers, for vacuum domination, and for curvature domination (tobs,tΛ , and tc, respectively.) In this work we restrict ourselves to bubbles with positive cosmological constant, Λ . We find that in the case of the causal patch cutoff, when the bubble universes have p +1 large spatial dimensions with p ≥2 , the shape of the probability distribution is such that we obtain the coincidence of time scales tobs˜tΛ˜tc . Moreover, the size of the cosmological constant is related to the size of the landscape. However, the exact shape of the probability distribution is different in the case p =2 , compared to p ≥3 . In the case of the fat geodesic measure, the result is even more robust: the shape of the probability distribution is the same for all p ≥2 , and we once again obtain the coincidence tobs˜tΛ˜tc . These results require only very mild conditions on the prior probability of the distribution of vacua in the landscape. Our work shows that the observed double coincidence of time scales is a robust prediction even when the multiverse is generalized to be multidimensional; that this coincidence is not a consequence of our particular Universe being (3 +1 )-dimensional; and that this observable cannot be used to preferentially select one measure over another in a multidimensional multiverse.
Shape of growth-rate distribution determines the type of Non-Gibrat’s Property
NASA Astrophysics Data System (ADS)
Ishikawa, Atushi; Fujimoto, Shouji; Mizuno, Takayuki
2011-11-01
In this study, the authors examine exhaustive business data on Japanese firms, which cover nearly all companies in the mid- and large-scale ranges in terms of firm size, to reach several key findings on profits/sales distribution and business growth trends. Here, profits denote net profits. First, detailed balance is observed not only in profits data but also in sales data. Furthermore, the growth-rate distribution of sales has wider tails than the linear growth-rate distribution of profits in log-log scale. On the one hand, in the mid-scale range of profits, the probability of positive growth decreases and the probability of negative growth increases symmetrically as the initial value increases. This is called Non-Gibrat’s First Property. On the other hand, in the mid-scale range of sales, the probability of positive growth decreases as the initial value increases, while the probability of negative growth hardly changes. This is called Non-Gibrat’s Second Property. Under detailed balance, Non-Gibrat’s First and Second Properties are analytically derived from the linear and quadratic growth-rate distributions in log-log scale, respectively. In both cases, the log-normal distribution is inferred from Non-Gibrat’s Properties and detailed balance. These analytic results are verified by empirical data. Consequently, this clarifies the notion that the difference in shapes between growth-rate distributions of sales and profits is closely related to the difference between the two Non-Gibrat’s Properties in the mid-scale range.
Estimation of distribution overlap of urn models.
Hampton, Jerrad; Lladser, Manuel E
2012-01-01
A classical problem in statistics is estimating the expected coverage of a sample, which has had applications in gene expression, microbial ecology, optimization, and even numismatics. Here we consider a related extension of this problem to random samples of two discrete distributions. Specifically, we estimate what we call the dissimilarity probability of a sample, i.e., the probability of a draw from one distribution not being observed in [Formula: see text] draws from another distribution. We show our estimator of dissimilarity to be a [Formula: see text]-statistic and a uniformly minimum variance unbiased estimator of dissimilarity over the largest appropriate range of [Formula: see text]. Furthermore, despite the non-Markovian nature of our estimator when applied sequentially over [Formula: see text], we show it converges uniformly in probability to the dissimilarity parameter, and we present criteria when it is approximately normally distributed and admits a consistent jackknife estimator of its variance. As proof of concept, we analyze V35 16S rRNA data to discern between various microbial environments. Other potential applications concern any situation where dissimilarity of two discrete distributions may be of interest. For instance, in SELEX experiments, each urn could represent a random RNA pool and each draw a possible solution to a particular binding site problem over that pool. The dissimilarity of these pools is then related to the probability of finding binding site solutions in one pool that are absent in the other.
Diffusion of active chiral particles
NASA Astrophysics Data System (ADS)
Sevilla, Francisco J.
2016-12-01
The diffusion of chiral active Brownian particles in three-dimensional space is studied analytically, by consideration of the corresponding Fokker-Planck equation for the probability density of finding a particle at position x and moving along the direction v ̂ at time t , and numerically, by the use of Langevin dynamics simulations. The analysis is focused on the marginal probability density of finding a particle at a given location and at a given time (independently of its direction of motion), which is found from an infinite hierarchy of differential-recurrence relations for the coefficients that appear in the multipole expansion of the probability distribution, which contains the whole kinematic information. This approach allows the explicit calculation of the time dependence of the mean-squared displacement and the time dependence of the kurtosis of the marginal probability distribution, quantities from which the effective diffusion coefficient and the "shape" of the positions distribution are examined. Oscillations between two characteristic values were found in the time evolution of the kurtosis, namely, between the value that corresponds to a Gaussian and the one that corresponds to a distribution of spherical shell shape. In the case of an ensemble of particles, each one rotating around a uniformly distributed random axis, evidence is found of the so-called effect "anomalous, yet Brownian, diffusion," for which particles follow a non-Gaussian distribution for the positions yet the mean-squared displacement is a linear function of time.
A Search Model for Imperfectly Detected Targets
NASA Technical Reports Server (NTRS)
Ahumada, Albert
2012-01-01
Under the assumptions that 1) the search region can be divided up into N non-overlapping sub-regions that are searched sequentially, 2) the probability of detection is unity if a sub-region is selected, and 3) no information is available to guide the search, there are two extreme case models. The search can be done perfectly, leading to a uniform distribution over the number of searches required, or the search can be done with no memory, leading to a geometric distribution for the number of searches required with a success probability of 1/N. If the probability of detection P is less than unity, but the search is done otherwise perfectly, the searcher will have to search the N regions repeatedly until detection occurs. The number of searches is thus the sum two random variables. One is N times the number of full searches (a geometric distribution with success probability P) and the other is the uniform distribution over the integers 1 to N. The first three moments of this distribution were computed, giving the mean, standard deviation, and the kurtosis of the distribution as a function of the two parameters. The model was fit to the data presented last year (Ahumada, Billington, & Kaiwi, 2 required to find a single pixel target on a simulated horizon. The model gave a good fit to the three moments for all three observers.
Optimizing probability of detection point estimate demonstration
NASA Astrophysics Data System (ADS)
Koshti, Ajay M.
2017-04-01
The paper provides discussion on optimizing probability of detection (POD) demonstration experiments using point estimate method. The optimization is performed to provide acceptable value for probability of passing demonstration (PPD) and achieving acceptable value for probability of false (POF) calls while keeping the flaw sizes in the set as small as possible. POD Point estimate method is used by NASA for qualifying special NDE procedures. The point estimate method uses binomial distribution for probability density. Normally, a set of 29 flaws of same size within some tolerance are used in the demonstration. Traditionally largest flaw size in the set is considered to be a conservative estimate of the flaw size with minimum 90% probability and 95% confidence. The flaw size is denoted as α90/95PE. The paper investigates relationship between range of flaw sizes in relation to α90, i.e. 90% probability flaw size, to provide a desired PPD. The range of flaw sizes is expressed as a proportion of the standard deviation of the probability density distribution. Difference between median or average of the 29 flaws and α90 is also expressed as a proportion of standard deviation of the probability density distribution. In general, it is concluded that, if probability of detection increases with flaw size, average of 29 flaw sizes would always be larger than or equal to α90 and is an acceptable measure of α90/95PE. If NDE technique has sufficient sensitivity and signal-to-noise ratio, then the 29 flaw-set can be optimized to meet requirements of minimum required PPD, maximum allowable POF, requirements on flaw size tolerance about mean flaw size and flaw size detectability requirements. The paper provides procedure for optimizing flaw sizes in the point estimate demonstration flaw-set.
Theoretical cratering rates on Ida, Mathilde, Eros and Gaspra
NASA Astrophysics Data System (ADS)
Jeffers, S. V.; Asher, D. J.; Bailey, M. E.
2002-11-01
We investigate the main influences on crater size distributions, by deriving results for the four example target objects, (951) Gaspra, (243) Ida, (253) Mathilde and (433) Eros. The dynamical history of each of these asteroids is modelled using the MERCURY (Chambers 1999) numerical integrator. The use of an efficient, Öpik-type, collision code enables the calculation of a velocity histogram and the probability of impact. This when combined with a crater scaling law and an impactor size distribution, through a Monte Carlo method, results in a crater size distribution. The resulting crater probability distributions are in good agreement with observed crater distributions on these asteroids.
Velocity distributions among colliding asteroids
NASA Technical Reports Server (NTRS)
Bottke, William F., Jr.; Nolan, Michael C.; Greenberg, Richard; Kolvoord, Robert A.
1994-01-01
The probability distribution for impact velocities between two given asteroids is wide, non-Gaussian, and often contains spikes according to our new method of analysis in which each possible orbital geometry for collision is weighted according to its probability. An average value would give a good representation only if the distribution were smooth and narrow. Therefore, the complete velocity distribution we obtain for various asteroid populations differs significantly from published histograms of average velocities. For all pairs among the 682 asteroids in the main-belt with D greater than 50 km, we find that our computed velocity distribution is much wider than previously computed histograms of average velocities. In this case, the most probable impact velocity is approximately 4.4 km/sec, compared with the mean impact velocity of 5.3 km/sec. For cases of a single asteroid (e.g., Gaspra or Ida) relative to an impacting population, the distribution we find yields lower velocities than previously reported by others. The width of these velocity distributions implies that mean impact velocities must be used with caution when calculating asteroid collisional lifetimes or crater-size distributions. Since the most probable impact velocities are lower than the mean, disruption events may occur less frequently than previously estimated. However, this disruption rate may be balanced somewhat by an apparent increase in the frequency of high-velocity impacts between asteroids. These results have implications for issues such as asteroidal disruption rates, the amount/type of impact ejecta available for meteoritical delivery to the Earth, and the geology and evolution of specific asteroids like Gaspra.
The Pan-Pacific Planet Search. VII. The Most Eccentric Planet Orbiting a Giant Star
NASA Astrophysics Data System (ADS)
Wittenmyer, Robert A.; Jones, M. I.; Horner, Jonathan; Kane, Stephen R.; Marshall, J. P.; Mustill, A. J.; Jenkins, J. S.; Pena Rojas, P. A.; Zhao, Jinglin; Villaver, Eva; Butler, R. P.; Clark, Jake
2017-12-01
Radial velocity observations from three instruments reveal the presence of a 4 M Jup planet candidate orbiting the K giant HD 76920. HD 76920b has an orbital eccentricity of 0.856 ± 0.009, making it the most eccentric planet known to orbit an evolved star. There is no indication that HD 76920 has an unseen binary companion, suggesting a scattering event rather than Kozai oscillations as a probable culprit for the observed eccentricity. The candidate planet currently approaches to about four stellar radii from its host star, and is predicted to be engulfed on a ∼100 Myr timescale due to the combined effects of stellar evolution and tidal interactions.
SIMP J2154–1055: A NEW LOW-GRAVITY L4β BROWN DWARF CANDIDATE MEMBER OF THE ARGUS ASSOCIATION
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gagné, Jonathan; Lafrenière, David; Doyon, René
We present SIMP J21543454–1055308, a new L4β brown dwarf identified in the SIMP survey that displays signs of low gravity in its near-infrared spectrum. Using BANYAN II, we show that it is a candidate member of the Argus association, albeit with a 21% probability that it is a contaminant from the field. Measurements of radial velocity and parallax will be needed to verify its membership. If it is a member of Argus (age 30-50 Myr), then this object would have a planetary mass of 10 ± 0.5 M {sub Jup}.
The Probability Distribution for a Biased Spinner
ERIC Educational Resources Information Center
Foster, Colin
2012-01-01
This article advocates biased spinners as an engaging context for statistics students. Calculating the probability of a biased spinner landing on a particular side makes valuable connections between probability and other areas of mathematics. (Contains 2 figures and 1 table.)
NASA Astrophysics Data System (ADS)
Gao, Haixia; Li, Ting; Xiao, Changming
2016-05-01
When a simple system is in its nonequilibrium state, it will shift to its equilibrium state. Obviously, in this process, there are a series of nonequilibrium states. With the assistance of Bayesian statistics and hyperensemble, a probable probability distribution of these nonequilibrium states can be determined by maximizing the hyperensemble entropy. It is known that the largest probability is the equilibrium state, and the far a nonequilibrium state is away from the equilibrium one, the smaller the probability will be, and the same conclusion can also be obtained in the multi-state space. Furthermore, if the probability stands for the relative time the corresponding nonequilibrium state can stay, then the velocity of a nonequilibrium state returning back to its equilibrium can also be determined through the reciprocal of the derivative of this probability. It tells us that the far away the state from the equilibrium is, the faster the returning velocity will be; if the system is near to its equilibrium state, the velocity will tend to be smaller and smaller, and finally tends to 0 when it gets the equilibrium state.
Study on probability distributions for evolution in modified extremal optimization
NASA Astrophysics Data System (ADS)
Zeng, Guo-Qiang; Lu, Yong-Zai; Mao, Wei-Jie; Chu, Jian
2010-05-01
It is widely believed that the power-law is a proper probability distribution being effectively applied for evolution in τ-EO (extremal optimization), a general-purpose stochastic local-search approach inspired by self-organized criticality, and its applications in some NP-hard problems, e.g., graph partitioning, graph coloring, spin glass, etc. In this study, we discover that the exponential distributions or hybrid ones (e.g., power-laws with exponential cutoff) being popularly used in the research of network sciences may replace the original power-laws in a modified τ-EO method called self-organized algorithm (SOA), and provide better performances than other statistical physics oriented methods, such as simulated annealing, τ-EO and SOA etc., from the experimental results on random Euclidean traveling salesman problems (TSP) and non-uniform instances. From the perspective of optimization, our results appear to demonstrate that the power-law is not the only proper probability distribution for evolution in EO-similar methods at least for TSP, the exponential and hybrid distributions may be other choices.
Rapidly assessing the probability of exceptionally high natural hazard losses
NASA Astrophysics Data System (ADS)
Gollini, Isabella; Rougier, Jonathan
2014-05-01
One of the objectives in catastrophe modeling is to assess the probability distribution of losses for a specified period, such as a year. From the point of view of an insurance company, the whole of the loss distribution is interesting, and valuable in determining insurance premiums. But the shape of the righthand tail is critical, because it impinges on the solvency of the company. A simple measure of the risk of insolvency is the probability that the annual loss will exceed the company's current operating capital. Imposing an upper limit on this probability is one of the objectives of the EU Solvency II directive. If a probabilistic model is supplied for the loss process, then this tail probability can be computed, either directly, or by simulation. This can be a lengthy calculation for complex losses. Given the inevitably subjective nature of quantifying loss distributions, computational resources might be better used in a sensitivity analysis. This requires either a quick approximation to the tail probability or an upper bound on the probability, ideally a tight one. We present several different bounds, all of which can be computed nearly instantly from a very general event loss table. We provide a numerical illustration, and discuss the conditions under which the bound is tight. Although we consider the perspective of insurance and reinsurance companies, exactly the same issues concern the risk manager, who is typically very sensitive to large losses.
Dopkins, Stephen; Varner, Kaitlin; Hoyer, Darin
2017-10-01
In word recognition semantic priming of test words increased the false-alarm rate and the mean of confidence ratings to lures. Such priming also increased the standard deviation of confidence ratings to lures and the slope of the z-ROC function, suggesting that the priming increased the standard deviation of the lure evidence distribution. The Unequal Variance Signal Detection (UVSD) model interpreted the priming as increasing the standard deviation of the lure evidence distribution. Without additional parameters the Dual Process Signal Detection (DPSD) model could only accommodate the results by fitting the data for related and unrelated primes separately, interpreting the priming, implausibly, as decreasing the probability of target recollection (DPSD). With an additional parameter, for the probability of false (lure) recollection the model could fit the data for related and unrelated primes together, interpreting the priming as increasing the probability of false recollection. These results suggest that DPSD estimates of target recollection probability will decrease with increases in the lure confidence/evidence standard deviation unless a parameter is included for false recollection. Unfortunately the size of a given lure confidence/evidence standard deviation relative to other possible lure confidence/evidence standard deviations is often unspecified by context. Hence the model often has no way of estimating false recollection probability and thereby correcting its estimates of target recollection probability.
THE BROWN DWARF KINEMATICS PROJECT (BDKP). IV. RADIAL VELOCITIES OF 85 LATE-M AND L DWARFS WITH MagE
DOE Office of Scientific and Technical Information (OSTI.GOV)
Burgasser, Adam J.; Logsdon, Sarah E.; Gagné, Jonathan
2015-09-15
Radial velocity measurements are presented for 85 late M- and L-type very low-mass stars and brown dwarfs obtained with the Magellan Echellette spectrograph. Targets primarily have distances within 20 pc of the Sun, with more distant sources selected for their unusual spectral energy distributions. We achieved precisions of 2–3 km s{sup −1}, and combined these with astrometric and spectrophotometric data to calculate UVW velocities. Most are members of the thin disk of the Galaxy, and velocity dispersions indicate a mean age of 5.2 ± 0.2 Gyr for sources within 20 pc. We find signficantly different kinematic ages between late-M dwarfsmore » (4.0 ± 0.2 Gyr) and L dwarfs (6.5 ± 0.4 Gyr) in our sample that are contrary to predictions from prior simulations. This difference appears to be driven by a dispersed population of unusually blue L dwarfs which may be more prevalent in our local volume-limited sample than in deeper magnitude-limited surveys. The L dwarfs exhibit an asymmetric U velocity distribution with a net inward flow, similar to gradients recently detected in local stellar samples. Simulations incorporating brown dwarf evolution and Galactic orbital dynamics are unable to reproduce the velocity asymmetry, suggesting non-axisymmetric perturbations or two distinct L dwarf populations. We also find the L dwarfs to have a kinematic age-activity correlation similar to more massive stars. We identify several sources with low surface gravities, and two new substellar candidate members of nearby young moving groups: the astrometric binary DENIS J08230313–4912012AB, a low-probability member of the β Pictoris Moving Group; and 2MASS J15104786–2818174, a moderate-probability member of the 30–50 Myr Argus Association.« less
QTL fine mapping with Bayes C(π): a simulation study.
van den Berg, Irene; Fritz, Sébastien; Boichard, Didier
2013-06-19
Accurate QTL mapping is a prerequisite in the search for causative mutations. Bayesian genomic selection models that analyse many markers simultaneously should provide more accurate QTL detection results than single-marker models. Our objectives were to (a) evaluate by simulation the influence of heritability, number of QTL and number of records on the accuracy of QTL mapping with Bayes Cπ and Bayes C; (b) estimate the QTL status (homozygous vs. heterozygous) of the individuals analysed. This study focussed on the ten largest detected QTL, assuming they are candidates for further characterization. Our simulations were based on a true dairy cattle population genotyped for 38,277 phased markers. Some of these markers were considered biallelic QTL and used to generate corresponding phenotypes. Different numbers of records (4387 and 1500), heritability values (0.1, 0.4 and 0.7) and numbers of QTL (10, 100 and 1000) were studied. QTL detection was based on the posterior inclusion probability for individual markers, or on the sum of the posterior inclusion probabilities for consecutive markers, estimated using Bayes C or Bayes Cπ. The QTL status of the individuals was derived from the contrast between the sums of the SNP allelic effects of their chromosomal segments. The proportion of markers with null effect (π) frequently did not reach convergence, leading to poor results for Bayes Cπ in QTL detection. Fixing π led to better results. Detection of the largest QTL was most accurate for medium to high heritability, for low to moderate numbers of QTL, and with a large number of records. The QTL status was accurately inferred when the distribution of the contrast between chromosomal segment effects was bimodal. QTL detection is feasible with Bayes C. For QTL detection, it is recommended to use a large dataset and to focus on highly heritable traits and on the largest QTL. QTL statuses were inferred based on the distribution of the contrast between chromosomal segment effects.
Earth reencounter probabilities for aborted space disposal of hazardous nuclear waste
NASA Technical Reports Server (NTRS)
Friedlander, A. L.; Feingold, H.
1977-01-01
A quantitative assessment is made of the long-term risk of earth reencounter and reentry associated with aborted disposal of hazardous material in the space environment. Numerical results are presented for 10 candidate disposal options covering a broad spectrum of disposal destinations and deployment propulsion systems. Based on representative models of system failure, the probability that a single payload will return and collide with earth within a period of 250,000 years is found to lie in the range .0002-.006. Proportionately smaller risk attaches to shorter time intervals. Risk-critical factors related to trajectory geometry and system reliability are identified as possible mechanisms of hazard reduction.
NASA Technical Reports Server (NTRS)
Wood, B. J.; Ablow, C. M.; Wise, H.
1973-01-01
For a number of candidate materials of construction for the dual air density explorer satellites the rate of oxygen atom loss by adsorption, surface reaction, and recombination was determined as a function of surface and temperature. Plain aluminum and anodized aluminum surfaces exhibit a collisional atom loss probability alpha .01 in the temperature range 140 - 360 K, and an initial sticking probability. For SiO coated aluminum in the same temperature range, alpha .001 and So .001. Atom-loss on gold is relatively rapid alpha .01. The So for gold varies between 0.25 and unity in the temperature range 360 - 140 K.
Occurrence probability of structured motifs in random sequences.
Robin, S; Daudin, J-J; Richard, H; Sagot, M-F; Schbath, S
2002-01-01
The problem of extracting from a set of nucleic acid sequences motifs which may have biological function is more and more important. In this paper, we are interested in particular motifs that may be implicated in the transcription process. These motifs, called structured motifs, are composed of two ordered parts separated by a variable distance and allowing for substitutions. In order to assess their statistical significance, we propose approximations of the probability of occurrences of such a structured motif in a given sequence. An application of our method to evaluate candidate promoters in E. coli and B. subtilis is presented. Simulations show the goodness of the approximations.
Probability distributions for Markov chain based quantum walks
NASA Astrophysics Data System (ADS)
Balu, Radhakrishnan; Liu, Chaobin; Venegas-Andraca, Salvador E.
2018-01-01
We analyze the probability distributions of the quantum walks induced from Markov chains by Szegedy (2004). The first part of this paper is devoted to the quantum walks induced from finite state Markov chains. It is shown that the probability distribution on the states of the underlying Markov chain is always convergent in the Cesaro sense. In particular, we deduce that the limiting distribution is uniform if the transition matrix is symmetric. In the case of a non-symmetric Markov chain, we exemplify that the limiting distribution of the quantum walk is not necessarily identical with the stationary distribution of the underlying irreducible Markov chain. The Szegedy scheme can be extended to infinite state Markov chains (random walks). In the second part, we formulate the quantum walk induced from a lazy random walk on the line. We then obtain the weak limit of the quantum walk. It is noted that the current quantum walk appears to spread faster than its counterpart-quantum walk on the line driven by the Grover coin discussed in literature. The paper closes with an outlook on possible future directions.
Digital simulation of an arbitrary stationary stochastic process by spectral representation.
Yura, Harold T; Hanson, Steen G
2011-04-01
In this paper we present a straightforward, efficient, and computationally fast method for creating a large number of discrete samples with an arbitrary given probability density function and a specified spectral content. The method relies on initially transforming a white noise sample set of random Gaussian distributed numbers into a corresponding set with the desired spectral distribution, after which this colored Gaussian probability distribution is transformed via an inverse transform into the desired probability distribution. In contrast to previous work, where the analyses were limited to auto regressive and or iterative techniques to obtain satisfactory results, we find that a single application of the inverse transform method yields satisfactory results for a wide class of arbitrary probability distributions. Although a single application of the inverse transform technique does not conserve the power spectra exactly, it yields highly accurate numerical results for a wide range of probability distributions and target power spectra that are sufficient for system simulation purposes and can thus be regarded as an accurate engineering approximation, which can be used for wide range of practical applications. A sufficiency condition is presented regarding the range of parameter values where a single application of the inverse transform method yields satisfactory agreement between the simulated and target power spectra, and a series of examples relevant for the optics community are presented and discussed. Outside this parameter range the agreement gracefully degrades but does not distort in shape. Although we demonstrate the method here focusing on stationary random processes, we see no reason why the method could not be extended to simulate non-stationary random processes. © 2011 Optical Society of America
Estimating alarm thresholds and the number of components in mixture distributions
NASA Astrophysics Data System (ADS)
Burr, Tom; Hamada, Michael S.
2012-09-01
Mixtures of probability distributions arise in many nuclear assay and forensic applications, including nuclear weapon detection, neutron multiplicity counting, and in solution monitoring (SM) for nuclear safeguards. SM data is increasingly used to enhance nuclear safeguards in aqueous reprocessing facilities having plutonium in solution form in many tanks. This paper provides background for mixture probability distributions and then focuses on mixtures arising in SM data. SM data can be analyzed by evaluating transfer-mode residuals defined as tank-to-tank transfer differences, and wait-mode residuals defined as changes during non-transfer modes. A previous paper investigated impacts on transfer-mode and wait-mode residuals of event marking errors which arise when the estimated start and/or stop times of tank events such as transfers are somewhat different from the true start and/or stop times. Event marking errors contribute to non-Gaussian behavior and larger variation than predicted on the basis of individual tank calibration studies. This paper illustrates evidence for mixture probability distributions arising from such event marking errors and from effects such as condensation or evaporation during non-transfer modes, and pump carryover during transfer modes. A quantitative assessment of the sample size required to adequately characterize a mixture probability distribution arising in any context is included.
NASA Astrophysics Data System (ADS)
Dubreuil, S.; Salaün, M.; Rodriguez, E.; Petitjean, F.
2018-01-01
This study investigates the construction and identification of the probability distribution of random modal parameters (natural frequencies and effective parameters) in structural dynamics. As these parameters present various types of dependence structures, the retained approach is based on pair copula construction (PCC). A literature review leads us to choose a D-Vine model for the construction of modal parameters probability distributions. Identification of this model is based on likelihood maximization which makes it sensitive to the dimension of the distribution, namely the number of considered modes in our context. To this respect, a mode selection preprocessing step is proposed. It allows the selection of the relevant random modes for a given transfer function. The second point, addressed in this study, concerns the choice of the D-Vine model. Indeed, D-Vine model is not uniquely defined. Two strategies are proposed and compared. The first one is based on the context of the study whereas the second one is purely based on statistical considerations. Finally, the proposed approaches are numerically studied and compared with respect to their capabilities, first in the identification of the probability distribution of random modal parameters and second in the estimation of the 99 % quantiles of some transfer functions.
Singh, A S; Shah, A; Brockmann, A
2018-02-01
In honey bees, continuous foraging at an artificial feeder induced a sustained upregulation of the immediate early genes early growth response protein 1 (Egr-1) and hormone receptor 38 (Hr38). This gene expression response was accompanied by an upregulation of several Egr-1 candidate downstream genes: ecdysone receptor (EcR), dopamine/ecdysteroid receptor (DopEcR), dopamine decarboxylase and dopamine receptor 2. Hr38, EcR and DopEcR are components of the ecdysteroid signalling pathway, which is highly probably involved in learning and memory processes in honey bees and other insects. Time-trained foragers still showed an upregulation of Egr-1 when the feeder was presented at an earlier time of the day, suggesting that the genomic response is more dependent on the food reward than training time. However, presentation of the feeder at the training time without food was still capable of inducing a transient increase in Egr-1 expression. Thus, learnt feeder cues, or even training time, probably affect Egr-1 expression. In contrast, whole brain Egr-1 expression changes did not differ between dancing and nondancing foragers. On the basis of our results we propose that food reward induced continuous foraging ultimately elicits a genomic response involving Egr-1 and Hr38 and their downstream genes. Furthermore this genomic response is highly probably involved in foraging-related learning and memory responses. © 2017 The Royal Entomological Society.
Probability distribution functions in turbulent convection
NASA Technical Reports Server (NTRS)
Balachandar, S.; Sirovich, L.
1991-01-01
Results of an extensive investigation of probability distribution functions (pdfs) for Rayleigh-Benard convection, in hard turbulence regime, are presented. It is shown that the pdfs exhibit a high degree of internal universality. In certain cases this universality is established within two Kolmogorov scales of a boundary. A discussion of the factors leading to the universality is presented.
Power-law tail probabilities of drainage areas in river basins
Veitzer, S.A.; Troutman, B.M.; Gupta, V.K.
2003-01-01
The significance of power-law tail probabilities of drainage areas in river basins was discussed. The convergence to a power law was not observed for all underlying distributions, but for a large class of statistical distributions with specific limiting properties. The article also discussed about the scaling properties of topologic and geometric network properties in river basins.
KINETICS OF LOW SOURCE REACTOR STARTUPS. PART II
DOE Office of Scientific and Technical Information (OSTI.GOV)
hurwitz, H. Jr.; MacMillan, D.B.; Smith, J.H.
1962-06-01
A computational technique is described for computation of the probability distribution of power level for a low source reactor startup. The technique uses a mathematical model, for the time-dependent probability distribution of neutron and precursor concentration, having finite neutron lifetime, one group of delayed neutron precursors, and no spatial dependence. Results obtained by the technique are given. (auth)
Generating an Empirical Probability Distribution for the Andrews-Pregibon Statistic.
ERIC Educational Resources Information Center
Jarrell, Michele G.
A probability distribution was developed for the Andrews-Pregibon (AP) statistic. The statistic, developed by D. F. Andrews and D. Pregibon (1978), identifies multivariate outliers. It is a ratio of the determinant of the data matrix with an observation deleted to the determinant of the entire data matrix. Although the AP statistic has been used…
Animating Statistics: A New Kind of Applet for Exploring Probability Distributions
ERIC Educational Resources Information Center
Kahle, David
2014-01-01
In this article, I introduce a novel applet ("module") for exploring probability distributions, their samples, and various related statistical concepts. The module is primarily designed to be used by the instructor in the introductory course, but it can be used far beyond it as well. It is a free, cross-platform, stand-alone interactive…
Hansen, John P
2003-01-01
Healthcare quality improvement professionals need to understand and use inferential statistics to interpret sample data from their organizations. In quality improvement and healthcare research studies all the data from a population often are not available, so investigators take samples and make inferences about the population by using inferential statistics. This three-part series will give readers an understanding of the concepts of inferential statistics as well as the specific tools for calculating confidence intervals for samples of data. This article, Part 2, describes probability, populations, and samples. The uses of descriptive and inferential statistics are outlined. The article also discusses the properties and probability of normal distributions, including the standard normal distribution.
NASA Astrophysics Data System (ADS)
Merdan, Ziya; Karakuş, Özlem
2016-11-01
The six dimensional Ising model with nearest-neighbor pair interactions has been simulated and verified numerically on the Creutz Cellular Automaton by using five bit demons near the infinite-lattice critical temperature with the linear dimensions L=4,6,8,10. The order parameter probability distribution for six dimensional Ising model has been calculated at the critical temperature. The constants of the analytical function have been estimated by fitting to probability function obtained numerically at the finite size critical point.
NASA Astrophysics Data System (ADS)
Annuar, A.
2015-09-01
We present updated results from an ongoing project to establish the most unbiased census of the Compton- thick active galactic nucleus (CTAGN) population and the intrinsic column density (NH) distribution of the overall AGN population in the local universe, using a sample of mid-infrared (mid-IR) selected AGN within 15 Mpc. We find that 20% of the AGN in the sample are bona-fide CTAGN based upon hard X-ray studies (E > 10 keV). More candidates are then selected using multiwavelength techniques, i.e. mid-IR:X-ray and optical [OIII]5007:X-ray flux ratios. Based on these analyses along with evidence from previous literature, we initially find a further 25% of potential candidates. We then observed two of these candidates, NGC 5643 and NGC 3486, using NuSTAR and is able to confirm the former as a CTAGN and rule out the latter as an obscured AGN. This constrains the total CTAGN population in the sample to 25-40%, though it could potentially be as high as 65% accounting for those that still lack data. Finally, we use these results to estimate the intrinsic NH distribution of the local AGN population. Two more of our CTAGN candidates are scheduled to be observed by NuSTAR, bringing the completeness of hard X-ray energy data of the sample to 65%. This work provides a well-defined local benchmark for AGN unification studies.
Neural correlates of the divergence of instrumental probability distributions.
Liljeholm, Mimi; Wang, Shuo; Zhang, June; O'Doherty, John P
2013-07-24
Flexible action selection requires knowledge about how alternative actions impact the environment: a "cognitive map" of instrumental contingencies. Reinforcement learning theories formalize this map as a set of stochastic relationships between actions and states, such that for any given action considered in a current state, a probability distribution is specified over possible outcome states. Here, we show that activity in the human inferior parietal lobule correlates with the divergence of such outcome distributions-a measure that reflects whether discrimination between alternative actions increases the controllability of the future-and, further, that this effect is dissociable from those of other information theoretic and motivational variables, such as outcome entropy, action values, and outcome utilities. Our results suggest that, although ultimately combined with reward estimates to generate action values, outcome probability distributions associated with alternative actions may be contrasted independently of valence computations, to narrow the scope of the action selection problem.
NASA Astrophysics Data System (ADS)
Xia, Xintao; Wang, Zhongyu
2008-10-01
For some methods of stability analysis of a system using statistics, it is difficult to resolve the problems of unknown probability distribution and small sample. Therefore, a novel method is proposed in this paper to resolve these problems. This method is independent of probability distribution, and is useful for small sample systems. After rearrangement of the original data series, the order difference and two polynomial membership functions are introduced to estimate the true value, the lower bound and the supper bound of the system using fuzzy-set theory. Then empirical distribution function is investigated to ensure confidence level above 95%, and the degree of similarity is presented to evaluate stability of the system. Cases of computer simulation investigate stable systems with various probability distribution, unstable systems with linear systematic errors and periodic systematic errors and some mixed systems. The method of analysis for systematic stability is approved.
A double hit model for the distribution of time to AIDS onset
NASA Astrophysics Data System (ADS)
Chillale, Nagaraja Rao
2013-09-01
Incubation time is a key epidemiologic descriptor of an infectious disease. In the case of HIV infection this is a random variable and is probably the longest one. The probability distribution of incubation time is the major determinant of the relation between the incidences of HIV infection and its manifestation to Aids. This is also one of the key factors used for accurate estimation of AIDS incidence in a region. The present article i) briefly reviews the work done, points out uncertainties in estimation of AIDS onset time and stresses the need for its precise estimation, ii) highlights some of the modelling features of onset distribution including immune failure mechanism, and iii) proposes a 'Double Hit' model for the distribution of time to AIDS onset in the cases of (a) independent and (b) dependent time variables of the two markers and examined the applicability of a few standard probability models.
Optimal moment determination in POME-copula based hydrometeorological dependence modelling
NASA Astrophysics Data System (ADS)
Liu, Dengfeng; Wang, Dong; Singh, Vijay P.; Wang, Yuankun; Wu, Jichun; Wang, Lachun; Zou, Xinqing; Chen, Yuanfang; Chen, Xi
2017-07-01
Copula has been commonly applied in multivariate modelling in various fields where marginal distribution inference is a key element. To develop a flexible, unbiased mathematical inference framework in hydrometeorological multivariate applications, the principle of maximum entropy (POME) is being increasingly coupled with copula. However, in previous POME-based studies, determination of optimal moment constraints has generally not been considered. The main contribution of this study is the determination of optimal moments for POME for developing a coupled optimal moment-POME-copula framework to model hydrometeorological multivariate events. In this framework, margins (marginals, or marginal distributions) are derived with the use of POME, subject to optimal moment constraints. Then, various candidate copulas are constructed according to the derived margins, and finally the most probable one is determined, based on goodness-of-fit statistics. This optimal moment-POME-copula framework is applied to model the dependence patterns of three types of hydrometeorological events: (i) single-site streamflow-water level; (ii) multi-site streamflow; and (iii) multi-site precipitation, with data collected from Yichang and Hankou in the Yangtze River basin, China. Results indicate that the optimal-moment POME is more accurate in margin fitting and the corresponding copulas reflect a good statistical performance in correlation simulation. Also, the derived copulas, capturing more patterns which traditional correlation coefficients cannot reflect, provide an efficient way in other applied scenarios concerning hydrometeorological multivariate modelling.
Simple summation rule for optimal fixation selection in visual search.
Najemnik, Jiri; Geisler, Wilson S
2009-06-01
When searching for a known target in a natural texture, practiced humans achieve near-optimal performance compared to a Bayesian ideal searcher constrained with the human map of target detectability across the visual field [Najemnik, J., & Geisler, W. S. (2005). Optimal eye movement strategies in visual search. Nature, 434, 387-391]. To do so, humans must be good at choosing where to fixate during the search [Najemnik, J., & Geisler, W.S. (2008). Eye movement statistics in humans are consistent with an optimal strategy. Journal of Vision, 8(3), 1-14. 4]; however, it seems unlikely that a biological nervous system would implement the computations for the Bayesian ideal fixation selection because of their complexity. Here we derive and test a simple heuristic for optimal fixation selection that appears to be a much better candidate for implementation within a biological nervous system. Specifically, we show that the near-optimal fixation location is the maximum of the current posterior probability distribution for target location after the distribution is filtered by (convolved with) the square of the retinotopic target detectability map. We term the model that uses this strategy the entropy limit minimization (ELM) searcher. We show that when constrained with human-like retinotopic map of target detectability and human search error rates, the ELM searcher performs as well as the Bayesian ideal searcher, and produces fixation statistics similar to human.
NASA Astrophysics Data System (ADS)
Yin, Bing; Li, Teng; Li, Jin-Feng; Yu, Yang; Li, Jian-Li; Wen, Zhen-Yi; Jiang, Zhen-Yi
2014-03-01
The first theoretical exploration of superhalogen properties of polynuclear structures based on pseudohalogen ligand is reported here via a case study on eight triply-bridged [Mg2(CN)5]- clusters. From our high-level ab initio results, all these clusters are superhalogens due to their high vertical electron detachment energies (VDE), of which the largest value is 8.67 eV at coupled-cluster single double triple (CCSD(T)) level. Although outer valence Green's function results are consistent with CCSD(T) in most cases, it overestimates the VDEs of three anions dramatically by more than 1 eV. Therefore, the combined usage of several theoretical methods is important for the accuracy of purely theoretical prediction of superhalogen properties of new structures. Spatial distribution of the extra electron of high-VDE anions here indicates two features: remarkable aggregation on bridging CN units and non-negligible distribution on every CN unit. These two features lower the potential and kinetic energies of the extra electron respectively and thus lead to high VDE. Besides superhalogen properties, the structures, relative stabilities and thermodynamic stabilities with respect to detachment of CN-1 were also investigated for these anions. The collection of these results indicates that polynuclear structures based on pseudohalogen ligand are promising candidates for new superhalogens with enhanced properties.
Yin, Bing; Li, Teng; Li, Jin-Feng; Yu, Yang; Li, Jian-Li; Wen, Zhen-Yi; Jiang, Zhen-Yi
2014-03-07
The first theoretical exploration of superhalogen properties of polynuclear structures based on pseudohalogen ligand is reported here via a case study on eight triply-bridged [Mg2(CN)5](-) clusters. From our high-level ab initio results, all these clusters are superhalogens due to their high vertical electron detachment energies (VDE), of which the largest value is 8.67 eV at coupled-cluster single double triple (CCSD(T)) level. Although outer valence Green's function results are consistent with CCSD(T) in most cases, it overestimates the VDEs of three anions dramatically by more than 1 eV. Therefore, the combined usage of several theoretical methods is important for the accuracy of purely theoretical prediction of superhalogen properties of new structures. Spatial distribution of the extra electron of high-VDE anions here indicates two features: remarkable aggregation on bridging CN units and non-negligible distribution on every CN unit. These two features lower the potential and kinetic energies of the extra electron respectively and thus lead to high VDE. Besides superhalogen properties, the structures, relative stabilities and thermodynamic stabilities with respect to detachment of CN(-1) were also investigated for these anions. The collection of these results indicates that polynuclear structures based on pseudohalogen ligand are promising candidates for new superhalogens with enhanced properties.
NASA Astrophysics Data System (ADS)
Obuchi, Tomoyuki; Cocco, Simona; Monasson, Rémi
2015-11-01
We consider the problem of learning a target probability distribution over a set of N binary variables from the knowledge of the expectation values (with this target distribution) of M observables, drawn uniformly at random. The space of all probability distributions compatible with these M expectation values within some fixed accuracy, called version space, is studied. We introduce a biased measure over the version space, which gives a boost increasing exponentially with the entropy of the distributions and with an arbitrary inverse `temperature' Γ . The choice of Γ allows us to interpolate smoothly between the unbiased measure over all distributions in the version space (Γ =0) and the pointwise measure concentrated at the maximum entropy distribution (Γ → ∞ ). Using the replica method we compute the volume of the version space and other quantities of interest, such as the distance R between the target distribution and the center-of-mass distribution over the version space, as functions of α =(log M)/N and Γ for large N. Phase transitions at critical values of α are found, corresponding to qualitative improvements in the learning of the target distribution and to the decrease of the distance R. However, for fixed α the distance R does not vary with Γ which means that the maximum entropy distribution is not closer to the target distribution than any other distribution compatible with the observable values. Our results are confirmed by Monte Carlo sampling of the version space for small system sizes (N≤ 10).
Singh, Anuradha; Mantri, Shrikant; Sharma, Monica; Chaudhury, Ashok; Tuli, Rakesh; Roy, Joy
2014-01-16
The cultivated bread wheat (Triticum aestivum L.) possesses unique flour quality, which can be processed into many end-use food products such as bread, pasta, chapatti (unleavened flat bread), biscuit, etc. The present wheat varieties require improvement in processing quality to meet the increasing demand of better quality food products. However, processing quality is very complex and controlled by many genes, which have not been completely explored. To identify the candidate genes whose expressions changed due to variation in processing quality and interaction (quality x development), genome-wide transcriptome studies were performed in two sets of diverse Indian wheat varieties differing for chapatti quality. It is also important to understand the temporal and spatial distributions of their expressions for designing tissue and growth specific functional genomics experiments. Gene-specific two-way ANOVA analysis of expression of about 55 K transcripts in two diverse sets of Indian wheat varieties for chapatti quality at three seed developmental stages identified 236 differentially expressed probe sets (10-fold). Out of 236, 110 probe sets were identified for chapatti quality. Many processing quality related key genes such as glutenin and gliadins, puroindolines, grain softness protein, alpha and beta amylases, proteases, were identified, and many other candidate genes related to cellular and molecular functions were also identified. The ANOVA analysis revealed that the expression of 56 of 110 probe sets was involved in interaction (quality x development). Majority of the probe sets showed differential expression at early stage of seed development i.e. temporal expression. Meta-analysis revealed that the majority of the genes expressed in one or a few growth stages indicating spatial distribution of their expressions. The differential expressions of a few candidate genes such as pre-alpha/beta-gliadin and gamma gliadin were validated by RT-PCR. Therefore, this study identified several quality related key genes including many other genes, their interactions (quality x development) and temporal and spatial distributions. The candidate genes identified for processing quality and information on temporal and spatial distributions of their expressions would be useful for designing wheat improvement programs for processing quality either by changing their expression or development of single nucleotide polymorphisms (SNPs) markers.
2014-01-01
Background The cultivated bread wheat (Triticum aestivum L.) possesses unique flour quality, which can be processed into many end-use food products such as bread, pasta, chapatti (unleavened flat bread), biscuit, etc. The present wheat varieties require improvement in processing quality to meet the increasing demand of better quality food products. However, processing quality is very complex and controlled by many genes, which have not been completely explored. To identify the candidate genes whose expressions changed due to variation in processing quality and interaction (quality x development), genome-wide transcriptome studies were performed in two sets of diverse Indian wheat varieties differing for chapatti quality. It is also important to understand the temporal and spatial distributions of their expressions for designing tissue and growth specific functional genomics experiments. Results Gene-specific two-way ANOVA analysis of expression of about 55 K transcripts in two diverse sets of Indian wheat varieties for chapatti quality at three seed developmental stages identified 236 differentially expressed probe sets (10-fold). Out of 236, 110 probe sets were identified for chapatti quality. Many processing quality related key genes such as glutenin and gliadins, puroindolines, grain softness protein, alpha and beta amylases, proteases, were identified, and many other candidate genes related to cellular and molecular functions were also identified. The ANOVA analysis revealed that the expression of 56 of 110 probe sets was involved in interaction (quality x development). Majority of the probe sets showed differential expression at early stage of seed development i.e. temporal expression. Meta-analysis revealed that the majority of the genes expressed in one or a few growth stages indicating spatial distribution of their expressions. The differential expressions of a few candidate genes such as pre-alpha/beta-gliadin and gamma gliadin were validated by RT-PCR. Therefore, this study identified several quality related key genes including many other genes, their interactions (quality x development) and temporal and spatial distributions. Conclusions The candidate genes identified for processing quality and information on temporal and spatial distributions of their expressions would be useful for designing wheat improvement programs for processing quality either by changing their expression or development of single nucleotide polymorphisms (SNPs) markers. PMID:24433256
Incremental classification learning for anomaly detection in medical images
NASA Astrophysics Data System (ADS)
Giritharan, Balathasan; Yuan, Xiaohui; Liu, Jianguo
2009-02-01
Computer-aided diagnosis usually screens thousands of instances to find only a few positive cases that indicate probable presence of disease.The amount of patient data increases consistently all the time. In diagnosis of new instances, disagreement occurs between a CAD system and physicians, which suggests inaccurate classifiers. Intuitively, misclassified instances and the previously acquired data should be used to retrain the classifier. This, however, is very time consuming and, in some cases where dataset is too large, becomes infeasible. In addition, among the patient data, only a small percentile shows positive sign, which is known as imbalanced data.We present an incremental Support Vector Machines(SVM) as a solution for the class imbalance problem in classification of anomaly in medical images. The support vectors provide a concise representation of the distribution of the training data. Here we use bootstrapping to identify potential candidate support vectors for future iterations. Experiments were conducted using images from endoscopy videos, and the sensitivity and specificity were close to that of SVM trained using all samples available at a given incremental step with significantly improved efficiency in training the classifier.
NASA Technical Reports Server (NTRS)
Schwartz, H. J.
1976-01-01
A Monte Carlo simulation process was used to develop the U.S. daily range requirements for an electric vehicle from probability distributions of trip lengths and frequencies and average annual mileage data. The analysis shows that a car in the U.S. with a practical daily range of 82 miles (132 km) can meet the needs of the owner on 95% of the days of the year, or at all times other than his long vacation trips. Increasing the range of the vehicle beyond this point will not make it more useful to the owner because it will still not provide intercity transportation. A daily range of 82 miles can be provided by an intermediate battery technology level characterized by an energy density of 30 to 50 watt-hours per pound (66 to 110 W-hr/kg). Candidate batteries in this class are nickel-zinc, nickel-iron, and iron-air. The implication of these results for the research goals of far-term battery systems suggests a shift in emphasis toward lower cost and greater life and away from high energy density.
Czyż, Daniel M.; Potluri, Lakshmi-Prasad; Jain-Gupta, Neeta; Riley, Sean P.; Martinez, Juan J.; Steck, Theodore L.; Crosson, Sean; Gabay, Joëlle E.
2014-01-01
ABSTRACT We sought a new approach to treating infections by intracellular bacteria, namely, by altering host cell functions that support their growth. We screened a library of 640 Food and Drug Administration (FDA)-approved compounds for agents that render THP-1 cells resistant to infection by four intracellular pathogens. We identified numerous drugs that are not antibiotics but were highly effective in inhibiting intracellular bacterial growth with limited toxicity to host cells. These compounds are likely to target three kinds of host functions: (i) G protein-coupled receptors, (ii) intracellular calcium signals, and (iii) membrane cholesterol distribution. The compounds that targeted G protein receptor signaling and calcium fluxes broadly inhibited Coxiella burnetii, Legionella pneumophila, Brucella abortus, and Rickettsia conorii, while those directed against cholesterol traffic strongly attenuated the intracellular growth of C. burnetii and L. pneumophila. These pathways probably support intracellular pathogen growth so that drugs that perturb them may be therapeutic candidates. Combining host- and pathogen-directed treatments is a strategy to decrease the emergence of drug-resistant intracellular bacterial pathogens. PMID:25073644
NASA Technical Reports Server (NTRS)
Majewski, Steven R.; Munn, Jeffrey A.; Hawley, Suzanne L.
1994-01-01
Radial velocities have been obtained for six of nine stars identified on the basis of similar distances and common, extreme transverse velocities in the proper motion survey of Majewski (1992) as a candidate halo moving group at the north Galactic pole. These radial velocities correspond to velocities perpendicular to the Galactic plane which span the range -48 +/- 21 to -128 +/- 9 km/sec (but a smaller range, -48 +/- 21 to -86 +/- 19 km/sec, when only our own measurements are considered), significantly different than the expected distribution, with mean 0 km/sec, for a random sample of either halo or thick disk stars. The probability of picking such a set of radial velocities at random is less than 1%. Thus the radial velocity data support the hypothesis that these stars constitute part of a halo moving group or star stream at a distance of approximately 4-5 kpc above the Galactic plane. If real, this moving group is evidence for halo phase space substructure which may be the fossil remains of a destroyed globular cluster, Galactic satellite, or Searle & Zinn (1978) 'fragment.'
A detection method for X-ray images based on wavelet transforms: the case of the ROSAT PSPC.
NASA Astrophysics Data System (ADS)
Damiani, F.; Maggio, A.; Micela, G.; Sciortino, S.
1996-02-01
The authors have developed a method based on wavelet transforms (WT) to detect efficiently sources in PSPC X-ray images. The multiscale approach typical of WT can be used to detect sources with a large range of sizes, and to estimate their size and count rate. Significance thresholds for candidate detections (found as local WT maxima) have been derived from a detailed study of the probability distribution of the WT of a locally uniform background. The use of the exposure map allows good detection efficiency to be retained even near PSPC ribs and edges. The algorithm may also be used to get upper limits to the count rate of undetected objects. Simulations of realistic PSPC images containing either pure background or background+sources were used to test the overall algorithm performances, and to assess the frequency of spurious detections (vs. detection threshold) and the algorithm sensitivity. Actual PSPC images of galaxies and star clusters show the algorithm to have good performance even in cases of extended sources and crowded fields.
Understanding extreme quasar optical variability with CRTS - I. Major AGN flares
NASA Astrophysics Data System (ADS)
Graham, Matthew J.; Djorgovski, S. G.; Drake, Andrew J.; Stern, Daniel; Mahabal, Ashish A.; Glikman, Eilat; Larson, Steve; Christensen, Eric
2017-10-01
There is a large degree of variety in the optical variability of quasars and it is unclear whether this is all attributable to a single (set of) physical mechanism(s). We present the results of a systematic search for major flares in active galactic nucleus (AGN) in the Catalina Real-time Transient Survey as part of a broader study into extreme quasar variability. Such flares are defined in a quantitative manner as being atop of the normal, stochastic variability of quasars. We have identified 51 events from over 900 000 known quasars and high-probability quasar candidates, typically lasting 900 d and with a median peak amplitude of Δm = 1.25 mag. Characterizing the flare profile with a Weibull distribution, we find that nine of the sources are well described by a single-point single-lens model. This supports the proposal by Lawrence et al. that microlensing is a plausible physical mechanism for extreme variability. However, we attribute the majority of our events to explosive stellar-related activity in the accretion disc: superluminous supernovae, tidal disruption events and mergers of stellar mass black holes.
Appetitive behavior, compulsivity, and neurochemistry in Prader-Willi syndrome.
Dimitropoulos, A; Feurer, I D; Roof, E; Stone, W; Butler, M G; Sutcliffe, J; Thompson, T
2000-01-01
Advances in genetic research have led to an increased understanding of genotype-phenotype relationships. Excessive eating and weight gain characteristic of Prader-Willi syndrome (PWS) have been the understandable focus of much of the research. The intense preoccupation with food, lack of satiation, and incessant food seeking are among the most striking features of PWS. It has become increasingly clear that the behavioral phenotype of PWS also includes symptoms similar to obsessive compulsive disorder, which in all probability interact with the incessant hunger and lack of satiation to engender the intense preoccupation and food seeking behavior that is characteristic of this disorder. Several lines of evidence suggest that genetic material on chromosome 15 may alter synthesis, release, metabolism, binding, intrinsic activity, or reuptake of specific neurotransmitters, or alter the receptor numbers and/or distribution involved in modulating feeding. Among the likely candidates are GABAnergic, serotonergic, and neuropeptidergic mechanisms. This review summarizes what is known about the appetitive behavior and compulsivity in PWS and discusses the possible mechanisms underlying these behaviors. MRDD Research Reviews 2000;6:125-130. Copyright 2000 Wiley-Liss, Inc.
NASA Astrophysics Data System (ADS)
Jiang, Quan; Zhong, Shan; Cui, Jie; Feng, Xia-Ting; Song, Leibo
2016-12-01
We investigated the statistical characteristics and probability distribution of the mechanical parameters of natural rock using triaxial compression tests. Twenty cores of Jinping marble were tested under each different levels of confining stress (i.e., 5, 10, 20, 30, and 40 MPa). From these full stress-strain data, we summarized the numerical characteristics and determined the probability distribution form of several important mechanical parameters, including deformational parameters, characteristic strength, characteristic strains, and failure angle. The statistical proofs relating to the mechanical parameters of rock presented new information about the marble's probabilistic distribution characteristics. The normal and log-normal distributions were appropriate for describing random strengths of rock; the coefficients of variation of the peak strengths had no relationship to the confining stress; the only acceptable random distribution for both Young's elastic modulus and Poisson's ratio was the log-normal function; and the cohesive strength had a different probability distribution pattern than the frictional angle. The triaxial tests and statistical analysis also provided experimental evidence for deciding the minimum reliable number of experimental sample and for picking appropriate parameter distributions to use in reliability calculations for rock engineering.
A novel automated spike sorting algorithm with adaptable feature extraction.
Bestel, Robert; Daus, Andreas W; Thielemann, Christiane
2012-10-15
To study the electrophysiological properties of neuronal networks, in vitro studies based on microelectrode arrays have become a viable tool for analysis. Although in constant progress, a challenging task still remains in this area: the development of an efficient spike sorting algorithm that allows an accurate signal analysis at the single-cell level. Most sorting algorithms currently available only extract a specific feature type, such as the principal components or Wavelet coefficients of the measured spike signals in order to separate different spike shapes generated by different neurons. However, due to the great variety in the obtained spike shapes, the derivation of an optimal feature set is still a very complex issue that current algorithms struggle with. To address this problem, we propose a novel algorithm that (i) extracts a variety of geometric, Wavelet and principal component-based features and (ii) automatically derives a feature subset, most suitable for sorting an individual set of spike signals. Thus, there is a new approach that evaluates the probability distribution of the obtained spike features and consequently determines the candidates most suitable for the actual spike sorting. These candidates can be formed into an individually adjusted set of spike features, allowing a separation of the various shapes present in the obtained neuronal signal by a subsequent expectation maximisation clustering algorithm. Test results with simulated data files and data obtained from chick embryonic neurons cultured on microelectrode arrays showed an excellent classification result, indicating the superior performance of the described algorithm approach. Copyright © 2012 Elsevier B.V. All rights reserved.
Andreo, Verónica; Glass, Gregory; Shields, Timothy; Provensal, Cecilia; Polop, Jaime
2011-09-01
We constructed a model to predict the potential distribution of Oligoryzomys longicaudatus, the reservoir of Andes virus (Genus: Hantavirus), in Argentina. We developed an extensive database of occurrence records from published studies and our own surveys and compared two methods to model the probability of O. longicaudatus presence; logistic regression and MaxEnt algorithm. The environmental variables used were tree, grass and bare soil cover from MODIS imagery and, altitude and 19 bioclimatic variables from WorldClim database. The models performances were evaluated and compared both by threshold dependent and independent measures. The best models included tree and grass cover, mean diurnal temperature range, and precipitation of the warmest and coldest seasons. The potential distribution maps for O. longicaudatus predicted the highest occurrence probabilities along the Andes range, from 32°S and narrowing southwards. They also predicted high probabilities for the south-central area of Argentina, reaching the Atlantic coast. The Hantavirus Pulmonary Syndrome cases coincided with mean occurrence probabilities of 95 and 77% for logistic and MaxEnt models, respectively. HPS transmission zones in Argentine Patagonia matched the areas with the highest probability of presence. Therefore, colilargos presence probability may provide an approximate risk of transmission and act as an early tool to guide control and prevention plans.
Properties of Starless Clumps through Protoclusters from the Bolocam Galactic Plane Survey
NASA Astrophysics Data System (ADS)
Svoboda, Brian E.; Shirley, Yancy
2014-07-01
High mass stars play a key role in the physical and chemical evolution of the interstellar medium, yet the evolution of physical properties for high-mass star-forming regions remains unclear. We sort a sample of ~4668 molecular cloud clumps from the Bolocam Galactic Plane Survey (BGPS) into different evolutionary stages by combining the BGPS 1.1 mm continuum and observational diagnostics of star-formation activity from a variety of Galactic plane surveys: 70 um compact sources, mid-IR color-selected YSOs, H2O and CH3OH masers, EGOs, and UCHII regions. We apply Monte Carlo techniques to distance probability distribution functions (DPDFs) in order to marginalize over the kinematic distance ambiguity and calculate distributions for derived quantities of clumps in different evolutionary stages. We also present a combined NH3 and H2O maser catalog for ~1590 clumps from the literature and our own GBT 100m observations. We identify a sub-sample of 440 dense clumps with no star-formation indicators, representing the largest and most robust sample of pre-protocluster candidates from a blind survey to date. Distributions of I(HCO+), I(N2H+), dv(HCO+), dv(N2H+), mass surface density, and kinetic temperature show strong progressions when separated by evolutionary stage. No progressions are found in size or dust mass; however, weak progressions are observed in area > 2 pc^2 and dust mass > 3 10^3 Msun. An observed breakdown occurs in the size-linewidth relationship and we find no improvement when sampling by evolutionary stage.
A single population of red globular clusters around the massive compact galaxy NGC 1277
NASA Astrophysics Data System (ADS)
Beasley, Michael A.; Trujillo, Ignacio; Leaman, Ryan; Montes, Mireia
2018-03-01
Massive galaxies are thought to form in two phases: an initial collapse of gas and giant burst of central star formation, followed by the later accretion of material that builds up their stellar and dark-matter haloes. The systems of globular clusters within such galaxies are believed to form in a similar manner. The initial central burst forms metal-rich (spectrally red) clusters, whereas more metal-poor (spectrally blue) clusters are brought in by the later accretion of less-massive satellites. This formation process is thought to result in the multimodal optical colour distributions that are seen in the globular cluster systems of massive galaxies. Here we report optical observations of the massive relic-galaxy candidate NGC 1277—a nearby, un-evolved example of a high-redshift ‘red nugget’ galaxy. We find that the optical colour distribution of the cluster system of NGC 1277 is unimodal and entirely red. This finding is in strong contrast to other galaxies of similar and larger stellar mass, the cluster systems of which always exhibit (and are generally dominated by) blue clusters. We argue that the colour distribution of the cluster system of NGC 1277 indicates that the galaxy has undergone little (if any) mass accretion after its initial collapse, and use simulations of possible merger histories to show that the stellar mass due to accretion is probably at most ten per cent of the total stellar mass of the galaxy. These results confirm that NGC 1277 is a genuine relic galaxy and demonstrate that blue clusters constitute an accreted population in present-day massive galaxies.
Aerosol Properties of the Atmospheres of Extrasolar Giant Planets
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lavvas, P.; Koskinen, T., E-mail: panayotis.lavvas@univ-reims.fr
2017-09-20
We use a model of aerosol microphysics to investigate the impact of high-altitude photochemical aerosols on the transmission spectra and atmospheric properties of close-in exoplanets, such as HD 209458 b and HD 189733 b. The results depend strongly on the temperature profiles in the middle and upper atmospheres, which are poorly understood. Nevertheless, our model of HD 189733 b, based on the most recently inferred temperature profiles, produces an aerosol distribution that matches the observed transmission spectrum. We argue that the hotter temperature of HD 209458 b inhibits the production of high-altitude aerosols and leads to the appearance of amore » clearer atmosphere than on HD 189733 b. The aerosol distribution also depends on the particle composition, photochemical production, and atmospheric mixing. Due to degeneracies among these inputs, current data cannot constrain the aerosol properties in detail. Instead, our work highlights the role of different factors in controlling the aerosol distribution that will prove useful in understanding different observations, including those from future missions. For the atmospheric mixing efficiency suggested by general circulation models, we find that the aerosol particles are small (∼nm) and probably spherical. We further conclude that a composition based on complex hydrocarbons (soots) is the most likely candidate to survive the high temperatures in hot-Jupiter atmospheres. Such particles would have a significant impact on the energy balance of HD 189733 b’s atmosphere and should be incorporated in future studies of atmospheric structure. We also evaluate the contribution of external sources to photochemical aerosol formation and find that their spectral signature is not consistent with observations.« less
A single population of red globular clusters around the massive compact galaxy NGC 1277.
Beasley, Michael A; Trujillo, Ignacio; Leaman, Ryan; Montes, Mireia
2018-03-22
Massive galaxies are thought to form in two phases: an initial collapse of gas and giant burst of central star formation, followed by the later accretion of material that builds up their stellar and dark-matter haloes. The systems of globular clusters within such galaxies are believed to form in a similar manner. The initial central burst forms metal-rich (spectrally red) clusters, whereas more metal-poor (spectrally blue) clusters are brought in by the later accretion of less-massive satellites. This formation process is thought to result in the multimodal optical colour distributions that are seen in the globular cluster systems of massive galaxies. Here we report optical observations of the massive relic-galaxy candidate NGC 1277-a nearby, un-evolved example of a high-redshift 'red nugget' galaxy. We find that the optical colour distribution of the cluster system of NGC 1277 is unimodal and entirely red. This finding is in strong contrast to other galaxies of similar and larger stellar mass, the cluster systems of which always exhibit (and are generally dominated by) blue clusters. We argue that the colour distribution of the cluster system of NGC 1277 indicates that the galaxy has undergone little (if any) mass accretion after its initial collapse, and use simulations of possible merger histories to show that the stellar mass due to accretion is probably at most ten per cent of the total stellar mass of the galaxy. These results confirm that NGC 1277 is a genuine relic galaxy and demonstrate that blue clusters constitute an accreted population in present-day massive galaxies.
Cowell, Robert G
2018-05-04
Current models for single source and mixture samples, and probabilistic genotyping software based on them used for analysing STR electropherogram data, assume simple probability distributions, such as the gamma distribution, to model the allelic peak height variability given the initial amount of DNA prior to PCR amplification. Here we illustrate how amplicon number distributions, for a model of the process of sample DNA collection and PCR amplification, may be efficiently computed by evaluating probability generating functions using discrete Fourier transforms. Copyright © 2018 Elsevier B.V. All rights reserved.
An Integrated Framework for Model-Based Distributed Diagnosis and Prognosis
2012-09-01
0 : t) denotes all measurements observed up to time t. The goal of prognosis is to determine the end of (use- ful) life ( EOL ) of a system, and/or its...remaining useful life (RUL). For a given fault, f , using the fault estimate, p(xf (t),θf (t)|y(0 : t)), a probability distribution of EOL , p(EOLf (tP...is stochas- tic, EOL /RUL are random variables and we represent them by probability distributions. The acceptable behavior of the system is expressed
NASA Astrophysics Data System (ADS)
Eem, Changkyoung; Kim, Iksu; Hong, Hyunki
2015-07-01
A method to estimate the environmental illumination distribution of a scene with gradient-based ray and candidate shadow maps is presented. In the shadow segmentation stage, we apply a Canny edge detector to the shadowed image by using a three-dimensional (3-D) augmented reality (AR) marker of a known size and shape. Then the hierarchical tree of the connected edge components representing the topological relation is constructed, and the connected components are merged, taking their hierarchical structures into consideration. A gradient-based ray that is perpendicular to the gradient of the edge pixel in the shadow image can be used to extract the shadow regions. In the light source detection stage, shadow regions with both a 3-D AR marker and the light sources are partitioned into candidate shadow maps. A simple logic operation between each candidate shadow map and the segmented shadow is used to efficiently compute the area ratio between them. The proposed method successively extracts the main light sources according to their relative contributions on the segmented shadows. The proposed method can reduce unwanted effects due to the sampling positions in the shadow region and the threshold values in the shadow edge detection.
Encounter risk analysis of rainfall and reference crop evapotranspiration in the irrigation district
NASA Astrophysics Data System (ADS)
Zhang, Jinping; Lin, Xiaomin; Zhao, Yong; Hong, Yang
2017-09-01
Rainfall and reference crop evapotranspiration are random but mutually affected variables in the irrigation district, and their encounter situation can determine water shortage risks under the contexts of natural water supply and demand. However, in reality, the rainfall and reference crop evapotranspiration may have different marginal distributions and their relations are nonlinear. In this study, based on the annual rainfall and reference crop evapotranspiration data series from 1970 to 2013 in the Luhun irrigation district of China, the joint probability distribution of rainfall and reference crop evapotranspiration are developed with the Frank copula function. Using the joint probability distribution, the synchronous-asynchronous encounter risk, conditional joint probability, and conditional return period of different combinations of rainfall and reference crop evapotranspiration are analyzed. The results show that the copula-based joint probability distributions of rainfall and reference crop evapotranspiration are reasonable. The asynchronous encounter probability of rainfall and reference crop evapotranspiration is greater than their synchronous encounter probability, and the water shortage risk associated with meteorological drought (i.e. rainfall variability) is more prone to appear. Compared with other states, there are higher conditional joint probability and lower conditional return period in either low rainfall or high reference crop evapotranspiration. For a specifically high reference crop evapotranspiration with a certain frequency, the encounter risk of low rainfall and high reference crop evapotranspiration is increased with the decrease in frequency. For a specifically low rainfall with a certain frequency, the encounter risk of low rainfall and high reference crop evapotranspiration is decreased with the decrease in frequency. When either the high reference crop evapotranspiration exceeds a certain frequency or low rainfall does not exceed a certain frequency, the higher conditional joint probability and lower conditional return period of various combinations likely cause a water shortage, but the water shortage is not severe.
NASA Astrophysics Data System (ADS)
Zhou, H.; Chen, B.; Han, Z. X.; Zhang, F. Q.
2009-05-01
The study on probability density function and distribution function of electricity prices contributes to the power suppliers and purchasers to estimate their own management accurately, and helps the regulator monitor the periods deviating from normal distribution. Based on the assumption of normal distribution load and non-linear characteristic of the aggregate supply curve, this paper has derived the distribution of electricity prices as the function of random variable of load. The conclusion has been validated with the electricity price data of Zhejiang market. The results show that electricity prices obey normal distribution approximately only when supply-demand relationship is loose, whereas the prices deviate from normal distribution and present strong right-skewness characteristic. Finally, the real electricity markets also display the narrow-peak characteristic when undersupply occurs.
Source clustering in the Hi-GAL survey determined using a minimum spanning tree method
NASA Astrophysics Data System (ADS)
Beuret, M.; Billot, N.; Cambrésy, L.; Eden, D. J.; Elia, D.; Molinari, S.; Pezzuto, S.; Schisano, E.
2017-01-01
Aims: The aims are to investigate the clustering of the far-infrared sources from the Herschel infrared Galactic Plane Survey (Hi-GAL) in the Galactic longitude range of -71 to 67 deg. These clumps, and their spatial distribution, are an imprint of the original conditions within a molecular cloud. This will produce a catalogue of over-densities. Methods: The minimum spanning tree (MST) method was used to identify the over-densities in two dimensions. The catalogue was further refined by folding in heliocentric distances, resulting in more reliable over-densities, which are cluster candidates. Results: We found 1633 over-densities with more than ten members. Of these, 496 are defined as cluster candidates because of the reliability of the distances, with a further 1137 potential cluster candidates. The spatial distributions of the cluster candidates are different in the first and fourth quadrants, with all clusters following the spiral structure of the Milky Way. The cluster candidates are fractal. The clump mass functions of the clustered and isolated are statistically indistinguishable from each other and are consistent with Kroupa's initial mass function. Hi-GAL is a key-project of the Herschel Space Observatory survey (Pilbratt et al. 2010) and uses the PACS (Poglitsch et al. 2010) and SPIRE (Griffin et al. 2010) cameras in parallel mode.The catalogues of cluster candidates and potential clusters are only available at the CDS via anonymous ftp to http://cdsarc.u-strasbg.fr (http://130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/597/A114
Time-correlated coincidences at the sudbury neutrino observatory: An antineutrino search
NASA Astrophysics Data System (ADS)
Shokair, Timothy Milad
This dissertation presents a search for antineutrinos in all three phases of data from the Sudbury Neutrino Observatory. This work presents a new method for detecting time correlated coincidences in water detectors. There are two separate searches: an outside search for the inverse beta decay of antineutrinos on protons and an inside search for the inverse beta decay of antineutrinos on deuterons. The inside search found 3 antineutrino candidates in Phase I with an expected background of 3.83+0.71-0.72 events, 28 antineutrino candidates in Phase II with an expected background of 21.25+3.72-3.75 events, 4 antineutrino candidates in Phase III with an expected background of 6.06 +/- 1.14 events. The outside search found 4 antineutrino candidates in Phase I with an expected background of 1.21+0.14-0.17 events, 8 antineutrino candidates in Phase II with an expected background of 9.77+1.06-1.34 events, 0 antineutrino candidates in Phase III with an expected background of 0.46 +/- 0.29 events. Including the expected contribution of antineutrinos from nuclear reactors after oscillations, a limit on the solar antineutrino flux is computed to be F8Bn¯ ≤ 2.5 x 103 cm-2s -1. Taking the flux limit and the measured 8B solar neutrino flux, a limit on the neutrino to antineutrino conversion probability of P(nu → nu) ≤ 5.0 x 10-4. These limits are the best limits from a water detector.
Differentially Private Frequent Sequence Mining via Sampling-based Candidate Pruning
Xu, Shengzhi; Cheng, Xiang; Li, Zhengyi; Xiong, Li
2016-01-01
In this paper, we study the problem of mining frequent sequences under the rigorous differential privacy model. We explore the possibility of designing a differentially private frequent sequence mining (FSM) algorithm which can achieve both high data utility and a high degree of privacy. We found, in differentially private FSM, the amount of required noise is proportionate to the number of candidate sequences. If we could effectively reduce the number of unpromising candidate sequences, the utility and privacy tradeoff can be significantly improved. To this end, by leveraging a sampling-based candidate pruning technique, we propose a novel differentially private FSM algorithm, which is referred to as PFS2. The core of our algorithm is to utilize sample databases to further prune the candidate sequences generated based on the downward closure property. In particular, we use the noisy local support of candidate sequences in the sample databases to estimate which sequences are potentially frequent. To improve the accuracy of such private estimations, a sequence shrinking method is proposed to enforce the length constraint on the sample databases. Moreover, to decrease the probability of misestimating frequent sequences as infrequent, a threshold relaxation method is proposed to relax the user-specified threshold for the sample databases. Through formal privacy analysis, we show that our PFS2 algorithm is ε-differentially private. Extensive experiments on real datasets illustrate that our PFS2 algorithm can privately find frequent sequences with high accuracy. PMID:26973430
DeVore, Matthew S.; Gull, Stephen F.; Johnson, Carey K.
2012-01-01
We describe a method for analysis of single-molecule Förster resonance energy transfer (FRET) burst measurements using classic maximum entropy. Classic maximum entropy determines the Bayesian inference for the joint probability describing the total fluorescence photons and the apparent FRET efficiency. The method was tested with simulated data and then with DNA labeled with fluorescent dyes. The most probable joint distribution can be marginalized to obtain both the overall distribution of fluorescence photons and the apparent FRET efficiency distribution. This method proves to be ideal for determining the distance distribution of FRET-labeled biomolecules, and it successfully predicts the shape of the recovered distributions. PMID:22338694
NASA Astrophysics Data System (ADS)
Saitoh, Kuniyasu; Magnanimo, Vanessa; Luding, Stefan
2017-10-01
Employing two-dimensional molecular dynamics (MD) simulations of soft particles, we study their non-affine responses to quasi-static isotropic compression where the effects of microscopic friction between the particles in contact and particle size distributions are examined. To quantify complicated restructuring of force-chain networks under isotropic compression, we introduce the conditional probability distributions (CPDs) of particle overlaps such that a master equation for distribution of overlaps in the soft particle packings can be constructed. From our MD simulations, we observe that the CPDs are well described by q-Gaussian distributions, where we find that the correlation for the evolution of particle overlaps is suppressed by microscopic friction, while it significantly increases with the increase of poly-dispersity.
Ordinal probability effect measures for group comparisons in multinomial cumulative link models.
Agresti, Alan; Kateri, Maria
2017-03-01
We consider simple ordinal model-based probability effect measures for comparing distributions of two groups, adjusted for explanatory variables. An "ordinal superiority" measure summarizes the probability that an observation from one distribution falls above an independent observation from the other distribution, adjusted for explanatory variables in a model. The measure applies directly to normal linear models and to a normal latent variable model for ordinal response variables. It equals Φ(β/2) for the corresponding ordinal model that applies a probit link function to cumulative multinomial probabilities, for standard normal cdf Φ and effect β that is the coefficient of the group indicator variable. For the more general latent variable model for ordinal responses that corresponds to a linear model with other possible error distributions and corresponding link functions for cumulative multinomial probabilities, the ordinal superiority measure equals exp(β)/[1+exp(β)] with the log-log link and equals approximately exp(β/2)/[1+exp(β/2)] with the logit link, where β is the group effect. Another ordinal superiority measure generalizes the difference of proportions from binary to ordinal responses. We also present related measures directly for ordinal models for the observed response that need not assume corresponding latent response models. We present confidence intervals for the measures and illustrate with an example. © 2016, The International Biometric Society.
Aitken, C G
1999-07-01
It is thought that, in a consignment of discrete units, a certain proportion of the units contain illegal material. A sample of the consignment is to be inspected. Various methods for the determination of the sample size are compared. The consignment will be considered as a random sample from some super-population of units, a certain proportion of which contain drugs. For large consignments, a probability distribution, known as the beta distribution, for the proportion of the consignment which contains illegal material is obtained. This distribution is based on prior beliefs about the proportion. Under certain specific conditions the beta distribution gives the same numerical results as an approach based on the binomial distribution. The binomial distribution provides a probability for the number of units in a sample which contain illegal material, conditional on knowing the proportion of the consignment which contains illegal material. This is in contrast to the beta distribution which provides probabilities for the proportion of a consignment which contains illegal material, conditional on knowing the number of units in the sample which contain illegal material. The interpretation when the beta distribution is used is much more intuitively satisfactory. It is also much more flexible in its ability to cater for prior beliefs which may vary given the different circumstances of different crimes. For small consignments, a distribution, known as the beta-binomial distribution, for the number of units in the consignment which are found to contain illegal material, is obtained, based on prior beliefs about the number of units in the consignment which are thought to contain illegal material. As with the beta and binomial distributions for large samples, it is shown that, in certain specific conditions, the beta-binomial and hypergeometric distributions give the same numerical results. However, the beta-binomial distribution, as with the beta distribution, has a more intuitively satisfactory interpretation and greater flexibility. The beta and the beta-binomial distributions provide methods for the determination of the minimum sample size to be taken from a consignment in order to satisfy a certain criterion. The criterion requires the specification of a proportion and a probability.
An artificial neural network to discover hypervelocity stars: candidates in Gaia DR1/TGAS
NASA Astrophysics Data System (ADS)
Marchetti, T.; Rossi, E. M.; Kordopatis, G.; Brown, A. G. A.; Rimoldi, A.; Starkenburg, E.; Youakim, K.; Ashley, R.
2017-09-01
The paucity of hypervelocity stars (HVSs) known to date has severely hampered their potential to investigate the stellar population of the Galactic Centre and the Galactic potential. The first Gaia data release (DR1, 2016 September 14) gives an opportunity to increase the current sample. The challenge is the disparity between the expected number of HVSs and that of bound background stars. We have applied a novel data mining algorithm based on machine learning techniques, an artificial neural network, to the Tycho-Gaia astrometric solution catalogue. With no pre-selection of data, we could exclude immediately ˜99 per cent of the stars in the catalogue and find 80 candidates with more than 90 per cent predicted probability to be HVSs, based only on their position, proper motions and parallax. We have cross-checked our findings with other spectroscopic surveys, determining radial velocities for 30 and spectroscopic distances for five candidates. In addition, follow-up observations have been carried out at the Isaac Newton Telescope for 22 stars, for which we obtained radial velocities and distance estimates. We discover 14 stars with a total velocity in the Galactic rest frame >400 km s-1, and five of these have a probability of >50 per cent of being unbound from the Milky Way. Tracing back their orbits in different Galactic potential models, we find one possible unbound HVS with v ˜ 520 km s-1, five bound HVSs and, notably, five runaway stars with median velocity between 400 and 780 km s-1. At the moment, uncertainties in the distance estimates and ages are too large to confirm the nature of our candidates by narrowing down their ejection location, and we wait for future Gaia releases to validate the quality of our sample. This test successfully demonstrates the feasibility of our new data-mining routine.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wu Xuebing; Wang Ran; Bian Fuyan
2011-09-15
The identification of quasars in the redshift range 2.2 < z < 3 is known to be very inefficient because the optical colors of such quasars are indistinguishable from those of stars. Recent studies have proposed using optical variability or near-infrared (near-IR) colors to improve the identification of the missing quasars in this redshift range. Here we present a case study combining both methods. We select a sample of 70 quasar candidates from variables in Sloan Digital Sky Survey (SDSS) Stripe 82, which are non-ultraviolet excess sources and have UKIDSS near-IR public data. They are clearly separated into two partsmore » on the Y - K/g - z color-color diagram, and 59 of them meet or lie close to a newly proposed Y - K/g - z selection criterion for z < 4 quasars. Of these 59 sources, 44 were previously identified as quasars in SDSS DR7, and 35 of them are quasars at 2.2 < z < 3. We present spectroscopic observations of 14 of 15 remaining quasar candidates using the Bok 2.3 m telescope and the MMT 6.5 m telescope, and successfully identify all of them as new quasars at z = 2.36-2.88. We also apply this method to a sample of 643 variable quasar candidates with SDSS-UKIDSS nine-band photometric data selected from 1875 new quasar candidates in SDSS Stripe 82 given by Butler and Bloom based on the time-series selections, and find that 188 of them are probably new quasars with photometric redshifts at 2.2 < z < 3. Our results indicate that the combination of optical variability and optical/near-IR colors is probably the most efficient way to find 2.2 < z < 3 quasars and is very helpful for constructing a complete quasar sample. We discuss its implications for ongoing and upcoming large optical and near-IR sky surveys.« less
The maximum entropy method of moments and Bayesian probability theory
NASA Astrophysics Data System (ADS)
Bretthorst, G. Larry
2013-08-01
The problem of density estimation occurs in many disciplines. For example, in MRI it is often necessary to classify the types of tissues in an image. To perform this classification one must first identify the characteristics of the tissues to be classified. These characteristics might be the intensity of a T1 weighted image and in MRI many other types of characteristic weightings (classifiers) may be generated. In a given tissue type there is no single intensity that characterizes the tissue, rather there is a distribution of intensities. Often this distributions can be characterized by a Gaussian, but just as often it is much more complicated. Either way, estimating the distribution of intensities is an inference problem. In the case of a Gaussian distribution, one must estimate the mean and standard deviation. However, in the Non-Gaussian case the shape of the density function itself must be inferred. Three common techniques for estimating density functions are binned histograms [1, 2], kernel density estimation [3, 4], and the maximum entropy method of moments [5, 6]. In the introduction, the maximum entropy method of moments will be reviewed. Some of its problems and conditions under which it fails will be discussed. Then in later sections, the functional form of the maximum entropy method of moments probability distribution will be incorporated into Bayesian probability theory. It will be shown that Bayesian probability theory solves all of the problems with the maximum entropy method of moments. One gets posterior probabilities for the Lagrange multipliers, and, finally, one can put error bars on the resulting estimated density function.
How Common are the Magellanic Clouds
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Lulu; Gerke, Brian F.; Wechsler, Risa H.
2011-05-20
We introduce a probabilistic approach to the problem of counting dwarf satellites around host galaxies in databases with limited redshift information. This technique is used to investigate the occurrence of satellites with luminosities similar to the Magellanic Clouds around hosts with properties similar to the Milky Way in the object catalog of the Sloan Digital Sky Survey. Our analysis uses data from SDSS Data Release 7, selecting candidate Milky-Way-like hosts from the spectroscopic catalog and candidate analogs of the Magellanic Clouds from the photometric catalog. Our principal result is the probability for a Milky-Way-like galaxy to host N{sub sat} closemore » satellites with luminosities similar to the Magellanic Clouds. We find that 81 percent of galaxies like the Milky Way have no such satellites within a radius of 150 kpc, 11 percent have one, and only 3.5 percent of hosts have two. The probabilities are robust to changes in host and satellite selection criteria, background-estimation technique, and survey depth. These results demonstrate that the Milky Way has significantly more satellites than a typical galaxy of its luminosity; this fact is useful for understanding the larger cosmological context of our home galaxy.« less
NASA Astrophysics Data System (ADS)
Koshinchanov, Georgy; Dimitrov, Dobri
2008-11-01
The characteristics of rainfall intensity are important for many purposes, including design of sewage and drainage systems, tuning flood warning procedures, etc. Those estimates are usually statistical estimates of the intensity of precipitation realized for certain period of time (e.g. 5, 10 min., etc) with different return period (e.g. 20, 100 years, etc). The traditional approach in evaluating the mentioned precipitation intensities is to process the pluviometer's records and fit probability distribution to samples of intensities valid for certain locations ore regions. Those estimates further become part of the state regulations to be used for various economic activities. Two problems occur using the mentioned approach: 1. Due to various factors the climate conditions are changed and the precipitation intensity estimates need regular update; 2. As far as the extremes of the probability distribution are of particular importance for the practice, the methodology of the distribution fitting needs specific attention to those parts of the distribution. The aim of this paper is to make review of the existing methodologies for processing the intensive rainfalls and to refresh some of the statistical estimates for the studied areas. The methodologies used in Bulgaria for analyzing the intensive rainfalls and produce relevant statistical estimates: The method of the maximum intensity, used in the National Institute of Meteorology and Hydrology to process and decode the pluviometer's records, followed by distribution fitting for each precipitation duration period; As the above, but with separate modeling of probability distribution for the middle and high probability quantiles. Method is similar to the first one, but with a threshold of 0,36 mm/min of intensity; Another method proposed by the Russian hydrologist G. A. Aleksiev for regionalization of estimates over some territory, improved and adapted by S. Gerasimov for Bulgaria; Next method is considering only the intensive rainfalls (if any) during the day with the maximal annual daily precipitation total for a given year; Conclusions are drown on the relevance and adequacy of the applied methods.
Smith, Erik A.; Sanocki, Chris A.; Lorenz, David L.; Jacobsen, Katrin E.
2017-12-27
Streamflow distribution maps for the Cannon River and St. Louis River drainage basins were developed by the U.S. Geological Survey, in cooperation with the Legislative-Citizen Commission on Minnesota Resources, to illustrate relative and cumulative streamflow distributions. The Cannon River was selected to provide baseline data to assess the effects of potential surficial sand mining, and the St. Louis River was selected to determine the effects of ongoing Mesabi Iron Range mining. Each drainage basin (Cannon, St. Louis) was subdivided into nested drainage basins: the Cannon River was subdivided into 152 nested drainage basins, and the St. Louis River was subdivided into 353 nested drainage basins. For each smaller drainage basin, the estimated volumes of groundwater discharge (as base flow) and surface runoff flowing into all surface-water features were displayed under the following conditions: (1) extreme low-flow conditions, comparable to an exceedance-probability quantile of 0.95; (2) low-flow conditions, comparable to an exceedance-probability quantile of 0.90; (3) a median condition, comparable to an exceedance-probability quantile of 0.50; and (4) a high-flow condition, comparable to an exceedance-probability quantile of 0.02.Streamflow distribution maps were developed using flow-duration curve exceedance-probability quantiles in conjunction with Soil-Water-Balance model outputs; both the flow-duration curve and Soil-Water-Balance models were built upon previously published U.S. Geological Survey reports. The selected streamflow distribution maps provide a proactive water management tool for State cooperators by illustrating flow rates during a range of hydraulic conditions. Furthermore, after the nested drainage basins are highlighted in terms of surface-water flows, the streamflows can be evaluated in the context of meeting specific ecological flows under different flow regimes and potentially assist with decisions regarding groundwater and surface-water appropriations. Presented streamflow distribution maps are foundational work intended to support the development of additional streamflow distribution maps that include statistical constraints on the selected flow conditions.
NASA Technical Reports Server (NTRS)
Hsu, J. P.
1983-01-01
The foundation of the quantum field theory is changed by introducing a new universal probability principle into field operators: one single inherent and invariant probability distribution P(/k/) is postulated for boson and fermion field oscillators. This can be accomplished only when one treats the four-dimensional symmetry from a broad viewpoint. Special relativity is too restrictive to allow such a universal probability principle. A radical length, R, appears in physics through the probability distribution P(/k/). The force between two point particles vanishes when their relative distance tends to zero. This appears to be a general property for all forces and resembles the property of asymptotic freedom. The usual infinities in vacuum fluctuations and in local interactions, however complicated they may be, are all removed from quantum field theories. In appendix A a simple finite and unitary theory of unified electroweak interactions is discussed without assuming Higgs scalar bosons.
Briggs, Andrew H; Ades, A E; Price, Martin J
2003-01-01
In structuring decision models of medical interventions, it is commonly recommended that only 2 branches be used for each chance node to avoid logical inconsistencies that can arise during sensitivity analyses if the branching probabilities do not sum to 1. However, information may be naturally available in an unconditional form, and structuring a tree in conditional form may complicate rather than simplify the sensitivity analysis of the unconditional probabilities. Current guidance emphasizes using probabilistic sensitivity analysis, and a method is required to provide probabilistic probabilities over multiple branches that appropriately represents uncertainty while satisfying the requirement that mutually exclusive event probabilities should sum to 1. The authors argue that the Dirichlet distribution, the multivariate equivalent of the beta distribution, is appropriate for this purpose and illustrate its use for generating a fully probabilistic transition matrix for a Markov model. Furthermore, they demonstrate that by adopting a Bayesian approach, the problem of observing zero counts for transitions of interest can be overcome.
Jeffrey H. Gove
2003-01-01
Many of the most popular sampling schemes used in forestry are probability proportional to size methods. These methods are also referred to as size biased because sampling is actually from a weighted form of the underlying population distribution. Length- and area-biased sampling are special cases of size-biased sampling where the probability weighting comes from a...
Surface Impact Simulations of Helium Nanodroplets
2015-06-30
mechanical delocalization of the individual helium atoms in the droplet and the quan- tum statistical effects that accompany the interchange of identical...incorporates the effects of atomic delocaliza- tion by treating individual atoms as smeared-out probability distributions that move along classical...probability density distributions to give effec- tive interatomic potential energy curves that have zero-point averaging effects built into them [25
ERIC Educational Resources Information Center
Lunsford, M. Leigh; Rowell, Ginger Holmes; Goodson-Espy, Tracy
2006-01-01
We applied a classroom research model to investigate student understanding of sampling distributions of sample means and the Central Limit Theorem in post-calculus introductory probability and statistics courses. Using a quantitative assessment tool developed by previous researchers and a qualitative assessment tool developed by the authors, we…
NASA Astrophysics Data System (ADS)
Iwakoshi, Takehisa; Hirota, Osamu
2014-10-01
This study will test an interpretation in quantum key distribution (QKD) that trace distance between the distributed quantum state and the ideal mixed state is a maximum failure probability of the protocol. Around 2004, this interpretation was proposed and standardized to satisfy both of the key uniformity in the context of universal composability and operational meaning of the failure probability of the key extraction. However, this proposal has not been verified concretely yet for many years while H. P. Yuen and O. Hirota have thrown doubt on this interpretation since 2009. To ascertain this interpretation, a physical random number generator was employed to evaluate key uniformity in QKD. In this way, we calculated statistical distance which correspond to trace distance in quantum theory after a quantum measurement is done, then we compared it with the failure probability whether universal composability was obtained. As a result, the degree of statistical distance of the probability distribution of the physical random numbers and the ideal uniformity was very large. It is also explained why trace distance is not suitable to guarantee the security in QKD from the view point of quantum binary decision theory.
SPACE WARPS - I. Crowdsourcing the discovery of gravitational lenses
NASA Astrophysics Data System (ADS)
Marshall, Philip J.; Verma, Aprajita; More, Anupreeta; Davis, Christopher P.; More, Surhud; Kapadia, Amit; Parrish, Michael; Snyder, Chris; Wilcox, Julianne; Baeten, Elisabeth; Macmillan, Christine; Cornen, Claude; Baumer, Michael; Simpson, Edwin; Lintott, Chris J.; Miller, David; Paget, Edward; Simpson, Robert; Smith, Arfon M.; Küng, Rafael; Saha, Prasenjit; Collett, Thomas E.
2016-01-01
We describe SPACE WARPS, a novel gravitational lens discovery service that yields samples of high purity and completeness through crowdsourced visual inspection. Carefully produced colour composite images are displayed to volunteers via a web-based classification interface, which records their estimates of the positions of candidate lensed features. Images of simulated lenses, as well as real images which lack lenses, are inserted into the image stream at random intervals; this training set is used to give the volunteers instantaneous feedback on their performance, as well as to calibrate a model of the system that provides dynamical updates to the probability that a classified image contains a lens. Low-probability systems are retired from the site periodically, concentrating the sample towards a set of lens candidates. Having divided 160 deg2 of Canada-France-Hawaii Telescope Legacy Survey imaging into some 430 000 overlapping 82 by 82 arcsec tiles and displaying them on the site, we were joined by around 37 000 volunteers who contributed 11 million image classifications over the course of eight months. This stage 1 search reduced the sample to 3381 images containing candidates; these were then refined in stage 2 to yield a sample that we expect to be over 90 per cent complete and 30 per cent pure, based on our analysis of the volunteers performance on training images. We comment on the scalability of the SPACE WARPS system to the wide field survey era, based on our projection that searches of 105 images could be performed by a crowd of 105 volunteers in 6 d.
Braggio, Simone; Montanari, Dino; Rossi, Tino; Ratti, Emiliangelo
2010-07-01
As a result of their wide acceptance and conceptual simplicity, drug-like concepts are having a major influence on the drug discovery process, particularly in the selection of the 'optimal' absorption, distribution, metabolism, excretion and toxicity and physicochemical parameters space. While they have an undisputable value when assessing the potential of lead series or in evaluating inherent risk of a portfolio of drug candidates, they result much less useful in weighing up compounds for the selection of the best potential clinical candidate. We introduce the concept of drug efficiency as a new tool both to guide the drug discovery program teams during the lead optimization phase and to better assess the developability potential of a drug candidate.
New Zealand’s Drug Development Industry
Lockhart, Michelle Marie; Babar, Zaheer-Ud-Din; Carswell, Christopher; Garg, Sanjay
2013-01-01
The pharmaceutical industry’s profitability depends on identifying and successfully developing new drug candidates while trying to contain the increasing costs of drug development. It is actively searching for new sources of innovative compounds and for mechanisms to reduce the enormous costs of developing new drug candidates. There is an opportunity for academia to further develop as a source of drug discovery. The rising levels of industry outsourcing also provide prospects for organisations that can reduce the costs of drug development. We explored the potential returns to New Zealand (NZ) from its drug discovery expertise by assuming a drug development candidate is out-licensed without clinical data and has anticipated peak global sales of $350 million. We also estimated the revenue from NZ’s clinical research industry based on a standard per participant payment to study sites and the number of industry-sponsored clinical trials approved each year. Our analyses found that NZ’s clinical research industry has generated increasing foreign revenue and appropriate policy support could ensure that this continues to grow. In addition the probability-based revenue from the out-licensing of a drug development candidate could be important for NZ if provided with appropriate policy and financial support. PMID:24065037
Option volatility and the acceleration Lagrangian
NASA Astrophysics Data System (ADS)
Baaquie, Belal E.; Cao, Yang
2014-01-01
This paper develops a volatility formula for option on an asset from an acceleration Lagrangian model and the formula is calibrated with market data. The Black-Scholes model is a simpler case that has a velocity dependent Lagrangian. The acceleration Lagrangian is defined, and the classical solution of the system in Euclidean time is solved by choosing proper boundary conditions. The conditional probability distribution of final position given the initial position is obtained from the transition amplitude. The volatility is the standard deviation of the conditional probability distribution. Using the conditional probability and the path integral method, the martingale condition is applied, and one of the parameters in the Lagrangian is fixed. The call option price is obtained using the conditional probability and the path integral method.
Gettings, Mark E.; Bultman, Mark W.
2005-01-01
Some aquifers of the southwestern Colorado Plateaus Province are deeply buried and overlain by several impermeable shale layers, and so recharge to the aquifer probably is mainly by seepage down penetrative-fracture systems. The purpose of this 2-year study, sponsored by the U.S. National Park Service, was to map candidate deep penetrative fractures over a 120,000-km2 area, using gravity and aeromagnetic-anomaly data together with surficial-fracture data. The study area was on the Colorado Plateau south of the Grand Canyon and west of Black Mesa; mapping was carried out at a scale of 1:250,000. The resulting database constitutes a spatially registered estimate of deep-fracture locations. Candidate penetrative fractures were located by spatial correlation of horizontal- gradient and analytic-signal maximums of gravity and magnetic anomalies with major surficial lineaments obtained from geologic, topographic, side-looking-airborne-radar, and satellite imagery. The maps define a subset of candidate penetrative fractures because of limitations in the data coverage and the analytical technique. In particular, the data and analytical technique used cannot predict whether the fractures are open or closed. Correlations were carried out by using image-processing software, such that every pixel on the resulting images was coded to uniquely identify which datasets are correlated. The technique correctly identified known and many new deep fracture systems. The resulting penetrative-fracture-distribution maps constitute an objectively obtained, repeatable dataset and a benchmark from which additional studies can begin. The maps also define in detail the tectonic fabrics of the southwestern Colorado Plateaus Province. Overlaying the correlated lineaments on the normalized-density-of-vegetation-index image reveals that many of these lineaments correlate with the boundaries of vegetation zones in drainages and canyons and so may be controlling near-surface water availability in some places. Many derivative products can be produced from the database, such as fracture-density-estimate maps, and maps with the number of correlations color-coded to estimate the possible quality of correlation. The database contained in this report is designed to be used in a geographic information system and image-processing systems, and most data layers are in georeferenced tagged image format (Geotiff) or ARC grids. The report includes 163 map plates and various metadata, supporting, and statistical diagram files.
NASA Astrophysics Data System (ADS)
Barrado, D.; de Gregorio Monsalvo, I.; Huélamo, N.; Morales-Calderón, M.; Bayo, A.; Palau, A.; Ruiz, M. T.; Rivière-Marichalar, P.; Bouy, H.; Morata, Ó.; Stauffer, J. R.; Eiroa, C.; Noriega-Crespo, A.
2018-04-01
Aims: The early evolutionary stage of brown dwarfs (BDs) is not very well characterized, especially during the embedded phase. Our goal is to gain insight into the dominant formation mechanism of very low-mass objects and BDs. Methods: We have conducted deep observations at 870 μm obtained with the LABOCA bolometer at the APEX telescope in order to identify young submillimeter (submm) sources in the Barnard 30 dark cloud. We have complemented these data with multi-wavelength observations from the optical to the far-IR and compiled complete spectral energy distributions in order to identify the counterparts, characterize the sources and to assess their membership to the association and stellar or substellar status based on the available photometric information. Results: We have identified 34 submm sources and a substantial number of possible and probable Barnard 30 members within each individual APEX/LABOCA beam. They can be classified into three distinct groups. First, 15 of these 34 have a clear optical or IR counterpart to the submm peak and nine of them are potential proto-BD candidates. Moreover, a substantial number of them could be multiple systems. A second group of 13 sources comprises candidate members with significant infrared excesses located away from the central submm emission. All of them include BD candidates, some displaying IR excess, but their association with submm emission is unclear. In addition, we have found six starless cores and, based on the total dust mass estimate, three might be pre-substellar (or pre-BDs) cores. Finally, the complete characterization of our APEX/LABOCA sources, focusing on those detected at 24 and/or 70 μm, indicates that in our sample of 34 submm sources there are, at least: two WTTs, four CTTs, five young stellar objects, eight proto-BD candidates (with another three dubious cases), and one very low luminosity objects. Conclusions: Our findings provide additional evidence concerning the BD formation mechanism, which seems to be a downsized version of the stellar formation. Tables 3-7 and reduced images (FITS files) are only available at the CDS via anonymous ftp to http://cdsarc.u-strasbg.fr (http://130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/vol/page
On the Structure of a Best Possible Crossover Selection Strategy in Genetic Algorithms
NASA Astrophysics Data System (ADS)
Lässig, Jörg; Hoffmann, Karl Heinz
The paper considers the problem of selecting individuals in the current population in genetic algorithms for crossover to find a solution with high fitness for a given optimization problem. Many different schemes have been described in the literature as possible strategies for this task but so far comparisons have been predominantly empirical. It is shown that if one wishes to maximize any linear function of the final state probabilities, e.g. the fitness of the best individual in the final population of the algorithm, then a best probability distribution for selecting an individual in each generation is a rectangular distribution over the individuals sorted in descending sequence by their fitness values. This means uniform probabilities have to be assigned to a group of the best individuals of the population but probabilities equal to zero to individuals with lower fitness, assuming that the probability distribution to choose individuals from the current population can be chosen independently for each iteration and each individual. This result is then generalized also to typical practically applied performance measures, such as maximizing the expected fitness value of the best individual seen in any generation.
The high-energy X-ray spectrum of black hole candidate GX 339-4 during a transition
NASA Technical Reports Server (NTRS)
Dolan, J. F.; Crannell, C. J.; Dennis, B. R.; Orwig, L. E.
1987-01-01
The X-ray emitting system GX 339-4 contains one of the prime candidates for a stellar mass-sized black hole. Determining the observational similarities and differences between the members of this group is of value in specifying which characteristics can be used to identify systems containing a black hole, especially those for which no mass determination can be made. The first observations of the E greater than 20 keV spectrum of GX 339-4 during a transition between luminosity states are reported here. The hard spectral state is the lower luminosity state of the system. GX 339-4 has a power-low spectrum above 20 keV which pivots during transitions between distinct luminosity states. The only other X-ray sources known to exhibit this behavior, Cyg XR-1 and (probably) A0620-00, are leading candidates for systems containing a black hole component based on their measured spectrocopic mass function.
A Limited Comparison of the Thermal Durability of Polyimide Candidate Matrix Polymers with PMR-15
NASA Technical Reports Server (NTRS)
Bowles, Kenneth J.; Papadopoulos, Demetrios S.; Scheiman, Daniel A.; Inghram, Linda L.; McCorkle, Linda S.; Klans, Ojars V.
2003-01-01
Studies were conducted with six different candidate high-temperature neat matrix resin specimens of varied geometric shapes to investigate the mechanisms involved in the thermal degradation of polyimides like PMR-15. The metrics for assessing the quality of these candidates were chosen to be glass transition temperature (T(sub g)), thermo-oxidative stability, dynamic mechanical properties, microstructural changes, and dimensional stability. The processing and mechanical properties were not investigated in the study reported herein. The dimensional changes and surface layer growth were measured and recorded. The data were in agreement with earlier published data. An initial weight increase reaction was observed to be dominating at the lower temperatures. However, at the more elevated temperatures, the weight loss reactions were prevalent and probably masked the weight gain reaction. These data confirmed the findings of the existence of an initial weight gain reaction previously reported. Surface- and core-dependent weight losses were shown to control the polymer degradation at the higher temperatures.
High throughput nonparametric probability density estimation.
Farmer, Jenny; Jacobs, Donald
2018-01-01
In high throughput applications, such as those found in bioinformatics and finance, it is important to determine accurate probability distribution functions despite only minimal information about data characteristics, and without using human subjectivity. Such an automated process for univariate data is implemented to achieve this goal by merging the maximum entropy method with single order statistics and maximum likelihood. The only required properties of the random variables are that they are continuous and that they are, or can be approximated as, independent and identically distributed. A quasi-log-likelihood function based on single order statistics for sampled uniform random data is used to empirically construct a sample size invariant universal scoring function. Then a probability density estimate is determined by iteratively improving trial cumulative distribution functions, where better estimates are quantified by the scoring function that identifies atypical fluctuations. This criterion resists under and over fitting data as an alternative to employing the Bayesian or Akaike information criterion. Multiple estimates for the probability density reflect uncertainties due to statistical fluctuations in random samples. Scaled quantile residual plots are also introduced as an effective diagnostic to visualize the quality of the estimated probability densities. Benchmark tests show that estimates for the probability density function (PDF) converge to the true PDF as sample size increases on particularly difficult test probability densities that include cases with discontinuities, multi-resolution scales, heavy tails, and singularities. These results indicate the method has general applicability for high throughput statistical inference.
High throughput nonparametric probability density estimation
Farmer, Jenny
2018-01-01
In high throughput applications, such as those found in bioinformatics and finance, it is important to determine accurate probability distribution functions despite only minimal information about data characteristics, and without using human subjectivity. Such an automated process for univariate data is implemented to achieve this goal by merging the maximum entropy method with single order statistics and maximum likelihood. The only required properties of the random variables are that they are continuous and that they are, or can be approximated as, independent and identically distributed. A quasi-log-likelihood function based on single order statistics for sampled uniform random data is used to empirically construct a sample size invariant universal scoring function. Then a probability density estimate is determined by iteratively improving trial cumulative distribution functions, where better estimates are quantified by the scoring function that identifies atypical fluctuations. This criterion resists under and over fitting data as an alternative to employing the Bayesian or Akaike information criterion. Multiple estimates for the probability density reflect uncertainties due to statistical fluctuations in random samples. Scaled quantile residual plots are also introduced as an effective diagnostic to visualize the quality of the estimated probability densities. Benchmark tests show that estimates for the probability density function (PDF) converge to the true PDF as sample size increases on particularly difficult test probability densities that include cases with discontinuities, multi-resolution scales, heavy tails, and singularities. These results indicate the method has general applicability for high throughput statistical inference. PMID:29750803