A check-up for the statistical Parton model
NASA Astrophysics Data System (ADS)
Buccella, Franco; Sohaily, Sozha
2015-11-01
We compare the Parton distributions deduced in the framework of a quantum statistical approach for both the longitudinal and transverse degrees of freedom with the unpolarized distributions measured at HERA and with the polarized ones proposed in a previous paper, which have been shown to be in very good agreement also with the results of experiments performed after that proposal. The agreement with HERA data in correspondence to very similar values for the “temperature” and the “potentials” found in the previous work gives a robust confirm of the statistical model. The unpolarized distributions are compared also with the result of NNPDF. The free parameters are fixed mainly by data in the range (0.1, 0.5) for the x variable, where the valence Partons dominate, and in the small x region for the diffractive contribution. This feature makes the parametrization proposed here very attractive.
Strangeness asymmetry of the nucleon in the statistical parton model
NASA Astrophysics Data System (ADS)
Bourrely, Claude; Soffer, Jacques; Buccella, Franco
2007-04-01
We extend to the strange quarks and antiquarks, the statistical approach of parton distributions and we calculate the strange quark asymmetry s -sbar. We find that the asymmetry is small, positive in the low x region and negative in the high x region. In this framework, the polarized strange quarks and antiquarks distributions, which are obtained simultaneously, are found to be both negative for all x values.
A statistical approach for polarized parton distributions
NASA Astrophysics Data System (ADS)
Bourrely, C.; Soffer, J.; Buccella, F.
2002-04-01
A global next-to-leading order QCD analysis of unpolarized and polarized deep-inelastic scattering data is performed with parton distributions constructed in a statistical physical picture of the nucleon. The chiral properties of QCD lead to strong relations between quarks and antiquarks distributions and the importance of the Pauli exclusion principle is also emphasized. We obtain a good description, in a broad range of x and Q^2, of all measured structure functions in terms of very few free parameters. We stress the fact that at RHIC-BNL the ratio of the unpolarized cross sections for the production of W^+ and W^- in pp collisions will directly probe the behavior of the bar d(x) / bar u(x) ratio for x ≥ 0.2, a definite and important test for the statistical model. Finally, we give specific predictions for various helicity asymmetries for the W^±, Z production in pp collisions at high energies, which will be measured with forthcoming experiments at RHIC-BNL and which are sensitive tests of the statistical model for Δ bar u(x) and Δ bar d(x).
The statistical parton distributions: status and prospects
NASA Astrophysics Data System (ADS)
Bourrely, C.; Soffer, J.; Buccella, F.
2005-06-01
New experimental results on polarized structure functions, cross sections for e^{±}p neutral and charge current reactions and ν (bar{ν}) charge current on isoscalar targets are compared with predictions using the statistical parton distributions, which were previously determined. New data on cross sections for Drell-Yan processes, single-jet data in pbar{p} collisions and inclusive π^0 production data in pp collisions are also compared with predictions from this theoretical approach. The good agreement which we find with all these tests against experiment strengthens our opinion on the relevance of the role of quantum statistics for parton distributions. We will also discuss the prospects of this physical framework.
NASA Astrophysics Data System (ADS)
Dean, David Jarvis
1991-02-01
The purpose of this dissertation is to develop a dynamical 3 + 1-dimensional model of interacting hadrons in relativistic collisions. The model incorporates the valence quark structure functions of the hadrons into the dynamical Nambu-Goto string picture. The nucleon is viewed as an ensemble average of various initial string configurations such that the flavor averaged valence quark structure function is reproduced. A stochastic decay mechanism is also developed and applied to string fragmentation (hadronization). The interaction e^+e^-togamma ^{*}to q| q is studied at energies from sqrt{s} = 14 to 30 GeV, and decay parameters are chosen such that the correct experimental multiplicity of particles is obtained. Transverse momentum production is obtained by dynamically generating q| q pairs according to a phenomenological momentum distribution. The interaction mechanism between two colliding nucleons is based on a quark-quark scattering and exchange. The quark scattering cross section is parameterized to reproduce the experimental results. The interactions coupled with the hadronization mechanism successfully reproduce many of the observed inclusive distributions. These include, the charged particle, rapidity, scaled parallel momentum, and p_| distributions. At the present stage of numerical calculations p_ | < 1.1 GeV region has been studied. The model interaction is capable of investigating higher p_| values, which require better statistics and more computing time. Using this interaction, pp collisions at sqrt{s} = 19.4 and 53 GeV are studied and reasonable fits to data are obtained. A further application of the model involves the study of the nuclear attenuation effects observed in e^-A when compared to e ^-p collision experiments. These effects are also observed in the string-parton calculation. At energies of v < 10 GeV the nuclear medium influences the hadronization process. At higher energies the effect is negligible.
Modeling Nucleon Generalized Parton Distributions
Radyushkin, Anatoly V.
2013-05-01
We discuss building models for nucleon generalized parton distributions (GPDs) H and E that are based on the formalism of double distributions (DDs). We find that the usual "DD+D-term'' construction should be amended by an extra term, generated by GPD E(x,\\xi). Unlike the $D$-term, this function has support in the whole -1 < x< 1 region, and in general does not vanish at the border points|x|=\\xi.
Recent Tests for the Statistical Parton Distributions
NASA Astrophysics Data System (ADS)
Bourrely, Claude; Soffer, Jacques; Buccella, Franco
We compare some recent experimental results obtained at DESY, SLAC and Jefferson Lab., with the predictions of the statistical model, we have previously proposed. The result of this comparison is very satisfactory.
Modeling Nucleon Generalized Parton Distributions
Radyushkin, Anatoly V.
2013-05-01
We discuss building models for nucleon generalized parton distributions (GPDs) H and E that are based on the formalism of double distributions (DDs). We found that the usual "DD+D-term" construction should be amended by an extra term, xiE^1_+ (x,xi) built from the alpha/Beta moment of the DD e(Beta,alpha) that generates GPD E(x,xi). Unlike the D-term, this function has support in the whole -1< x<1 region, and in general does not vanish at the border points |x|=xi.
New Results in the Quantum Statistical Approach to Parton Distributions
NASA Astrophysics Data System (ADS)
Soffer, Jacques; Bourrely, Claude; Buccella, Franco
2015-02-01
We will describe the quantum statistical approach to parton distributions allowing to obtain simultaneously the unpolarized distributions and the helicity distributions. We will present some recent results, in particular related to the nucleon spin structure in QCD. Future measurements are challenging to check the validity of this novel physical framework.
Quantum Statistical Parton Distributions and the Spin Crisis
NASA Astrophysics Data System (ADS)
Buccella, F.; Miele, G.; Tancredi, N.
1996-10-01
Quantum statistical distributions for partons provide a fair description of deep inelastic scattering data at Q2 = 3 and 10 (GeV/c)2. Study of the polarized structure functions seems to suggest an alternative possible solution of the spin crisis based on the Pauli principle. In this scheme, in fact, it becomes apparent that the defects of the Gottfried sum rule and Ellis-Jaffe sum rule for the proton are strongly connected. This possibility finds particular evidence from the phenomenological observation that the relation Δu = 2$tilde{F} + u - d - 1 seems to be satisfied well by parton distributions.
The Transverse Momentum Dependent Statistical Parton Distributions Revisited
NASA Astrophysics Data System (ADS)
Bourrely, Claude; Buccella, Franco; Soffer, Jacques
2013-04-01
The extension of the statistical parton distributions to include their transverse momentum dependence (TMD) is revisited by considering that the proton target has a finite longitudinal momentum. The TMD will be generated by means of a transverse energy sum rule. The new results are mainly relevant for electron-proton inelastic collisions in the low Q2 region. We take into account the effects of the Melosh-Wigner rotation for the helicity distributions.
W± bosons production in the quantum statistical parton distributions approach
NASA Astrophysics Data System (ADS)
Bourrely, Claude; Buccella, Franco; Soffer, Jacques
2013-10-01
We consider W± gauge bosons production in connection with recent results from BNL-RHIC and FNAL-Tevatron and interesting predictions from the statistical parton distributions. They concern relevant aspects of the structure of the nucleon sea and the high-x region of the valence quark distributions. We also give predictions in view of future proton-neutron collisions experiments at BNL-RHIC.
Modeling the Pion Generalized Parton Distribution
NASA Astrophysics Data System (ADS)
Mezrag, C.
2016-02-01
We compute the pion Generalized Parton Distribution (GPD) in a valence dressed quarks approach. We model the Mellin moments of the GPD using Ansätze for Green functions inspired by the numerical solutions of the Dyson-Schwinger Equations (DSE) and the Bethe-Salpeter Equation (BSE). Then, the GPD is reconstructed from its Mellin moment using the Double Distribution (DD) formalism. The agreement with available experimental data is very good.
The partonic interpretation of reggeon theory models
NASA Astrophysics Data System (ADS)
Boreskov, K. G.; Kaidalov, A. B.; Khoze, V. A.; Martin, A. D.; Ryskin, M. G.
2005-12-01
We review the physical content of the two simplest models of reggeon field theory: namely the eikonal and the Schwimmer models. The AGK cutting rules are used to obtain the inclusive, the inelastic and the diffractive cross sections. The system of non-linear equations for these cross sections is written down and analytic expressions for its solution are obtained. We derive the rapidity gap dependence of the differential cross sections for diffractive dissociation in the Schwimmer model and in its eikonalized extension. The results are interpreted from the partonic viewpoint of the interaction at high energies.
How large is the gluon polarization in the statistical parton distributions approach?
Soffer, Jacques; Bourrely, Claude; Buccella, Franco
2015-04-10
We review the theoretical foundations of the quantum statistical approach to parton distributions and we show that by using some recent experimental results from Deep Inelastic Scattering, we are able to improve the description of the data by means of a new determination of the parton distributions. We will see that a large gluon polarization emerges, giving a significant contribution to the proton spin.
How large is the gluon polarization in the statistical parton distributions approach?
NASA Astrophysics Data System (ADS)
Soffer, Jacques; Bourrely, Claude; Buccella, Franco
2015-04-01
We review the theoretical foundations of the quantum statistical approach to parton distributions and we show that by using some recent experimental results from Deep Inelastic Scattering, we are able to improve the description of the data by means of a new determination of the parton distributions. We will see that a large gluon polarization emerges, giving a significant contribution to the proton spin.
The Polarized TMDs in the covariant parton model approach
A.V. Efremov, P. Schweitzer, O.V. Teryaev, P. Zavada
2011-05-01
We derive relations between polarized transverse momentum dependent distribution functions (TMDs) and the usual parton distribution functions (PDFs) in the 3D covariant parton model, which follow from Lorentz invariance and the assumption of a rotationally symmetric distribution of parton momenta in the nucleon rest frame. Using the known PDF $g_{1}^{q}(x)$ as input we predict the $x$- and $\\mathbf{p}_{T}$-dependence of all polarized twist-2 naively time-reversal even (T-even) TMDs.
New model for nucleon generalized parton distributions
Radyushkin, Anatoly V.
2014-01-01
We describe a new type of models for nucleon generalized parton distributions (GPDs) H and E. They are heavily based on the fact nucleon GPDs require to use two forms of double distribution (DD) representations. The outcome of the new treatment is that the usual DD+D-term construction should be amended by an extra term, {xi} E{sub +}{sup 1} (x,{xi}) which has the DD structure {alpha}/{beta} e({beta},{alpha}, with e({beta},{alpha}) being the DD that generates GPD E(x,{xi}). We found that this function, unlike the D-term, has support in the whole -1 <= x <= 1 region. Furthermore, it does not vanish at the border points |x|={xi}.
Fermi-Dirac statistics plus liquid description of quark partons
NASA Astrophysics Data System (ADS)
Buccella, F.; Miele, G.; Migliore, G.; Tibullo, V.
1995-12-01
A previous approach with Fermi-Dirac distributions for fermion partons is here improved to comply with the expected low x behaviour of structure functions. We are so able to get a fair description of the unpolarized and polarized structure functions of the nucleons as well as of neutrino data. We cannot reach definite conclusions, but confirm our suspicion of a relationship between the defects in Gottfried and spin sum rules.
Generalized Valon Model for Double Parton Distributions
NASA Astrophysics Data System (ADS)
Broniowski, Wojciech; Ruiz Arriola, Enrique; Golec-Biernat, Krzysztof
2016-03-01
We show how the double parton distributions may be obtained consistently from the many-body light-cone wave functions. We illustrate the method on the example of the pion with two Fock components. The procedure, by construction, satisfies the Gaunt-Stirling sum rules. The resulting single parton distributions of valence quarks and gluons are consistent with a phenomenological parametrization at a low scale.
Generalized Valon Model for Double Parton Distributions
NASA Astrophysics Data System (ADS)
Broniowski, Wojciech; Ruiz Arriola, Enrique; Golec-Biernat, Krzysztof
2016-06-01
We show how the double parton distributions may be obtained consistently from the many-body light-cone wave functions. We illustrate the method on the example of the pion with two Fock components. The procedure, by construction, satisfies the Gaunt-Stirling sum rules. The resulting single parton distributions of valence quarks and gluons are consistent with a phenomenological parametrization at a low scale.
A.V. Efremov, P. Schweitzer, O.V. Teryaev, P. Zavada
2011-03-01
We derive relations between transverse momentum dependent distribution functions (TMDs) and the usual parton distribution functions (PDFs) in the 3D covariant parton model, which follow from Lorentz invariance and the assumption of a rotationally symmetric distribution of parton momenta in the nucleon rest frame. Using the known PDFs f_1(x) and g_1(x) as input we predict the x- and pT-dependence of all twist-2 T-even TMDs.
Efremov, A. V.; Teryaev, O. V.; Schweitzer, P.; Zavada, P.
2011-03-01
We derive relations between transverse momentum dependent distribution functions and the usual parton distribution functions in the 3D covariant parton model, which follow from Lorentz invariance and the assumption of a rotationally symmetric distribution of parton momenta in the nucleon rest frame. Using the known parton distribution functions f{sub 1}{sup a}(x) and g{sub 1}{sup a}(x) as input we predict the x- and p{sub T}-dependence of all twist-2 T-even transverse momentum dependent distribution functions.
Nonperturbative approach to the parton model
NASA Astrophysics Data System (ADS)
Simonov, Yu. A.
2016-02-01
In this paper, the nonperturbative parton distributions, obtained from the Lorentz contracted wave functions, are analyzed in the formalism of many-particle Fock components and their properties are compared to the standard perturbative distributions. We show that the collinear and IR divergencies specific for perturbative evolution treatment are absent in the nonperturbative version, however for large momenta pi2 ≫ σ (string tension), the bremsstrahlung kinematics is restored. A preliminary discussion of possible nonperturbative effects in DIS and high energy scattering is given, including in particular a possible role of multihybrid states in creating ridge-type effects.
Independent pair parton interactions model of hadron interactions
NASA Astrophysics Data System (ADS)
Dremin, I. M.; Nechitailo, V. A.
2004-08-01
A model of independent pair parton interactions is proposed, according to which hadron interactions are represented by a set of independent binary parton collisions. The final multiplicity distribution is described by a convolution of the negative binomial distributions in each of the partonic collisions. As a result, it is given by a weighted sum of negative binomial distributions with parameters multiplied by the number of active pairs. Its shape and moments are considered. Experimental data on multiplicity distributions in high energy pp¯ processes are well fitted by these distributions. Predictions for the CERN Large Hadron Collider and higher energies are presented. The difference between e+e- and pp¯ processes is discussed.
Diffraction scattering and the parton model in QCD
White, A.
1985-01-01
Arguments are presented that the validity of the parton model for hadron scattering in QCD is directly related to the occurrence of the Critical Pomeron description of diffraction scattering. An attractive route suggested for Electroweak and Grand Unification is also briefly described.
NASA Astrophysics Data System (ADS)
Bourrely, Claude; Buccella, Franco; Soffer, Jacques
2011-04-01
We consider the extension of the statistical parton distributions to include their transverse momentum dependence, by using two different methods, one is based on our quantum statistical approach, the other on a relativistic covariant method. We take into account the effects of the Melosh-Wigner rotation for the polarized distributions. The results obtained can be compared with recent semi-inclusive deep inelastic scattering (DIS) data on the cross section and double longitudinal-spin asymmetries from JLab. We also give some predictions for future experiments on electron-neutron scattering.
Projective symmetry of partons in Kitaev's honeycomb model
NASA Astrophysics Data System (ADS)
Mellado, Paula
2015-03-01
Low-energy states of quantum spin liquids are thought to involve partons living in a gauge-field background. We study the spectrum of Majorana fermions of Kitaev's honeycomb model on spherical clusters. The gauge field endows the partons with half-integer orbital angular momenta. As a consequence, the multiplicities reflect not the point-group symmetries of the cluster, but rather its projective symmetries, operations combining physical and gauge transformations. The projective symmetry group of the ground state is the double cover of the point group. We acknowledge Fondecyt under Grant No. 11121397, Conicyt under Grant No. 79112004, and the Simons Foundation (P.M.); the Max Planck Society and the Alexander von Humboldt Foundation (O.P.); and the US DOE Grant No. DE-FG02-08ER46544 (O.T.).
Efremov, A. V.; Teryaev, O. V.; Schweitzer, P.; Zavada, P.
2009-07-01
Transverse parton momentum dependent distribution functions (TMDs) of the nucleon are studied in a covariant model, which describes the intrinsic motion of partons in terms of a covariant momentum distribution. The consistency of the approach is demonstrated, and model relations among TMDs are studied. As a by-product it is shown how the approach allows to formulate the nonrelativistic limit.
Frederico, T.; Pace, E.; Pasquini, B.; Salme, G.
2010-08-05
Longitudinal and transverse parton distributions for pion and nucleon are calculated from hadron vertexes obtained by a study of form factors within relativistic quark models. The relevance of the one-gluon-exchange dominance at short range for the behavior of the form factors at large momentum transfer and of the parton distributions at the end points is stressed.
Backward dilepton production in color dipole and parton models
Gay Ducati, Maria Beatriz; Graeve de Oliveira, Emmanuel
2010-03-01
The Drell-Yan dilepton production at backward rapidities is studied in proton-nucleus collisions at Relativistic Heavy Ion Collider and LHC energies by comparing two different approaches: the k{sub T} factorization at next-to-leading order with intrinsic transverse momentum and the same process formulated in the target rest frame, i.e., the color dipole approach. Our results are expressed in terms of the ratio between p(d)-A and p-p collisions as a function of transverse momentum and rapidity. Three nuclear parton distribution functions are used: EKS (Eskola, Kolhinen, and Ruuskanen), EPS08, and EPS09 and, in both approaches, dileptons show sensitivity to nuclear effects, specially regarding the intrinsic transverse momentum. Also, there is room to discriminate between formalisms: the color dipole approach lacks soft effects introduced by the intrinsic k{sub T}. Geometric scaling GBW (Golec-Biernat and Wusthoff) and BUW (Boer, Utermann, and Wessels) color dipole cross section models and also a DHJ (Dumitru, Hayashigaki, and Jalilian-Marian) model, which breaks geometric scaling, are used. No change in the ratio between collisions is observed, showing that this observable is not changed by the particular shape of the color dipole cross section. Furthermore, our k{sub T} factorization results are compared with color glass condensate results at forward rapidities: the results agree at Relativistic Heavy Ion Collider although disagree at LHC, mainly due to the different behavior of target gluon and quark shadowing.
Comparing multiparticle production within a two-component dual parton model with collider data
Hahn, K.; Ranft, J. )
1990-03-01
The dual parton model (DPM) is very successful in describing hadronic multiparticle production. The version of DPM presented includes both soft and hard mechanisms. The hard component is described according to the lowest-order perturbative QCD--parton-model cross section. The model is formulated in the form of a Monte Carlo event generator. Results obtained with this event generator are compared with data on inclusive reactions in the TeV energy range of the CERN and Fermilab hadron colliders.
Parton model for hA and AA collisions at high energies
NASA Astrophysics Data System (ADS)
Braun, M. A.
1991-02-01
The parton model for hA and AA interactions is developed for arbitrary dependence of parton amplitudes on energy. Conditions are studied under which the Glauber formula results for total cross sections. The fulfillment of the AGK rules is shown for all energies and registered particle momenta. Inclusive A'A cross sections in the forward hemisphere prove to be A' times larger than for NA collisions.
Charm quark energy loss in infinite QCD matter using a parton cascade model
NASA Astrophysics Data System (ADS)
Younus, Mohammed; Coleman-Smith, Christopher E.; Bass, Steffen A.; Srivastava, Dinesh K.
2015-02-01
We utilize the parton cascade model to study the evolution of charm quarks propagating through a thermal brick of QCD matter. We determine the energy loss and the transport coefficient q ̂ for charm quarks. The calculations are done at a constant temperature of 350 MeV and the results are compared to analytical calculations of heavy-quark energy loss in order to validate the applicability of using a parton cascade model for the study of heavy-quark dynamics in hot and dense QCD matter.
Pion generalized parton distributions within a fully covariant constituent quark model
NASA Astrophysics Data System (ADS)
Fanelli, Cristiano; Pace, Emanuele; Romanelli, Giovanni; Salmè, Giovanni; Salmistraro, Marco
2016-05-01
We extend the investigation of the generalized parton distribution for a charged pion within a fully covariant constituent quark model, in two respects: (1) calculating the tensor distribution and (2) adding the treatment of the evolution, needed for achieving a meaningful comparison with both the experimental parton distribution and the lattice evaluation of the so-called generalized form factors. Distinct features of our phenomenological covariant quark model are: (1) a 4D Ansatz for the pion Bethe-Salpeter amplitude, to be used in the Mandelstam formula for matrix elements of the relevant current operators, and (2) only two parameters, namely a quark mass assumed to be m_q=~220 MeV and a free parameter fixed through the value of the pion decay constant. The possibility of increasing the dynamical content of our covariant constituent quark model is briefly discussed in the context of the Nakanishi integral representation of the Bethe-Salpeter amplitude.
Diphoton production in the ADD model to NLO + parton shower accuracy at the LHC
NASA Astrophysics Data System (ADS)
Frederix, R.; Mandal, Manoj K.; Mathews, Prakash; Ravindran, V.; Seth, Satyajit; Torrielli, P.; Zaro, M.
2012-12-01
In this paper, we present the next-to-leading order predictions for diphoton production in the ADD model, matched to the HERWIG parton shower using the MC@NLO formalism. A selection of the results is presented for d = 2-6 extra dimensions, using generic cuts as well as analysis cuts mimicking the search strategies as pursued by the ATLAS and CMS experiments.
The Extension to the Transverse Momentum of the Statistical Parton Distributions
NASA Astrophysics Data System (ADS)
Buccella, F.
2006-02-01
The extension of the statistical approach to the transverse degrees of freedhom explains a moltiplicative factor, we were obliged to introduce in a previous work to comply with experiment for the Fermi-Dirac functions of the light quarks. It is possible to get light antiquark distributions similar to the ones proposed there.
The Extension to the Transverse Momentum of the Statistical Parton Distributions
NASA Astrophysics Data System (ADS)
Bourrely, Claude; Soffer, Jacques; Buccella, Franco
By extending the statistical distributions to the transverse degree of freedom, we account for a multiplicative factor in the Fermi-Dirac functions of the light quarks, we were led to introduce in a previous work to comply with experiment. We can also get light antiquark distributions, similar to those we proposed earlier.
Low-P/sub T/ hadron production and a valon-parton recombination model
Amiri, F.
1981-01-01
A variant of the recombination model which we call the valon-parton model is applied simultaneously to a variety of meson inclusive reactions with proton, pion and kaon beams in the kinematic region of low transverse momentum and intermediate values of longitudinal momentum fractions. It is found that the valon distributions in hadrons show no evidence for SU(3) breaking. There are some indications of substantial gluon dissociation contributions which we interpreted through a maximally enhanced sea. For proton induced reactions the model predictions are in excellent agreement with the data; meson initiated reactions indicate additional contributions are coming from resonances which are produced recombinantly and then decay into the observed mesons.
Low-P/sub T/ hadron production and a valon-parton recombination model
Amiri, F.
1981-12-01
A variant of the recombination model which we call the valon-parton model is applied simultaneously to a variety of meson inclusive reactions with proton, pion and kaon beams in the kinematic region of low transverse momentum and intermediate values of longitudinal momentum fractions. It is found that the valon distributions in hadrons show no evidence for SU(3) breaking. There are some indications of substantial gluon dissociation contributions which we interpreted through a maximally enhanced sea. For proton induced reactions the model predictions are in excellent agreement with the data; meson initiated reactions indicate additional contributions are coming from resonances which are produced recombinantly and then decay into the observed mesons.
Particle Identification in the Dynamical String-Parton Model of Relativistic Heavy-Ion Collisions
NASA Astrophysics Data System (ADS)
Malov, D. E.; Umar, A. S.; Ernst, D. J.; Dean, D. J.
The dynamical string-parton model for relativistic heavy-ion collisions is generalized to include particle identification of the final-state hadrons by phenomenologically quantizing the masses of the classical strings which result from string breaking. General features of the Nambu-Gotō strings are used to motivate a model that identifies a mass window near the physical mass of a meson, and does not allow the string to decay further if its mass falls within the window. Data from e+e- collisions in the region √ {s} =10 to 30 GeV are well reproduced by this model.
Energy loss in a partonic transport model including bremsstrahlung processes
Fochler, Oliver; Greiner, Carsten; Xu Zhe
2010-08-15
A detailed investigation of the energy loss of gluons that traverse a thermal gluonic medium simulated within the perturbative QCD-based transport model BAMPS (a Boltzmann approach to multiparton scatterings) is presented in the first part of this work. For simplicity the medium response is neglected in these calculations. The energy loss from purely elastic interactions is compared with the case where radiative processes are consistently included based on the matrix element by Gunion and Bertsch. From this comparison, gluon multiplication processes gg{yields}ggg are found to be the dominant source of energy loss within the approach employed here. The consequences for the quenching of gluons with high transverse momentum in fully dynamic simulations of Au+Au collisions at the BNL Relativistic Heavy Ion Collider (RHIC) energy of {radical}(s)=200A GeV are discussed in the second major part of this work. The results for central collisions as discussed in a previous publication are revisited, and first results on the nuclear modification factor R{sub AA} for noncentral Au+Au collisions are presented. They show a decreased quenching compared to central collisions while retaining the same shape. The investigation of the elliptic flow v{sub 2} is extended up to nonthermal transverse momenta of 10 GeV, exhibiting a maximum v{sub 2} at roughly 4 to 5 GeV and a subsequent decrease. Finally the sensitivity of the aforementioned results on the specific implementation of the effective modeling of the Landau-Pomeranchuk-Migdal (LPM) effect via a formation-time-based cutoff is explored.
NASA Astrophysics Data System (ADS)
Gavai, Rajiv V.; Venugopalan, Raju
1996-11-01
Using weak coupling methods McLerran and Venugopalan expressed the parton distributions in large nuclei as correlation functions of a two-dimensional Euclidean field theory. The theory has the dimensionful coupling g2μ, where μ2~A1/3 is the valence quark color charge squared per unit area. We use a lattice regularization to investigate these correlation functions both analytically and numerically for the simplified case of SU(2) gauge theory. In weak coupling (g2μL<<5), where L is the transverse size of the nucleus, the numerical results agree with the analytic lattice weak coupling results. For g2μL>>5, no solutions exist at O(a4) where a is the lattice spacing. This suggests an ill-defined infrared behavior for the two-dimensional theory. A recent proposal of Jalilian-Marian, Kovner, McLerran, and Weigert for an analytic solution of the classical problem is discussed briefly.
Multiplicity distributions up to √s = 540 GeV in the dual parton model
NASA Astrophysics Data System (ADS)
Capella, A.; Tran Thanh van, J.
1982-08-01
We compute the average charge multiplicities and dispersions in proton-proton and antiproton-proton interactions up to SPS collider energies in the framework of a multi-chain dual parton model. The corresponding data for deep inelastic lepton-proton and e+e- reactions are used as the sole input. The height of the central plateau is also computed. Laboratoire associé au Centre National de la Recherche Scientifique. Postal address: Bâtiment 211, Université Paris-Sud 91405 Orsay, France.
Transverse momentum dependent parton distributions in a light-cone quark model
NASA Astrophysics Data System (ADS)
Pasquini, B.; Cazzaniga, S.; Boffi, S.
2008-08-01
The leading twist transverse momentum dependent parton distributions (TMDs) are studied in a light-cone description of the nucleon where the Fock expansion is truncated to consider only valence quarks. General analytic expressions are derived in terms of the six amplitudes needed to describe the three-quark sector of the nucleon light-cone wave function. Numerical calculations for the T-even TMDs are presented in a light-cone constituent quark model, and the role of the so-called pretzelosity is investigated to produce a nonspherical shape of the nucleon.
Thermalization of parton spectra in the colour-flux-tube model
NASA Astrophysics Data System (ADS)
Ryblewski, Radoslaw
2016-09-01
A detailed study of thermalization of the momentum spectra of partons produced via decays of colour flux tubes due to the Schwinger tunnelling mechanism is presented. The collisions between particles are included in the relaxation-time approximation specified by different values of the shear viscosity to entropy density ratio. At first we show that, to a good approximation, the transverse-momentum spectra of the produced partons are exponential, irrespective of the assumed value of the viscosity of the system and the freeze-out time. This thermal-like behaviour may be attributed to specific properties of the Schwinger tunnelling process. In the next step, in order to check the approach of the system towards genuine local equilibrium, we compare the local slope of the model transverse-momentum spectra with the local slope of the fully equilibrated reference spectra characterized by the effective temperature that reproduces the energy density of the system. We find that the viscosity corresponding to the anti-de Sitter/conformal field theory lower bound is necessary for thermalization of the system within about two fermis.
Low-p/sub T/ hadron production and a valon-parton recombination model
Amiri, F.; Williams, P.K.
1981-11-01
We apply a variant of the recombination model which we call the ''valon-parton model'' simultaneously to a variety of meson inclusive reactions with proton, pion, and kaon beams in the kinematic region of low transverse momentum and intermediate values of longitudinal-momentum fractions. We find that the valon distributions in hadrons show no evidence for SU(3) breaking. There are some indications of substantial gluon-dissociation contributions which we interpret through a ''maximally enhanced sea.'' For proton-induced reactions the model predictions are in excellent agreement with data; meson-induced reactions indicate additional contributions are coming from resonances which are produced recombinantly and then decay into the observed mesons.
An upgraded issue of the parton and hadron cascade model, PACIAE 2.2
NASA Astrophysics Data System (ADS)
Zhou, Dai-Mei; Yan, Yu-Liang; Li, Xing-Long; Li, Xiao-Mei; Dong, Bao-Guo; Cai, Xu; Sa, Ben-Hao
2015-08-01
The parton and hadron cascade model PACIAE 2.1 (cf. Comput. Phys. Commun. 184 (2013) 1476) has been upgraded to the new issue of PACIAE 2.2. By this new issue the lepton-nucleon and lepton-nucleus (inclusive) deep inelastic scatterings can also be investigated. As an example, the PACIAE 2.2 model is enabled to calculate the specific charged hadron multiplicity in the e-+p and e-+D semi-inclusive deep-inelastic scattering at 27.6 GeV electron beam energy. The calculated results are well comparing with the corresponding HERMES data. Additionally, the effect of model parameters α and β in the Lund string fragmentation function on the multiplicity is studied.
Parton distribution in pseudoscalar mesons with a light-front constituent quark model
NASA Astrophysics Data System (ADS)
de Melo, J. P. B. C.; Ahmed, Isthiaq; Tsushima, Kazuo
2016-05-01
We compute the distribution amplitudes of the pion and kaon in the light-front constituent quark model with the symmetric quark-bound state vertex function [1, 2, 3]. In the calculation we explicitly include the flavor-SU(3) symmetry breaking effect in terms of the constituent quark masses of the up (down) and strange quarks. To calculate the kaon parton distribution functions (PDFs), we use both the conditions in the light-cone wave function, i.e., when s ¯ quark is on-shell, and when u quark is on-shell, and make a comparison between them. The kaon PDFs calculated in the two different conditions clearly show asymmetric behaviour due to the flavor SU(3)-symmetry breaking implemented by the quark masses [4, 5].
Charge symmetry at the partonic level
Londergan, J. T.; Peng, J. C.; Thomas, A. W.
2010-07-01
This review article discusses the experimental and theoretical status of partonic charge symmetry. It is shown how the partonic content of various structure functions gets redefined when the assumption of charge symmetry is relaxed. We review various theoretical and phenomenological models for charge symmetry violation in parton distribution functions. We summarize the current experimental upper limits on charge symmetry violation in parton distributions. A series of experiments are presented, which might reveal partonic charge symmetry violation, or alternatively might lower the current upper limits on parton charge symmetry violation.
Are partons confined tachyons?
Noyes, H.P.
1996-03-01
The author notes that if hadrons are gravitationally stabilized ``black holes``, as discrete physics suggests, it is possible that partons, and in particular quarks, could be modeled as tachyons, i.e. particles having v{sup 2} > c{sup 2}, without conflict with the observational fact that neither quarks nor tachyons have appeared as ``free particles``. Some consequences of this model are explored.
Statistical validation of system models
Barney, P.; Ferregut, C.; Perez, L.E.; Hunter, N.F.; Paez, T.L.
1997-01-01
It is common practice in system analysis to develop mathematical models for system behavior. Frequently, the actual system being modeled is also available for testing and observation, and sometimes the test data are used to help identify the parameters of the mathematical model. However, no general-purpose technique exists for formally, statistically judging the quality of a model. This paper suggests a formal statistical procedure for the validation of mathematical models of systems when data taken during operation of the system are available. The statistical validation procedure is based on the bootstrap, and it seeks to build a framework where a statistical test of hypothesis can be run to determine whether or not a mathematical model is an acceptable model of a system with regard to user-specified measures of system behavior. The approach to model validation developed in this study uses experimental data to estimate the marginal and joint confidence intervals of statistics of interest of the system. These same measures of behavior are estimated for the mathematical model. The statistics of interest from the mathematical model are located relative to the confidence intervals for the statistics obtained from the experimental data. These relative locations are used to judge the accuracy of the mathematical model. An extension of the technique is also suggested, wherein randomness may be included in the mathematical model through the introduction of random variable and random process terms. These terms cause random system behavior that can be compared to the randomness in the bootstrap evaluation of experimental system behavior. In this framework, the stochastic mathematical model can be evaluated. A numerical example is presented to demonstrate the application of the technique.
NASA Astrophysics Data System (ADS)
Tuppan, Sam; Budnik, Garrett; Fox, Jordan
2014-09-01
The Meson Cloud Model (MCM) has proven to be a natural explanation for strangeness in the proton because of meson-baryon splitting into kaon-hyperon pairs. Total strangeness is predicted by integrated splitting functions, which represent the probability that the proton will fluctuate into a given meson-baryon pair. However, the momentum distributions s (x) and s (x) in the proton are determined from convolution integrals that depend on the parton distribution functions (PDFs) used for the mesons and baryons in the MCM. Theoretical calculations of these momentum distributions use many different forms for these PDFs. In our investigation, we calculate PDFs for K, K*, Λ, and Σ from two-body wave functions in a Light Cone Model (LCM) of the hadrons. We use these PDFs in conjunction with the MCM to create a hybrid model and compare our results to other theoretical calculations, experimental data from NuTeV, HERMES, ATLAS, and global parton distribution analyses. The Meson Cloud Model (MCM) has proven to be a natural explanation for strangeness in the proton because of meson-baryon splitting into kaon-hyperon pairs. Total strangeness is predicted by integrated splitting functions, which represent the probability that the proton will fluctuate into a given meson-baryon pair. However, the momentum distributions s (x) and s (x) in the proton are determined from convolution integrals that depend on the parton distribution functions (PDFs) used for the mesons and baryons in the MCM. Theoretical calculations of these momentum distributions use many different forms for these PDFs. In our investigation, we calculate PDFs for K, K*, Λ, and Σ from two-body wave functions in a Light Cone Model (LCM) of the hadrons. We use these PDFs in conjunction with the MCM to create a hybrid model and compare our results to other theoretical calculations, experimental data from NuTeV, HERMES, ATLAS, and global parton distribution analyses. This research has been supported in part by the
Statistical modelling of software reliability
NASA Technical Reports Server (NTRS)
Miller, Douglas R.
1991-01-01
During the six-month period from 1 April 1991 to 30 September 1991 the following research papers in statistical modeling of software reliability appeared: (1) A Nonparametric Software Reliability Growth Model; (2) On the Use and the Performance of Software Reliability Growth Models; (3) Research and Development Issues in Software Reliability Engineering; (4) Special Issues on Software; and (5) Software Reliability and Safety.
Transition from hadronic to partonic interactions for a composite spin-1/2 model of a nucleon
Tjon, J. A.; Wallace, S. J.
2000-12-01
A simple model of a composite nucleon is developed in which a fermion and a boson, representing quark and diquark constituents of the nucleon, form a bound state owing to a contact interaction. Photon and pion couplings to the quark provide vertex functions for the photon and pion interactions with the composite nucleon. By a suitable choice of cutoff parameters of the model, realistic electromagnetic form factors are obtained for the proton. When a pseudoscalar pion-quark coupling is used, the pion-nucleon coupling is predominantly pseudovector. A virtual photopion amplitude is considered in which there are two types of contributions: hadronic contributions where the photon and pion interactions have an intervening propagator of the nucleon or its excited states, and contactlike contributions where the photon and pion interactions occur within a single vertex. At large Q, the contactlike contributions are dominant. The model nucleon exhibits scaling behavior in deep-inelastic scattering and the normalization of the parton distribution provides a rough normalization of the contactlike contributions. Calculations for the virtual photopion amplitude are performed using kinematics appropriate to its occurrence as a meson-exchange current in electron-deuteron scattering. The results suggest that the contactlike terms can dominate the meson-exchange current for Q>1GeV/c. There is a direct connection of the contactlike terms to the off-forward parton distributions of the model nucleon.
Statistics by Example, Finding Models.
ERIC Educational Resources Information Center
Mosteller, Frederick; And Others
This booklet, part of a series of four which provide problems in probability and statistics for the secondary school level, is aimed at aiding the student in developing models as structure for data and in learning how to change models to fit real-life problems. Twelve different problem situations arising from biology, business, English, physical…
Violation of KNO scaling and the NBD phenomenon in the framework of the statistical bootstrap model
NASA Astrophysics Data System (ADS)
Kokoulina, E. S.; Kuvshinov, V. I.
1991-05-01
The connection is considered of multiplicity distributions in three stages: partonic, hadronization, and hadronic. An interpretation of the LoPHD parameter is found. It is shown that under specific hypotheses on the form of the mass spectrum, the statistical bootstrap model leads to the negative binomial distribution (NBD) at the hadronic stage of development of the multiple production process with specific analytic dependences of the parameters of the NBD.
NASA Astrophysics Data System (ADS)
Bellm, Johannes; Plätzer, Simon; Richardson, Peter; Siódmok, Andrzej; Webster, Stephen
2016-08-01
We report on the possibility of reweighting parton-shower Monte Carlo predictions for scale variations in the parton-shower algorithm. The method is based on a generalization of the Sudakov veto algorithm. We demonstrate the feasibility of this approach using example physical distributions. Implementations are available for both of the parton-shower modules in the Herwig 7 event generator.
Kovalenko, V. N.
2013-10-15
The soft part of proton-proton interaction is considered within a phenomenological model that involves the formation of color strings. Under the assumption that an elementary collision is associated with the interaction of two color dipoles, the total inelastic cross section and the multiplicity of charged particles are estimated in order to fix model parameters. Particular attention is given to modeling of exclusive parton distributions with allowance for the energy-conservation law and for fixing the center of mass, which are necessary for describing correlations. An algorithm that describes the fusion of strings in the transverse plane and which takes into account their finite rapidity width is developed. The influence of string-fusion effects on long-range correlations is found within this mechanism.
Improved model for statistical alignment
Miklos, I.; Toroczkai, Z.
2001-01-01
The statistical approach to molecular sequence evolution involves the stochastic modeling of the substitution, insertion and deletion processes. Substitution has been modeled in a reliable way for more than three decades by using finite Markov-processes. Insertion and deletion, however, seem to be more difficult to model, and thc recent approaches cannot acceptably deal with multiple insertions and deletions. A new method based on a generating function approach is introduced to describe the multiple insertion process. The presented algorithm computes the approximate joint probability of two sequences in 0(13) running time where 1 is the geometric mean of the sequence lengths.
Statistical Modelling of Compound Floods
NASA Astrophysics Data System (ADS)
Bevacqua, Emanuele; Maraun, Douglas; Vrac, Mathieu; Widmann, Martin; Manning, Colin
2016-04-01
In the recent special report of the Intergovernmental Panel on Climate Change (IPCC) on extreme events it has been highlighted that an important class of extreme events has received little attention so far: so-called compound events (CEs) (Seneviratne et al., 2012). Compound events (CEs) are multivariate extreme events in which the individual contributing events might not be extreme themselves, but their joint occurrence causes an extreme impact. Following Leonard et al., 2013, we define events as CEs only when the contributing events are statistically dependent. For many events analysed so far, the contributing events have not been statistically dependent (e.g. the floods in Rotterdam, Van den Brink et al., 2005). Two typical examples of CEs are severe drought in conjunction with a heatwave, and storm surges coinciding with heavy rain that cause the so-called Compound Floods in the lower section of a river. We develop a multivariate statistical model to represent and analyse the physical mechanisms driving CEs, and to quantify the risk associated with these events. The model is based on pair-copula construction theory, which has the advantage of building joint probability distributions modeling the marginal distributions separately from the dependence structure among variables. This allows to analyse the individual contributing variables underlying the CE separately to their dependence structure. Here is presented an application of the statistical model for Compound Floods, based on a conceptual case study. For these particular events it is not trivial to find satisfying data. Usually, water level stations are not present in the area of the river where both the influence of the sea and river are seen. The main reason being that this critical area is small and stakeholders have little interest in measuring both effect from the sea and from the river. For these reasons we have developed a conceptual case study which allows us to vary the system's physical parameters
Statistical models for seismic magnitude
NASA Astrophysics Data System (ADS)
Christoffersson, Anders
1980-02-01
In this paper some statistical models in connection with seismic magnitude are presented. Two main situations are treated. The first deals with the estimation of magnitude for an event, using a fixed network of stations and taking into account the detection and bias properties of the individual stations. The second treats the problem of estimating seismicity, and detection and bias properties of individual stations. The models are applied to analyze the magnitude bias effects for an earthquake aftershock sequence from Japan, as recorded by a hypothetical network of 15 stations. It is found that network magnitudes computed by the conventional averaging technique are considerably biased, and that a maximum likelihood approach using instantaneous noise-level estimates for non-detecting stations gives the most consistent magnitude estimates. Finally, the models are applied to evaluate the detection characteristics and associated seismicity as recorded by three VELA arrays: UBO (Uinta Basin), TFO (Tonto Forest) and WMO (Wichita Mountains).
Equilibrium statistical-thermal models in high-energy physics
NASA Astrophysics Data System (ADS)
Tawfik, Abdel Nasser
2014-05-01
We review some recent highlights from the applications of statistical-thermal models to different experimental measurements and lattice QCD thermodynamics that have been made during the last decade. We start with a short review of the historical milestones on the path of constructing statistical-thermal models for heavy-ion physics. We discovered that Heinz Koppe formulated in 1948, an almost complete recipe for the statistical-thermal models. In 1950, Enrico Fermi generalized this statistical approach, in which he started with a general cross-section formula and inserted into it, the simplifying assumptions about the matrix element of the interaction process that likely reflects many features of the high-energy reactions dominated by density in the phase space of final states. In 1964, Hagedorn systematically analyzed the high-energy phenomena using all tools of statistical physics and introduced the concept of limiting temperature based on the statistical bootstrap model. It turns to be quite often that many-particle systems can be studied with the help of statistical-thermal methods. The analysis of yield multiplicities in high-energy collisions gives an overwhelming evidence for the chemical equilibrium in the final state. The strange particles might be an exception, as they are suppressed at lower beam energies. However, their relative yields fulfill statistical equilibrium, as well. We review the equilibrium statistical-thermal models for particle production, fluctuations and collective flow in heavy-ion experiments. We also review their reproduction of the lattice QCD thermodynamics at vanishing and finite chemical potential. During the last decade, five conditions have been suggested to describe the universal behavior of the chemical freeze-out parameters. The higher order moments of multiplicity have been discussed. They offer deep insights about particle production and to critical fluctuations. Therefore, we use them to describe the freeze-out parameters
Statistical models for trisomic phenotypes
Lamb, N.E.; Sherman, S.L.; Feingold, E.
1996-01-01
Certain genetic disorders are rare in the general population but more common in individuals with specific trisomies, which suggests that the genes involved in the etiology of these disorders may be located on the trisomic chromosome. As with all aneuploid syndromes, however, a considerable degree of variation exists within each phenotype so that any given trait is present only among a subset of the trisomic population. We have previously presented a simple gene-dosage model to explain this phenotypic variation and developed a strategy to map genes for such traits. The mapping strategy does not depend on the simple model but works in theory under any model that predicts that affected individuals have an increased likelihood of disomic homozygosity at the trait locus. This paper explores the robustness of our mapping method by investigating what kinds of models give an expected increase in disomic homozygosity. We describe a number of basic statistical models for trisomic phenotypes. Some of these are logical extensions of standard models for disomic phenotypes, and some are more specific to trisomy. Where possible, we discuss genetic mechanisms applicable to each model. We investigate which models and which parameter values give an expected increase in disomic homozygosity in individuals with the trait. Finally, we determine the sample sizes required to identify the increased disomic homozygosity under each model. Most of the models we explore yield detectable increases in disomic homozygosity for some reasonable range of parameter values, usually corresponding to smaller trait frequencies. It therefore appears that our mapping method should be effective for a wide variety of moderately infrequent traits, even though the exact mode of inheritance is unlikely to be known. 21 refs., 8 figs., 1 tab.
Simulations of Statistical Model Fits to RHIC Data
NASA Astrophysics Data System (ADS)
Llope, W. J.
2013-04-01
The application of statistical model fits to experimentally measured particle multiplicity ratios allows inferences of the average values of temperatures, T, baryochemical potentials, μB, and other quantities at chemical freeze-out. The location of the boundary between the hadronic and partonic regions in the (μB,T) phase diagram, and the possible existence of a critical point, remains largely speculative. The search for a critical point using the moments of the particle multiplicity distributions in tightly centrality constrained event samples makes the tacit assumption that the variances in the (μB,T) values in these samples is sufficiently small to tightly localize the events in the phase diagram. This and other aspects were explored in simulations by coupling the UrQMD transport model to the statistical model code Thermus. The phase diagram trajectories of individual events versus the time in fm/c was calculated versus the centrality and beam energy. The variances of the (μB,T) values at freeze-out, even in narrow centrality bins, are seen to be relatively large. This suggests that a new way to constrain the events on the phase diagram may lead to more sensitive searches for the possible critical point.
Nuclear Parton Distribution Functions
I. Schienbein, J.Y. Yu, C. Keppel, J.G. Morfin, F. Olness, J.F. Owens
2009-06-01
We study nuclear effects of charged current deep inelastic neutrino-iron scattering in the framework of a {chi}{sup 2} analysis of parton distribution functions (PDFs). We extract a set of iron PDFs which are used to compute x{sub Bj}-dependent and Q{sup 2}-dependent nuclear correction factors for iron structure functions which are required in global analyses of free nucleon PDFs. We compare our results with nuclear correction factors from neutrino-nucleus scattering models and correction factors for charged-lepton--iron scattering. We find that, except for very high x{sub Bj}, our correction factors differ in both shape and magnitude from the correction factors of the models and charged-lepton scattering.
NASA Astrophysics Data System (ADS)
Nam, Seung-il
2012-10-01
We investigate the parton-distribution functions (PDFs) for the positively charged pion and kaon at a low renormalization scale ˜1GeV. To this end, we employ the gauge-invariant effective chiral action from the nonlocal chiral-quark model, resulting in the vector currents being conserved. All the model parameters are determined phenomenologically with the normalization condition for PDF and the empirical values for the pseudoscalar meson weak-decay constants. We consider the momentum dependence of the effective quark mass properly within the model calculations. It turns out that the leading local contribution provides about 70% of the total strength for PDF, whereas the nonlocal one, which is newly taken into account in this work for the gauge invariance, does the rest. High-Q2 evolution to 27GeV2 is performed for the valance-quark distribution function, using the Dokshitzer-Gribov-Lipatov-Altarelli-Parisi equation. The moments for the pion and kaon valance-quark distribution functions are also computed. The numerical results are compared with the empirical data and theoretical estimations, and show qualitatively agreement with them.
PACIAE 2.1: An updated issue of the parton and hadron cascade model PACIAE 2.0
NASA Astrophysics Data System (ADS)
Sa, Ben-Hao; Zhou, Dai-Mei; Yan, Yu-Liang; Dong, Bao-Guo; Cai, Xu
2013-05-01
We have updated the parton and hadron cascade model PACIAE 2.0 (cf. Ben-Hao Sa, Dai-Mei Zhou, Yu-Liang Yan, Xiao-Mei Li, Sheng-Qin Feng, Bao-Guo Dong, Xu Cai, Comput. Phys. Comm. 183 (2012) 333.) to the new issue of PACIAE 2.1. The PACIAE model is based on PYTHIA. In the PYTHIA model, once the hadron transverse momentum pT is randomly sampled in the string fragmentation, the px and py components are originally put on the circle with radius pT randomly. Now it is put on the circumference of ellipse with half major and minor axes of pT(1+δp) and pT(1-δp), respectively, in order to better investigate the final state transverse momentum anisotropy. New version program summaryManuscript title: PACIAE 2.1: An updated issue of the parton and hadron cascade model PACIAE 2.0 Authors: Ben-Hao Sa, Dai-Mei Zhou, Yu-Liang Yan, Bao-Guo Dong, and Xu Cai Program title: PACIAE version 2.1 Journal reference: Catalogue identifier: Licensing provisions: none Programming language: FORTRAN 77 or GFORTRAN Computer: DELL Studio XPS and others with a FORTRAN 77 or GFORTRAN compiler Operating system: Linux or Windows with FORTRAN 77 or GFORTRAN compiler RAM: ≈ 1GB Number of processors used: Supplementary material: Keywords: relativistic nuclear collision; PYTHIA model; PACIAE model Classification: 11.1, 17.8 External routines/libraries: Subprograms used: Catalogue identifier of previous version: aeki_v1_0* Journal reference of previous version: Comput. Phys. Comm. 183(2012)333. Does the new version supersede the previous version?: Yes* Nature of problem: PACIAE is based on PYTHIA. In the PYTHIA model, once the hadron transverse momentum(pT)is randomly sampled in the string fragmentation, thepxandpycomponents are randomly placed on the circle with radius ofpT. This strongly cancels the final state transverse momentum asymmetry developed dynamically. Solution method: Thepxandpycomponent of hadron in the string fragmentation is now randomly placed on the circumference of an ellipse with
STATISTICAL MODELS FOR WATER MAIN FAILURES
A detailed statistical analysis of pipe break records from New Haven, Connecticut, and Cincinnati, Ohio, water distribution systems focussed on deriving predictive models for pipe failure probabilities at the individual pipe level. The statistical methodology of the proportional ...
Statistical modeling of software reliability
NASA Technical Reports Server (NTRS)
Miller, Douglas R.
1992-01-01
This working paper discusses the statistical simulation part of a controlled software development experiment being conducted under the direction of the System Validation Methods Branch, Information Systems Division, NASA Langley Research Center. The experiment uses guidance and control software (GCS) aboard a fictitious planetary landing spacecraft: real-time control software operating on a transient mission. Software execution is simulated to study the statistical aspects of reliability and other failure characteristics of the software during development, testing, and random usage. Quantification of software reliability is a major goal. Various reliability concepts are discussed. Experiments are described for performing simulations and collecting appropriate simulated software performance and failure data. This data is then used to make statistical inferences about the quality of the software development and verification processes as well as inferences about the reliability of software versions and reliability growth under random testing and debugging.
Unraveling hadron structure with generalized parton distributions
Andrei Belitsky; Anatoly Radyushkin
2004-10-01
The recently introduced generalized parton distributions have emerged as a universal tool to describe hadrons in terms of quark and gluonic degrees of freedom. They combine the features of form factors, parton densities and distribution amplitudes - the functions used for a long time in studies of hadronic structure. Generalized parton distributions are analogous to the phase-space Wigner quasi-probability function of non-relativistic quantum mechanics which encodes full information on a quantum-mechanical system. We give an extensive review of main achievements in the development of this formalism. We discuss physical interpretation and basic properties of generalized parton distributions, their modeling and QCD evolution in the leading and next-to-leading orders. We describe how these functions enter a wide class of exclusive reactions, such as electro- and photo-production of photons, lepton pairs, or mesons.
From many body wee partons dynamics to perfect fluid: a standard model for heavy ion collisions
Venugopalan, R.
2010-07-22
We discuss a standard model of heavy ion collisions that has emerged both from experimental results of the RHIC program and associated theoretical developments. We comment briefly on the impact of early results of the LHC program on this picture. We consider how this standard model of heavy ion collisions could be solidified or falsified in future experiments at RHIC, the LHC and a future Electro-Ion Collider.
Statistical Modeling of SAR Images: A Survey
Gao, Gui
2010-01-01
Statistical modeling is essential to SAR (Synthetic Aperture Radar) image interpretation. It aims to describe SAR images through statistical methods and reveal the characteristics of these images. Moreover, statistical modeling can provide a technical support for a comprehensive understanding of terrain scattering mechanism, which helps to develop algorithms for effective image interpretation and creditable image simulation. Numerous statistical models have been developed to describe SAR image data, and the purpose of this paper is to categorize and evaluate these models. We first summarize the development history and the current researching state of statistical modeling, then different SAR image models developed from the product model are mainly discussed in detail. Relevant issues are also discussed. Several promising directions for future research are concluded at last. PMID:22315568
Statistical modeling of electrical components: Final report
Jolly, R.L.
1988-07-01
A method of forecasting production yields based on SPICE (University of California at Berkeley) circuit simulation and Monte Carlo techniques was evaluated. This method involved calculating functionally accurate component models using statistical techniques and using these component models in a SPICE electrical circuit simulation program. The results of the simulation program allow production yields to be calculated using standard statistical techniques.
Measurement of parton shower observables with OPAL
NASA Astrophysics Data System (ADS)
Fischer, N.; Gieseke, S.; Kluth, S.; Plätzer, S.; Skands, P.
2016-07-01
A study of QCD coherence is presented based on a sample of about 397,000 e+e- hadronic annihilation events collected at √s = 91 GeV with the OPAL detector at LEP. The study is based on four recently proposed observables that are sensitive to coherence effects in the perturbative regime. The measurement of these observables is presented, along with a comparison with the predictions of different parton shower models. The models include both conventional parton shower models and dipole antenna models. Different ordering variables are used to investigate their influence on the predictions.
Automated statistical modeling of analytical measurement systems
Jacobson, J J
1992-08-01
The statistical modeling of analytical measurement systems at the Idaho Chemical Processing Plant (ICPP) has been completely automated through computer software. The statistical modeling of analytical measurement systems is one part of a complete quality control program used by the Remote Analytical Laboratory (RAL) at the ICPP. The quality control program is an integration of automated data input, measurement system calibration, database management, and statistical process control. The quality control program and statistical modeling program meet the guidelines set forth by the American Society for Testing Materials and American National Standards Institute. A statistical model is a set of mathematical equations describing any systematic bias inherent in a measurement system and the precision of a measurement system. A statistical model is developed from data generated from the analysis of control standards. Control standards are samples which are made up at precise known levels by an independent laboratory and submitted to the RAL. The RAL analysts who process control standards do not know the values of those control standards. The object behind statistical modeling is to describe real process samples in terms of their bias and precision and, to verify that a measurement system is operating satisfactorily. The processing of control standards gives us this ability.
SINGULARITIES OF GENERALIZED PARTON DISTRIBUTIONS
Anatoly Radyushkin
2012-12-01
We discuss recent developments in building models for generalized parton distributions (GPDs) that are based on the formalism of double distributions (DDs). A special attention is given to a careful analysis of the singularity structure of DDs. The DD formalism is applied to construction of a model GPDs with a singular Regge behavior. Within the developed DD-based approach, we discuss the structure of GPD sum rules. It is shown that separation of DDs into the so-called ``plus'' part and the D-term part may be treated as a renormalization procedure for the GPD sum rules. This approach is compared with an alternative prescription based on analytic regularization.
Different Manhattan project: automatic statistical model generation
NASA Astrophysics Data System (ADS)
Yap, Chee Keng; Biermann, Henning; Hertzmann, Aaron; Li, Chen; Meyer, Jon; Pao, Hsing-Kuo; Paxia, Salvatore
2002-03-01
We address the automatic generation of large geometric models. This is important in visualization for several reasons. First, many applications need access to large but interesting data models. Second, we often need such data sets with particular characteristics (e.g., urban models, park and recreation landscape). Thus we need the ability to generate models with different parameters. We propose a new approach for generating such models. It is based on a top-down propagation of statistical parameters. We illustrate the method in the generation of a statistical model of Manhattan. But the method is generally applicable in the generation of models of large geographical regions. Our work is related to the literature on generating complex natural scenes (smoke, forests, etc) based on procedural descriptions. The difference in our approach stems from three characteristics: modeling with statistical parameters, integration of ground truth (actual map data), and a library-based approach for texture mapping.
Topology for statistical modeling of petascale data.
Pascucci, Valerio; Mascarenhas, Ajith Arthur; Rusek, Korben; Bennett, Janine Camille; Levine, Joshua; Pebay, Philippe Pierre; Gyulassy, Attila; Thompson, David C.; Rojas, Joseph Maurice
2011-07-01
This document presents current technical progress and dissemination of results for the Mathematics for Analysis of Petascale Data (MAPD) project titled 'Topology for Statistical Modeling of Petascale Data', funded by the Office of Science Advanced Scientific Computing Research (ASCR) Applied Math program. Many commonly used algorithms for mathematical analysis do not scale well enough to accommodate the size or complexity of petascale data produced by computational simulations. The primary goal of this project is thus to develop new mathematical tools that address both the petascale size and uncertain nature of current data. At a high level, our approach is based on the complementary techniques of combinatorial topology and statistical modeling. In particular, we use combinatorial topology to filter out spurious data that would otherwise skew statistical modeling techniques, and we employ advanced algorithms from algebraic statistics to efficiently find globally optimal fits to statistical models. This document summarizes the technical advances we have made to date that were made possible in whole or in part by MAPD funding. These technical contributions can be divided loosely into three categories: (1) advances in the field of combinatorial topology, (2) advances in statistical modeling, and (3) new integrated topological and statistical methods.
Accardi, Alberto; Owens, Jeff F.
2013-07-01
Three new sets of next-to-leading order parton distribution functions (PDFs) are presented, determined by global fits to a wide variety of data for hard scattering processes. The analysis includes target mass and higher twist corrections needed for the description of deep-inelastic scattering data at large x and low Q^2, and nuclear corrections for deuterium targets. The PDF sets correspond to three different models for the nuclear effects, and provide a more realistic uncertainty range for the d quark PDF compared with previous fits. Applications to weak boson production at colliders are also discussed.
KOVCHEGOV,Y.V.
2000-04-25
The authors derive an equation determining the small-x evolution of the F{sub 2} structure function of a large nucleus which resumes a cascade of gluons in the leading logarithmic approximation using Mueller's color dipole model. In the traditional language it corresponds to resummation of the pomeron fan diagrams, originally conjectured in the GLR equation. The authors show that the solution of the equation describes the physics of structure functions at high partonic densities, thus allowing them to gain some understanding of the most interesting and challenging phenomena in small-x physics--saturation.
TOPICS IN THEORY OF GENERALIZED PARTON DISTRIBUTIONS
Radyushkin, Anatoly V.
2013-05-01
Several topics in the theory of generalized parton distributions (GPDs) are reviewed. First, we give a brief overview of the basics of the theory of generalized parton distributions and their relationship with simpler phenomenological functions, viz. form factors, parton densities and distribution amplitudes. Then, we discuss recent developments in building models for GPDs that are based on the formalism of double distributions (DDs). A special attention is given to a careful analysis of the singularity structure of DDs. The DD formalism is applied to construction of a model GPDs with a singular Regge behavior. Within the developed DD-based approach, we discuss the structure of GPD sum rules. It is shown that separation of DDs into the so-called ``plus'' part and the $D$-term part may be treated as a renormalization procedure for the GPD sum rules. This approach is compared with an alternative prescription based on analytic regularization.
Generalized parton distributions in AdS/QCD
Vega, Alfredo; Schmidt, Ivan; Gutsche, Thomas; Lyubovitskij, Valery E.
2011-02-01
The nucleon helicity-independent generalized parton distributions of quarks are calculated in the zero skewness case, in the framework of the anti-de Sitter/QCD model. The present approach is based on a matching procedure of sum rules relating the electromagnetic form factors to generalized parton distributions and anti-de Sitter modes.
Exponential family models and statistical genetics.
Palmgren, J
2000-02-01
This article describes the evolution of applied exponential family models, starting at 1972, the year of publication of the seminal papers on generalized linear models and on Cox regression, and leading to multivariate (i) marginal models and inference based on estimating equations and (ii) random effects models and Bayesian simulation-based posterior inference. By referring to recent work in genetic epidemiology, on semiparametric methods for linkage analysis and on transmission/disequilibrium tests for haplotype transmission this paper illustrates the potential for the recent advances in applied probability and statistics to contribute to new and unified tools for statistical genetics. Finally, it is emphasized that there is a need for well-defined postgraduate education paths in medical statistics in the year 2000 and thereafter. PMID:10826159
Statistical Modeling for Radiation Hardness Assurance
NASA Technical Reports Server (NTRS)
Ladbury, Raymond L.
2014-01-01
We cover the models and statistics associated with single event effects (and total ionizing dose), why we need them, and how to use them: What models are used, what errors exist in real test data, and what the model allows us to say about the DUT will be discussed. In addition, how to use other sources of data such as historical, heritage, and similar part and how to apply experience, physics, and expert opinion to the analysis will be covered. Also included will be concepts of Bayesian statistics, data fitting, and bounding rates.
Statistical Model of Evaporating Multicomponent Fuel Drops
NASA Technical Reports Server (NTRS)
Harstad, Kenneth; LeClercq, Patrick; Bellan, Josette
2007-01-01
An improved statistical model has been developed to describe the chemical composition of an evaporating multicomponent- liquid drop and of the mixture of gases surrounding the drop. The model is intended for use in computational simulations of the evaporation and combustion of sprayed liquid fuels, which are typically mixtures of as many as hundreds of different hydrocarbon compounds. The present statistical model is an approximation designed to afford results that are accurate enough to contribute to understanding of the simulated physical and chemical phenomena, without imposing an unduly large computational burden.
Structured statistical models of inductive reasoning.
Kemp, Charles; Tenenbaum, Joshua B
2009-01-01
Everyday inductive inferences are often guided by rich background knowledge. Formal models of induction should aim to incorporate this knowledge and should explain how different kinds of knowledge lead to the distinctive patterns of reasoning found in different inductive contexts. This article presents a Bayesian framework that attempts to meet both goals and describes [corrected] 4 applications of the framework: a taxonomic model, a spatial model, a threshold model, and a causal model. Each model makes probabilistic inferences about the extensions of novel properties, but the priors for the 4 models are defined over different kinds of structures that capture different relationships between the categories in a domain. The framework therefore shows how statistical inference can operate over structured background knowledge, and the authors argue that this interaction between structure and statistics is critical for explaining the power and flexibility of human reasoning. PMID:19159147
Partonic Transverse Momentum Distributions
Rossi, Patrizia
2010-08-04
In recent years parton distributions have been generalized to account also for transverse degrees of freedom and new sets of more general distributions, Transverse Momentum Dependent (TMD) parton distributions and fragmentation functions were introduced. Different experiments worldwide (HERMES, COMPASS, CLAS, JLab-Hall A) have measurements of TMDs in semi-inclusive DIS processes as one of their main focuses of research. TMD studies are also an important part of the present and future Drell-Yan experiments at RICH and JPARC and GSI, respectively, Studies of TMDs are also one of the main driving forces of the Jefferson Lab (JLab) 12 GeV upgrade project. Progress in phenomenology and theory is flourishing as well. In this talk an overview of the latest developments in studies of TMDs will be given and newly released results, ongoing activities, as well as planned near term and future measurements will be discussed.
Dynamics of hot and dense nuclear and partonic matter
Bratkovskaya, E. L.; Cassing, W.; Linnyk, O.; Konchakovski, V. P.; Voronyuk, V.; Ozvenchuk, V.
2012-06-15
The dynamics of hot and dense nuclear matter is discussed from the microscopic transport point of view. The basic concepts of the Hadron-String-Dynamical transport model (HSD)-derived from Kadanoff-Baym equations in phase phase-are presented as well as 'highlights' of HSD results for different observables in heavy-ion collisions from 100 A MeV (SIS) to 21 A TeV(RHIC) energies. Furthermore, a novel extension of the HSD model for the description of the partonic phase-the Parton-Hadron-String-Dynamics (PHSD) approach-is introduced. PHSD includes a nontrivial partonic equation of state-in line with lattice QCD-as well as covariant transition rates from partonic to hadronic degrees of freedom. The sensitivity of hadronic observables to the partonic phase is demonstrated for relativistic heavy-ion collisions from the FAIR/NICA up to the RHIC energy regime.
Statistical label fusion with hierarchical performance models
Asman, Andrew J.; Dagley, Alexander S.; Landman, Bennett A.
2014-01-01
Label fusion is a critical step in many image segmentation frameworks (e.g., multi-atlas segmentation) as it provides a mechanism for generalizing a collection of labeled examples into a single estimate of the underlying segmentation. In the multi-label case, typical label fusion algorithms treat all labels equally – fully neglecting the known, yet complex, anatomical relationships exhibited in the data. To address this problem, we propose a generalized statistical fusion framework using hierarchical models of rater performance. Building on the seminal work in statistical fusion, we reformulate the traditional rater performance model from a multi-tiered hierarchical perspective. This new approach provides a natural framework for leveraging known anatomical relationships and accurately modeling the types of errors that raters (or atlases) make within a hierarchically consistent formulation. Herein, we describe several contributions. First, we derive a theoretical advancement to the statistical fusion framework that enables the simultaneous estimation of multiple (hierarchical) performance models within the statistical fusion context. Second, we demonstrate that the proposed hierarchical formulation is highly amenable to the state-of-the-art advancements that have been made to the statistical fusion framework. Lastly, in an empirical whole-brain segmentation task we demonstrate substantial qualitative and significant quantitative improvement in overall segmentation accuracy. PMID:24817809
Statistical Hot Spot Model for Explosive Detonation
Nichols III, A L
2004-05-10
The Non-local Thermodynamic Equilibrium Statistical Hot Spot Model (NLTE SHS), a new model for explosive detonation, is described. In this model, the formation, ignition, propagation, and extinction of hot spots is explicitly modeled. The equation of state of the explosive mixture is treated with a nonlocal equilibrium thermodynamic assumption. A methodology for developing the parameters for the model is discussed, and applied to the detonation velocity diameter effect. Examination of these results indicates where future improvements to the model can be made.
Statistical Hot Spot Model for Explosive Detonation
Nichols, III, A L
2005-07-14
The Non-local Thermodynamic Equilibrium Statistical Hot Spot Model (NLTE SHS), a new model for explosive detonation, is described. In this model, the formation, ignition, propagation, and extinction of hot spots is explicitly modeled. The equation of state of the explosive mixture is treated with a non-local equilibrium thermodynamic assumption. A methodology for developing the parameters for the model is discussed, and applied to the detonation velocity diameter effect. Examination of these results indicates where future improvements to the model can be made.
Statistical modeling of the arterial vascular tree
NASA Astrophysics Data System (ADS)
Beck, Thomas; Godenschwager, Christian; Bauer, Miriam; Bernhardt, Dominik; Dillmann, Rüdiger
2011-03-01
Automatic examination of medical images becomes increasingly important due to the rising amount of data. Therefore automated methods are required which combine anatomical knowledge and robust segmentation to examine the structure of interest. We propose a statistical model of the vascular tree based on vascular landmarks and unbranched vessel sections. An undirected graph provides anatomical topology, semantics, existing landmarks and attached vessel sections. The atlas was built using semi-automatically generated geometric models of various body regions ranging from carotid arteries to the lower legs. Geometric models contain vessel centerlines as well as orthogonal cross-sections in equidistant intervals with the vessel contour having the form of a polygon path. The geometric vascular model is supplemented by anatomical landmarks which are not necessarily related to the vascular system. These anatomical landmarks define point correspondences which are used for registration with a Thin-Plate-Spline interpolation. After the registration process, the models were merged to form the statistical model which can be mapped to unseen images based on a subset of anatomical landmarks. This approach provides probability distributions for the location of landmarks, vessel-specific geometric properties including shape, expected radii and branching points and vascular topology. The applications of this statistical model include model-based extraction of the vascular tree which greatly benefits from vessel-specific geometry description and variation ranges. Furthermore, the statistical model can be applied as a basis for computer aided diagnosis systems as indicator for pathologically deformed vessels and the interaction with the geometric model is significantly more user friendly for physicians through anatomical names.
Statistical transmutation in doped quantum dimer models.
Lamas, C A; Ralko, A; Cabra, D C; Poilblanc, D; Pujol, P
2012-07-01
We prove a "statistical transmutation" symmetry of doped quantum dimer models on the square, triangular, and kagome lattices: the energy spectrum is invariant under a simultaneous change of statistics (i.e., bosonic into fermionic or vice versa) of the holes and of the signs of all the dimer resonance loops. This exact transformation enables us to define the duality equivalence between doped quantum dimer Hamiltonians and provides the analytic framework to analyze dynamical statistical transmutations. We investigate numerically the doping of the triangular quantum dimer model with special focus on the topological Z(2) dimer liquid. Doping leads to four (instead of two for the square lattice) inequivalent families of Hamiltonians. Competition between phase separation, superfluidity, supersolidity, and fermionic phases is investigated in the four families. PMID:23031119
Topology for Statistical Modeling of Petascale Data.
Bennett, Janine Camille; Pebay, Philippe Pierre; Pascucci, Valerio; Levine, Joshua; Gyulassy, Attila; Rojas, Joseph Maurice
2014-07-01
This document presents current technical progress and dissemination of results for the Mathematics for Analysis of Petascale Data (MAPD) project titled "Topology for Statistical Modeling of Petascale Data", funded by the Office of Science Advanced Scientific Computing Research (ASCR) Applied Math program.
MICROARRAY DATA ANALYSIS USING MULTIPLE STATISTICAL MODELS
Microarray Data Analysis Using Multiple Statistical Models
Wenjun Bao1, Judith E. Schmid1, Amber K. Goetz1, Ming Ouyang2, William J. Welsh2,Andrew I. Brooks3,4, ChiYi Chu3,Mitsunori Ogihara3,4, Yinhe Cheng5, David J. Dix1. 1National Health and Environmental Effects Researc...
Observer models for statistically-defined backgrounds
NASA Astrophysics Data System (ADS)
Burgess, Arthur E.
1994-04-01
Investigation of human signal-detection performance for noise- limited tasks with statistically defined signal or image parameters represents a step towards clinical realism. However, the ideal observer procedure is then usually nonlinear, and analysis becomes mathematically intractable. Two linear but suboptimal observer models, the Hotelling observer and the non- prewhitening (NPW) matched filter, have been proposed for mathematical convenience. Experiments by Rolland and Barrett involving detection of signals in white noise superimposed on statistically defined backgrounds showed that the Hotelling model gave a good fit while the simple NPW matched filter gave a poor fit. It will be shown that the NPW model can be modified to fit their data by adding a spatial frequency filter of shape similar to the human contrast sensitivity function. The best fit is obtained using an eye filter model, E(f) equals f1.3 exp(-cf2) with c selected to give a peak at 4 cycles per degree.
Statistical modeling of space shuttle environmental data
NASA Technical Reports Server (NTRS)
Tubbs, J. D.; Brewer, D. W.
1983-01-01
Statistical models which use a class of bivariate gamma distribution are examined. Topics discussed include: (1) the ratio of positively correlated gamma varieties; (2) a method to determine if unequal shape parameters are necessary in bivariate gamma distribution; (3) differential equations for modal location of a family of bivariate gamma distribution; and (4) analysis of some wind gust data using the analytical results developed for modeling application.
Statistical physical models of cellular motility
NASA Astrophysics Data System (ADS)
Banigan, Edward J.
Cellular motility is required for a wide range of biological behaviors and functions, and the topic poses a number of interesting physical questions. In this work, we construct and analyze models of various aspects of cellular motility using tools and ideas from statistical physics. We begin with a Brownian dynamics model for actin-polymerization-driven motility, which is responsible for cell crawling and "rocketing" motility of pathogens. Within this model, we explore the robustness of self-diffusiophoresis, which is a general mechanism of motility. Using this mechanism, an object such as a cell catalyzes a reaction that generates a steady-state concentration gradient that propels the object in a particular direction. We then apply these ideas to a model for depolymerization-driven motility during bacterial chromosome segregation. We find that depolymerization and protein-protein binding interactions alone are sufficient to robustly pull a chromosome, even against large loads. Next, we investigate how forces and kinetics interact during eukaryotic mitosis with a many-microtubule model. Microtubules exert forces on chromosomes, but since individual microtubules grow and shrink in a force-dependent way, these forces lead to bistable collective microtubule dynamics, which provides a mechanism for chromosome oscillations and microtubule-based tension sensing. Finally, we explore kinematic aspects of cell motility in the context of the immune system. We develop quantitative methods for analyzing cell migration statistics collected during imaging experiments. We find that during chronic infection in the brain, T cells run and pause stochastically, following the statistics of a generalized Levy walk. These statistics may contribute to immune function by mimicking an evolutionarily conserved efficient search strategy. Additionally, we find that naive T cells migrating in lymph nodes also obey non-Gaussian statistics. Altogether, our work demonstrates how physical
Pitfalls in statistical landslide susceptibility modelling
NASA Astrophysics Data System (ADS)
Schröder, Boris; Vorpahl, Peter; Märker, Michael; Elsenbeer, Helmut
2010-05-01
The use of statistical methods is a well-established approach to predict landslide occurrence probabilities and to assess landslide susceptibility. This is achieved by applying statistical methods relating historical landslide inventories to topographic indices as predictor variables. In our contribution, we compare several new and powerful methods developed in machine learning and well-established in landscape ecology and macroecology for predicting the distribution of shallow landslides in tropical mountain rainforests in southern Ecuador (among others: boosted regression trees, multivariate adaptive regression splines, maximum entropy). Although these methods are powerful, we think it is necessary to follow a basic set of guidelines to avoid some pitfalls regarding data sampling, predictor selection, and model quality assessment, especially if a comparison of different models is contemplated. We therefore suggest to apply a novel toolbox to evaluate approaches to the statistical modelling of landslide susceptibility. Additionally, we propose some methods to open the "black box" as an inherent part of machine learning methods in order to achieve further explanatory insights into preparatory factors that control landslides. Sampling of training data should be guided by hypotheses regarding processes that lead to slope failure taking into account their respective spatial scales. This approach leads to the selection of a set of candidate predictor variables considered on adequate spatial scales. This set should be checked for multicollinearity in order to facilitate model response curve interpretation. Model quality assesses how well a model is able to reproduce independent observations of its response variable. This includes criteria to evaluate different aspects of model performance, i.e. model discrimination, model calibration, and model refinement. In order to assess a possible violation of the assumption of independency in the training samples or a possible
Statistical models of lunar rocks and regolith
NASA Technical Reports Server (NTRS)
Marcus, A. H.
1973-01-01
The mathematical, statistical, and computational approaches used in the investigation of the interrelationship of lunar fragmental material, regolith, lunar rocks, and lunar craters are described. The first two phases of the work explored the sensitivity of the production model of fragmental material to mathematical assumptions, and then completed earlier studies on the survival of lunar surface rocks with respect to competing processes. The third phase combined earlier work into a detailed statistical analysis and probabilistic model of regolith formation by lithologically distinct layers, interpreted as modified crater ejecta blankets. The fourth phase of the work dealt with problems encountered in combining the results of the entire project into a comprehensive, multipurpose computer simulation model for the craters and regolith. Highlights of each phase of research are given.
Statistical Models of Adaptive Immune populations
NASA Astrophysics Data System (ADS)
Sethna, Zachary; Callan, Curtis; Walczak, Aleksandra; Mora, Thierry
The availability of large (104-106 sequences) datasets of B or T cell populations from a single individual allows reliable fitting of complex statistical models for naïve generation, somatic selection, and hypermutation. It is crucial to utilize a probabilistic/informational approach when modeling these populations. The inferred probability distributions allow for population characterization, calculation of probability distributions of various hidden variables (e.g. number of insertions), as well as statistical properties of the distribution itself (e.g. entropy). In particular, the differences between the T cell populations of embryonic and mature mice will be examined as a case study. Comparing these populations, as well as proposed mixed populations, provides a concrete exercise in model creation, comparison, choice, and validation.
NASA Astrophysics Data System (ADS)
Asaturyan, R.; Ent, R.; Mkrtchyan, H.; Navasardyan, T.; Tadevosyan, V.; Adams, G. S.; Ahmidouch, A.; Angelescu, T.; Arrington, J.; Asaturyan, A.; Baker, O. K.; Benmouna, N.; Bertoncini, C.; Blok, H. P.; Boeglin, W. U.; Bosted, P. E.; Breuer, H.; Christy, M. E.; Connell, S. H.; Cui, Y.; Dalton, M. M.; Danagoulian, S.; Day, D.; Dunne, J. A.; Dutta, D.; El Khayari, N.; Fenker, H. C.; Frolov, V. V.; Gan, L.; Gaskell, D.; Hafidi, K.; Hinton, W.; Holt, R. J.; Horn, T.; Huber, G. M.; Hungerford, E.; Jiang, X.; Jones, M.; Joo, K.; Kalantarians, N.; Kelly, J. J.; Keppel, C. E.; Kubarovsky, V.; Li, Y.; Liang, Y.; Mack, D.; Malace, S. P.; Markowitz, P.; McGrath, E.; McKee, P.; Meekins, D. G.; Mkrtchyan, A.; Moziak, B.; Niculescu, G.; Niculescu, I.; Opper, A. K.; Ostapenko, T.; Reimer, P. E.; Reinhold, J.; Roche, J.; Rock, S. E.; Schulte, E.; Segbefia, E.; Smith, C.; Smith, G. R.; Stoler, P.; Tang, L.; Ungaro, M.; Uzzle, A.; Vidakovic, S.; Villano, A.; Vulcan, W. F.; Wang, M.; Warren, G.; Wesselmann, F. R.; Wojtsekhowski, B.; Wood, S. A.; Xu, C.; Yuan, L.; Zheng, X.
2012-01-01
A large set of cross sections for semi-inclusive electroproduction of charged pions (π±) from both proton and deuteron targets was measured. The data are in the deep-inelastic scattering region with invariant mass squared W2>4 GeV2 (up to ≈7 GeV2) and range in four-momentum transfer squared 2
Asaturyan, R.; Ent, R.; Mkrtchyan, H.; Navasardyan, T.; Tadevosyan, V.; Adams, G. S.; Ahmidouch, A.; Angelescu, T.; Arrington, J.; Asaturyan, A.; et al
2012-01-01
A large set of cross sections for semi-inclusive electroproduction of charged pions (π±) from both proton and deuteron targets was measured. The data are in the deep-inelastic scattering region with invariant mass squared W2 > 4 GeV2 and range in four-momentum transfer squared 2 < Q2 < 4 (GeV/c)2, and cover a range in the Bjorken scaling variable 0.2 < x < 0.6. The fractional energy of the pions spans a range 0.3 < z < 1, with small transverse momenta with respect to the virtual-photon direction, Pt2 < 0.2 (GeV/c)2. The invariant mass that goes undetected, Mx or W',more » is in the nucleon resonance region, W' < 2 GeV. The new data conclusively show the onset of quark-hadron duality in this process, and the relation of this phenomenon to the high-energy factorization ansatz of electron-quark scattering and subsequent quark → pion production mechanisms. The x, z and Pt2 dependences of several ratios (the ratios of favored-unfavored fragmentation functions, charged pion ratios, deuteron-hydrogen and aluminum-deuteron ratios for π+ and π-) have been studied. The ratios are found to be in good agreement with expectations based upon a high-energy quark-parton model description. We find the azimuthal dependences to be small, as compared to exclusive pion electroproduction, and consistent with theoretical expectations based on tree-level factorization in terms of transverse-momentum-dependent parton distribution and fragmentation functions. In the context of a simple model, the initial transverse momenta of d quarks are found to be slightly smaller than for u quarks, while the transverse momentum width of the favored fragmentation function is about the same as for the unfavored one, and both fragmentation widths are larger than the quark widths.« less
Radar scattering statistics for digital terrain models
NASA Astrophysics Data System (ADS)
Wilson, Kelce; Patrick, Dale; Blair, James
2005-05-01
The statistic results for a digital terrain model are presented that closely match measurements for 77% of the 189 possible combinations of 7 radar bands, 3 polarizations, and 9 terrain types. The model produces realistic backscatter coefficient values for the scenarios over all incidence angles from normal to grazing. The generator was created using measured data sets reported in the Handbook of Radar Scattering Statistics for Terrain covering L, C, S, X, Ka, Ku, and W frequency bands; HH, HV, and VV polarizations; and soil and rock, shrub, tree, short vegetation, grass, dry snow, wet snow, road surface, and urban area terrain types. The first two statistical moments match published values precisely, and a Chi-Square histogram test failed to reject the generator at a 95% confidence level for the 146 terrain models implemented. A Sea State model provides the grazing angle extension for predictions beyond the available measurements. This work will contain a comprehensive set of plots of mean and standard deviation versus incidence angle.
Uphoff, Jan; Senzel, Florian; Fochler, Oliver; Wesp, Christian; Xu, Zhe; Greiner, Carsten
2015-03-20
The quark gluon plasma produced in ultrarelativistic heavy-ion collisions exhibits remarkable features. It behaves like a nearly perfect liquid with a small shear viscosity to entropy density ratio and leads to the quenching of highly energetic particles. We show that both effects can be understood for the first time within one common framework. Employing the parton cascade Boltzmann approach to multiparton scatterings, the microscopic interactions and the space-time evolution of the quark gluon plasma are calculated by solving the relativistic Boltzmann equation. Based on cross sections obtained from perturbative QCD with explicitly taking the running coupling into account, we calculate the nuclear modification factor and elliptic flow in ultrarelativistic heavy-ion collisions. With only one single parameter associated with coherence effects of medium-induced gluon radiation, the experimental data of both observables can be understood on a microscopic level. Furthermore, we show that perturbative QCD interactions with a running coupling lead to a sufficiently small shear viscosity to entropy density ratio of the quark gluon plasma, which provides a microscopic explanation for the observations stated by hydrodynamic calculations. PMID:25839262
Statistical Modeling of Retinal Optical Coherence Tomography.
Amini, Zahra; Rabbani, Hossein
2016-06-01
In this paper, a new model for retinal Optical Coherence Tomography (OCT) images is proposed. This statistical model is based on introducing a nonlinear Gaussianization transform to convert the probability distribution function (pdf) of each OCT intra-retinal layer to a Gaussian distribution. The retina is a layered structure and in OCT each of these layers has a specific pdf which is corrupted by speckle noise, therefore a mixture model for statistical modeling of OCT images is proposed. A Normal-Laplace distribution, which is a convolution of a Laplace pdf and Gaussian noise, is proposed as the distribution of each component of this model. The reason for choosing Laplace pdf is the monotonically decaying behavior of OCT intensities in each layer for healthy cases. After fitting a mixture model to the data, each component is gaussianized and all of them are combined by Averaged Maximum A Posterior (AMAP) method. To demonstrate the ability of this method, a new contrast enhancement method based on this statistical model is proposed and tested on thirteen healthy 3D OCTs taken by the Topcon 3D OCT and five 3D OCTs from Age-related Macular Degeneration (AMD) patients, taken by Zeiss Cirrus HD-OCT. Comparing the results with two contending techniques, the prominence of the proposed method is demonstrated both visually and numerically. Furthermore, to prove the efficacy of the proposed method for a more direct and specific purpose, an improvement in the segmentation of intra-retinal layers using the proposed contrast enhancement method as a preprocessing step, is demonstrated. PMID:26800532
Statistical models for operational risk management
NASA Astrophysics Data System (ADS)
Cornalba, Chiara; Giudici, Paolo
2004-07-01
The Basel Committee on Banking Supervision has released, in the last few years, recommendations for the correct determination of the risks to which a banking organization is subject. This concerns, in particular, operational risks, which are all those management events that may determine unexpected losses. It is necessary to develop valid statistical models to measure and, consequently, predict, such operational risks. In the paper we present the possible approaches, including our own proposal, which is based on Bayesian networks.
NASA Astrophysics Data System (ADS)
Shi, Shusu
2009-10-01
The measurement of event anisotropy, often called v2, provides a powerful tool for studying the properties of hot and dense medium created in high-energy nuclear collisions. The important discoveries of partonic collectivity and the brand-new process for hadronization - quark coalescence were obtained through a systematic analysis of the v2 for 200 GeV Au+Au collisions at RHIC [1]. However, early dynamic information might be masked by later hadronic rescatterings. Multistrange hadrons (φ, ξ and φ) with their large mass and presumably small hadronic cross sections should be less sensitive to hadronic rescattering in the later stage of the collisions and therefore a good probe of the early stage of the collision. We will present the measurement of v2 of π, p, KS^0, λ, ξ, φ and φ in heavy ion collisions. In minimum-bias Au+Au collisions at √sNN = 200 GeV, a significant amount of elliptic flow, almost identical to other mesons and baryons, is observed for φ and φ. Experimental observations of pT dependence of v2 of identified particles at RHIC support partonic collectivity. [4pt] [1] B. I. Abelev et al., (STAR Collaboration), Phys. Rev. C 77, 054901 (2008).
Statistical strength models for composites with discontinuities
Cheng, Ming-Chih.
1991-01-01
This thesis investigates the microfracture process in laminated composites based on a problematic approach. Two probabilistic fracture models, a statistical model and a stochastic model, were developed for predicting the failure stress in composites containing geometric discontinuities such as holes and cracks. A two-parameter Weibull distribution of fiber strength and the elastic properties of matrix materials are used in the analysis. The results show that the fracture stress is related to notch size by a power law similar to that proposed by Mar and Lin. Experiments were conducted to verify the probabilistic fracture models developed. The strength reduction factors for boron/epoxy (0/-45/0/45)s laminates were obtained and the fiber damage near the hole edge was studied. The experimental data agree well with the predictions from the probabilistic fracture models.
A simple statistical model for geomagnetic reversals
NASA Technical Reports Server (NTRS)
Constable, Catherine
1990-01-01
The diversity of paleomagnetic records of geomagnetic reversals now available indicate that the field configuration during transitions cannot be adequately described by simple zonal or standing field models. A new model described here is based on statistical properties inferred from the present field and is capable of simulating field transitions like those observed. Some insight is obtained into what one can hope to learn from paleomagnetic records. In particular, it is crucial that the effects of smoothing in the remanence acquisition process be separated from true geomagnetic field behavior. This might enable us to determine the time constants associated with the dominant field configuration during a reversal.
Computational Statistical Methods for Social Network Models
Hunter, David R.; Krivitsky, Pavel N.; Schweinberger, Michael
2013-01-01
We review the broad range of recent statistical work in social network models, with emphasis on computational aspects of these methods. Particular focus is applied to exponential-family random graph models (ERGM) and latent variable models for data on complete networks observed at a single time point, though we also briefly review many methods for incompletely observed networks and networks observed at multiple time points. Although we mention far more modeling techniques than we can possibly cover in depth, we provide numerous citations to current literature. We illustrate several of the methods on a small, well-known network dataset, Sampson’s monks, providing code where possible so that these analyses may be duplicated. PMID:23828720
Statistical Trajectory Models for Phonetic Recognition.
NASA Astrophysics Data System (ADS)
Goldenthal, William David
The main goal of this work is to develop an alternative methodology for acoustic-phonetic modelling of speech sounds. The approach utilizes a segment-based framework to capture the dynamical behavior and statistical dependencies of the acoustic attributes used to represent the speech waveform. Temporal behavior is modelled explicitly by creating dynamic tracks of the acoustic attributes used to represent the waveform, and by estimating the spatio-temporal correlation structure of the resulting errors. The tracks serve as templates from which synthetic segments of the acoustic attributes are generated. Scoring of an hypothesized phonetic segment is then based on the error between the measured acoustic attributes and the synthetic segments generated for each phonetic model. Phonetic contextual influences are accounted for in two ways. First, context-dependent biphone tracks are created for each phonetic model. These tracks are then merged as needed to generate triphone tracks. The error statistics are pooled over all the contexts for each phonetic model. This allows for the creation of a large number of contextual models (e.g., 2,500) without compromising the robustness of the statistical parameter estimates. The resulting triphone coverage is over 99.5%. The second method of accounting for context involves creating tracks of the transitions between phones. By clustering these tracks, complete models are constructed of over 200 "canonical" transitions. The transition models help in two ways. First, the transition scores are incorporated into the scoring framework to help determine the phonetic identity of the two phones involved. Secondly, they are used to determine likely segment boundaries within an utterance. This reduces the search space during phonetic recognition. Phonetic classification experiments are performed which demonstrate the importance of the temporal correlation information in the speech signal. A complete phonetic recognition system, incorporating
Tetraquark production in double parton scattering
NASA Astrophysics Data System (ADS)
Carvalho, F.; Cazaroto, E. R.; Gonçalves, V. P.; Navarra, F. S.
2016-02-01
We develop a model to study tetraquark production in hadronic collisions. We focus on double parton scattering and formulate a version of the color evaporation model for the production of the X (3872 ) and of the T4 c tetraquark, a state composed by the c c ¯c c ¯ quarks. We find that the production cross section grows rapidly with the collision energy √{s } and make predictions for the forthcoming higher energy data of the LHC.
Statistical model with a standard Γ distribution
NASA Astrophysics Data System (ADS)
Patriarca, Marco; Chakraborti, Anirban; Kaski, Kimmo
2004-07-01
We study a statistical model consisting of N basic units which interact with each other by exchanging a physical entity, according to a given microscopic random law, depending on a parameter λ . We focus on the equilibrium or stationary distribution of the entity exchanged and verify through numerical fitting of the simulation data that the final form of the equilibrium distribution is that of a standard Gamma distribution. The model can be interpreted as a simple closed economy in which economic agents trade money and a saving criterion is fixed by the saving propensity λ . Alternatively, from the nature of the equilibrium distribution, we show that the model can also be interpreted as a perfect gas at an effective temperature T(λ) , where particles exchange energy in a space with an effective dimension D(λ) .
Statistical Modeling of Single Target Cell Encapsulation
Moon, SangJun; Ceyhan, Elvan; Gurkan, Umut Atakan; Demirci, Utkan
2011-01-01
High throughput drop-on-demand systems for separation and encapsulation of individual target cells from heterogeneous mixtures of multiple cell types is an emerging method in biotechnology that has broad applications in tissue engineering and regenerative medicine, genomics, and cryobiology. However, cell encapsulation in droplets is a random process that is hard to control. Statistical models can provide an understanding of the underlying processes and estimation of the relevant parameters, and enable reliable and repeatable control over the encapsulation of cells in droplets during the isolation process with high confidence level. We have modeled and experimentally verified a microdroplet-based cell encapsulation process for various combinations of cell loading and target cell concentrations. Here, we explain theoretically and validate experimentally a model to isolate and pattern single target cells from heterogeneous mixtures without using complex peripheral systems. PMID:21814548
The statistical mechanics of several Hamiltonian models
NASA Astrophysics Data System (ADS)
Lee, Chi-Lun
This thesis has two major parts. The first part concerns studies of the equilibrium thermodynamics on different models using a self-consistent Ornstein-Zernike approximation (SCOZA). For most approximate correlation-function theories there exists an inconsistency for thermodynamic quantities evaluated from different thermodynamic routes. In SCOZA one solves this inconsistency through a renormalization procedure, which is based on the enforcement of thermodynamic consistency for quantities evaluated from the energy and the compressibility routes. This procedure has resulted in remarkable accuracy of thermodynamics for most phase regions. We apply several versions of SCOZA to study different models such as the two-dimensional lattice gas, the hard-core Yukawa fluid, and the polymer fluid. Our main objective is to develop a simple non-perturbative approximation that can give accurate results for thermodynamic quantities even when the system stays very close to its critical point. The second part is focused on a study of the protein-folding dynamics using a statistical energy landscape theory. A protein molecule is modelled as a heterogeneous polymer with randomized interaction energies characterized by a statistical distribution. This results in an funnel-like energy landscape with local fluctuations (roughness) and an overall bias towards the folded state. With the introduction of an order parameter, the direction of folding can be characterized. The statistical energy landscape is then mapped into a one-dimensional continuous-time random walk along the order parameter, in which the dynamics is represented through a generalized Fokker-Planck equation. By solving the equation numerically we find a transition from exponential to non-exponential kinetics in the distribution of the first-passage time to the folded state. In our results the non-exponential kinetics has a distribution which resembles a truncated Levy distribution in time.
Some useful statistical methods for model validation.
Marcus, A H; Elias, R W
1998-01-01
Although formal hypothesis tests provide a convenient framework for displaying the statistical results of empirical comparisons, standard tests should not be used without consideration of underlying measurement error structure. As part of the validation process, predictions of individual blood lead concentrations from models with site-specific input parameters are often compared with blood lead concentrations measured in field studies that also report lead concentrations in environmental media (soil, dust, water, paint) as surrogates for exposure. Measurements of these environmental media are subject to several sources of variability, including temporal and spatial sampling, sample preparation and chemical analysis, and data entry or recording. Adjustments for measurement error must be made before statistical tests can be used to empirically compare environmental data with model predictions. This report illustrates the effect of measurement error correction using a real dataset of child blood lead concentrations for an undisclosed midwestern community. We illustrate both the apparent failure of some standard regression tests and the success of adjustment of such tests for measurement error using the SIMEX (simulation-extrapolation) procedure. This procedure adds simulated measurement error to model predictions and then subtracts the total measurement error, analogous to the method of standard additions used by analytical chemists. Images Figure 1 Figure 3 PMID:9860913
Quiet Sun coronal heating: A statistical model
NASA Astrophysics Data System (ADS)
Krasnoselskikh, V.; Podladchikova, O.; Lefebvre, B.; Vilmer, N.
2002-02-01
Recent observations of Krucker & Benz (\\cite{Krucker98}) give strong support to Parker's hypothesis (\\cite∥) that small-scale dissipative events make up the main contribution to quiet Sun coronal heating. They also showed that these small-scale events are associated not only with the magnetic network, but also with the cell interiors (Benz & Krucker \\cite{Benz98}). Taking into account in addition the results of the analysis performed by Priest with co-authors (\\cite{pr1}) who demonstrated that the heating is quasi-homogeneous along the arcs, we come to the conclusion that the sources driving these dissipative events are also small-scale sources. Typically they are of the order of or smaller than the linear scale of the events observed, that is <2000 km. To describe statistical properties of quiet Sun corona heating by microflares, nanoflares, and even smaller events, we consider a cellular automata model subject to uniform small-scale driving and dissipation. The model consists of two elements, the magnetic field source supposed to be associated with the small scale hydrodynamic turbulence convected from the photosphere and local dissipation of small scale currents. The dissipation is assumed to be provided by either anomalous resistivity, when the current density exceeds a certain threshold value, or by the magnetic reconnection. The main problem considered is how the statistical characteristics of dissipated energy flow depend upon characteristics of the magnetic field source and on physical mechanism responsible for the magnetic field dissipation. As the threshold value of current is increased, we observe the transition from Gaussian statistics to power-law type. In addition, we find that the dissipation provided by reconnection results in stronger deviations from Gaussian distribution.
Statistical model semiquantitatively approximates arabinoxylooligosaccharides' structural diversity.
Dotsenko, Gleb; Nielsen, Michael Krogsgaard; Lange, Lene
2016-05-13
A statistical model describing the random distribution of substituted xylopyranosyl residues in arabinoxylooligosaccharides is suggested and compared with existing experimental data. Structural diversity of arabinoxylooligosaccharides of various length, originating from different arabinoxylans (wheat flour arabinoxylan (arabinose/xylose, A/X = 0.47); grass arabinoxylan (A/X = 0.24); wheat straw arabinoxylan (A/X = 0.15); and hydrothermally pretreated wheat straw arabinoxylan (A/X = 0.05)), is semiquantitatively approximated using the proposed model. The suggested approach can be applied not only for prediction and quantification of arabinoxylooligosaccharides' structural diversity, but also for estimate of yield and selection of the optimal source of arabinoxylan for production of arabinoxylooligosaccharides with desired structural features. PMID:27043469
Asaturyan, R.; Ent, R.; Mkrtchyan, H.; Navasardyan, T.; Tadevosyan, V.; Adams, G. S.; Ahmidouch, A.; Angelescu, T.; Arrington, J.; Asaturyan, A.; Baker, O. K.; Benmouna, N.; Bertoncini, C.; Blok, H. P.; Boeglin, W. U.; Bosted, P. E.; Breuer, H.; Christy, M. E.; Connell, S. H.; Cui, Y.; Dalton, M. M.; Danagoulian, S.; Day, D.; Dunne, J. A.; Dutta, D.; El Khayari, N.; Fenker, H. C.; Frolov, V. V.; Gan, L.; Gaskell, D.; Hafidi, K.; Hinton, W.; Holt, R. J.; Horn, T.; Huber, G. M.; Hungerford, E.; Jiang, X.; Jones, M.; Joo, K.; Kalantarians, N.; Kelly, J. J.; Keppel, C. E.; Kubarovsky, V.; Li, Y.; Liang, Y.; Mack, D.; Malace, S. P.; Markowitz, P.; McGrath, E.; McKee, P.; Meekins, D. G.; Mkrtchyan, A.; Moziak, B.; Niculescu, G.; Niculescu, I.; Opper, A. K.; Ostapenko, T.; Reimer, P. E.; Reinhold, J.; Roche, J.; Rock, S. E.; Schulte, E.; Segbefia, E.; Smith, C.; Smith, G. R.; Stoler, P.; Tang, L.; Ungaro, M.; Uzzle, A.; Vidakovic, S.; Villano, A.; Vulcan, W. F.; Wang, M.; Warren, G.; Wesselmann, F. R.; Wojtsekhowski, B.; Wood, S. A.; Xu, C.; Yuan, L.; Zheng, X.
2012-01-01
A large set of cross sections for semi-inclusive electroproduction of charged pions (π^{±}) from both proton and deuteron targets was measured. The data are in the deep-inelastic scattering region with invariant mass squared W^{2} > 4 GeV^{2} and range in four-momentum transfer squared 2 < Q^{2} < 4 (GeV/c)^{2}, and cover a range in the Bjorken scaling variable 0.2 < x < 0.6. The fractional energy of the pions spans a range 0.3 < z < 1, with small transverse momenta with respect to the virtual-photon direction, P_{t}^{2} < 0.2 (GeV/c)^{2}. The invariant mass that goes undetected, M_{x} or W', is in the nucleon resonance region, W' < 2 GeV. The new data conclusively show the onset of quark-hadron duality in this process, and the relation of this phenomenon to the high-energy factorization ansatz of electron-quark scattering and subsequent quark → pion production mechanisms. The x, z and P_{t}^{2} dependences of several ratios (the ratios of favored-unfavored fragmentation functions, charged pion ratios, deuteron-hydrogen and aluminum-deuteron ratios for π^{+} and π^{-}) have been studied. The ratios are found to be in good agreement with expectations based upon a high-energy quark-parton model description. We find the azimuthal dependences to be small, as compared to exclusive pion electroproduction, and consistent with theoretical expectations based on tree-level factorization in terms of transverse-momentum-dependent parton distribution and fragmentation functions. In the context of a simple model, the initial transverse momenta of d quarks are found to be slightly smaller than for u quarks, while the transverse momentum width of the favored fragmentation function is about the same as for the unfavored one, and both fragmentation widths are larger than the quark widths.
Statistical palaeomagnetic field modelling and symmetry considerations
NASA Astrophysics Data System (ADS)
Hulot, G.; Bouligand, C.
2005-06-01
In the present paper, we address symmetry issues in the context of the so-called giant gaussian process (GGP) modelling approach, currently used to statistically analyse the present and past magnetic field of the Earth at times of stable polarity. We first recall the principle of GGP modelling, and for the first time derive the complete and exact constraints a GGP model should satisfy if it is to satisfy statistical spherical, axisymmetrical or equatorially symmetric properties. We note that as often correctly claimed by the authors, many simplifying assumptions used so far to ease the GGP modelling amount to make symmetry assumptions, but not always exactly so, because previous studies did not recognize that symmetry assumptions do not systematically require a lack of cross-correlations between Gauss coefficients. We further note that GGP models obtained so far for the field over the past 5Myr clearly reveal some spherical symmetry breaking properties in both the mean and the fluctuating field (as defined by the covariance matrix of the model) and some equatorial symmetry breaking properties in the mean field. Non-zonal terms found in the mean field of some models and mismatches between variances defining the fluctuating field (in models however not defined in a consistent way) would further suggest that axial symmetry also is broken. The meaning of this is discussed. Spherical symmetry breaking trivially testifies for the influence of the rotation of the Earth on the geodynamo (a long-recognized fact). Axial symmetry breaking, if confirmed, could hardly be attributed to anything else but some influence of the core-mantle boundary (CMB) conditions on the geodynamo (also a well-known fact). By contrast, equatorial symmetry breaking (in particular the persistence of an axial mean quadrupole) may not trivially be considered as evidence of some influence of CMB conditions. To establish this, one would need to better investigate whether or not this axial quadrupole has
Statistical Mechanics Model of Solids with Defects
NASA Astrophysics Data System (ADS)
Kaufman, M.; Walters, P. A.; Ferrante, J.
1997-03-01
Previously(M.Kaufman, J.Ferrante,NASA Tech. Memor.,1996), we examined the phase diagram for the failure of a solid under isotropic expansion and compression as a function of stress and temperature with the "springs" modelled by the universal binding energy relation (UBER)(J.H.Rose, J.R.Smith, F.Guinea, J.Ferrante, Phys.Rev.B29, 2963 (1984)). In the previous calculation we assumed that the "springs" failed independently and that the strain is uniform. In the present work, we have extended this statistical model of mechanical failure by allowing for correlations between "springs" and for thermal fluctuations in strains. The springs are now modelled in the harmonic approximation with a failure threshold energy E0, as an intermediate step in future studies to reinclude the full non-linear dependence of the UBER for modelling the interactions. We use the Migdal-Kadanoff renormalization-group method to determine the phase diagram of the model and to compute the free energy.
NASA Astrophysics Data System (ADS)
Knospe, A. G.; Markert, C.; Werner, K.; Steinheimer, J.; Bleicher, M.
2016-01-01
We study the production of hadronic resonances and their interaction in the partonic and hadronic medium using the EPOS3 model, which employs the UrQMD model for the description of the hadronic phase. We investigate the centrality dependence of the yields and momentum distributions for various resonances [ρ (770) 0 , K*(892) 0 , ϕ (1020 ) , Δ (1232) ++ , Σ (1385) ± , Λ (1520 ) , Ξ (1530) 0 and their antiparticles] in Pb-Pb collisions at √{sN N}= 2.76 TeV. The predictions for K*(892) 0 and ϕ (1020 ) will be compared with the experimental data from the ALICE collaboration. The observed signal suppression of the K*(892) 0 with increasing centrality will be discussed with respect to the resonance interaction in the hadronic medium. The mean transverse momentum and other particle ratios such as ϕ (1020 )/p and (Ω +Ω ¯) /ϕ (1020 ) will be discussed with respect to additional contributions from the hadronic medium interactions.
Parton Saturation and the Color Glass Condensate
NASA Astrophysics Data System (ADS)
Kovchegov, Yuri V.
2007-03-01
We review recent developments in the field of parton saturation and the Color Glass Condensate. We discuss the classical gluon fields of McLerran-Venugopalan model. We explain how small- x non-linear quantum evolution corrections can be included into the total cross section for deep inelastic scattering. We proceed by reviewing saturation physics predictions for the particle production in p( d) A collisions and conclude by demonstrating how such predictions were confirmed by the RHIC experiments.
Generalized Parton Distributions and their Singularities
Anatoly Radyushkin
2011-04-01
A new approach to building models of generalized parton distributions (GPDs) is discussed that is based on the factorized DD (double distribution) Ansatz within the single-DD formalism. The latter was not used before, because reconstructing GPDs from the forward limit one should start in this case with a very singular function $f(\\beta)/\\beta$ rather than with the usual parton density $f(\\beta)$. This results in a non-integrable singularity at $\\beta=0$ exaggerated by the fact that $f(\\beta)$'s, on their own, have a singular $\\beta^{-a}$ Regge behavior for small $\\beta$. It is shown that the singularity is regulated within the GPD model of Szczepaniak et al., in which the Regge behavior is implanted through a subtracted dispersion relation for the hadron-parton scattering amplitude. It is demonstrated that using proper softening of the quark-hadron vertices in the regions of large parton virtualities results in model GPDs $H(x,\\xi)$ that are finite and continuous at the "border point'' $x=\\xi$. Using a simple input forward distribution, we illustrate the implementation of the new approach for explicit construction of model GPDs. As a further development, a more general method of regulating the $\\beta=0$ singularities is proposed that is based on the separation of the initial single DD $f(\\beta, \\alpha)$ into the "plus'' part $[f(\\beta,\\alpha)]_{+}$ and the $D$-term. It is demonstrated that the "DD+D'' separation method allows to (re)derive GPD sum rules that relate the difference between the forward distribution $f(x)=H(x,0)$ and the border function $H(x,x)$ with the $D$-term function $D(\\alpha)$.
Access to generalized parton distributions at COMPASS
Nowak, Wolf-Dieter
2015-04-10
A brief experimentalist's introduction to Generalized Parton Distributions (GPDs) is given. Recent COMPASS results are shown on transverse target-spin asymmetries in hard exclusive ρ{sup 0} production and their interpretation in terms of a phenomenological model as indication for chiral-odd, transverse GPDs is discussed. For deeply virtual Compton scattering, it is briefly outlined how to access GPDs and projections are shown for future COMPASS measurements.
Generalized parton distributions and their singularities
Radyushkin, A. V.
2011-04-01
A new approach to building models of generalized parton distributions (GPDs) is discussed that is based on the factorized DD (double distribution) ansatz within the single-DD formalism. The latter was not used before, because reconstructing GPDs from the forward limit one should start in this case with a very singular function f({beta})/{beta} rather than with the usual parton density f({beta}). This results in a nonintegrable singularity at {beta}=0 exaggerated by the fact that f({beta})'s, on their own, have a singular {beta}{sup -a} Regge behavior for small {beta}. It is shown that the singularity is regulated within the GPD model of Szczepaniak et al., in which the Regge behavior is implanted through a subtracted dispersion relation for the hadron-parton scattering amplitude. It is demonstrated that using proper softening of the quark-hadron vertices in the regions of large parton virtualities results in model GPDs H(x,{xi}) that are finite and continuous at the 'border point' x={xi}. Using a simple input forward distribution, we illustrate implementation of the new approach for explicit construction of model GPDs. As a further development, a more general method of regulating the {beta}=0 singularities is proposed that is based on the separation of the initial single DD f({beta},{alpha}) into the 'plus' part [f({beta},{alpha})]{sub +} and the D term. It is demonstrated that the ''DD+D'' separation method allows one to (re)derive GPD sum rules that relate the difference between the forward distribution f(x)=H(x,0) and the border function H(x,x) with the D-term function D({alpha}).
Strongly interacting parton matter equilibration
Ozvenchuk, V.; Linnyk, O.; Bratkovskaya, E.; Gorenstein, M.; Cassing, W.
2012-07-15
We study the kinetic and chemical equilibration in 'infinite' parton matter within the Parton-Hadron-String Dynamics transport approach. The 'infinite' matter is simulated within a cubic box with periodic boundary conditions initialized at different energy densities. Particle abundances, kinetic energy distributions, and the detailed balance of the off-shell quarks and gluons in the strongly-interacting quarkgluon plasma are addressed and discussed.
Assessing Statistical Model Assumptions under Climate Change
NASA Astrophysics Data System (ADS)
Varotsos, Konstantinos V.; Giannakopoulos, Christos; Tombrou, Maria
2016-04-01
The majority of the studies assesses climate change impacts on air-quality using chemical transport models coupled to climate ones in an off-line mode, for various horizontal resolutions and different present and future time slices. A complementary approach is based on present-day empirical relations between air-pollutants and various meteorological variables which are then extrapolated to the future. However, the extrapolation relies on various assumptions such as that these relationships will retain their main characteristics in the future. In this study we focus on the ozone-temperature relationship. It is well known that among a number of meteorological variables, temperature is found to exhibit the highest correlation with ozone concentrations. This has led, in the past years, to the development and application of statistical models with which the potential impact of increasing future temperatures on various ozone statistical targets was examined. To examine whether the ozone-temperature relationship retains its main characteristics under warmer temperatures we analyze the relationship during the heatwaves events of 2003 and 2006 in Europe. More specifically, we use available gridded daily maximum temperatures (E-OBS) and hourly ozone observations from different non-urban stations (EMEP) within the areas that were impacted from the two heatwave events. In addition, we compare the temperature distributions of the two events with temperatures from two different future time periods 2021-2050 and 2071-2100 from a number of regional climate models developed under the framework of the Cordex initiative (http://www.cordex.org) with a horizontal resolution of 12 x 12km, based on different IPCC RCPs emissions scenarios. A statistical analysis is performed on the ozone-temperature relationship for each station and for the two aforementioned years which are then compared against the ozone-temperature relationships obtained from the rest of the available dataseries. The
Statistical Shape Modeling of Cam Femoroacetabular Impingement
Harris, Michael D.; Dater, Manasi; Whitaker, Ross; Jurrus, Elizabeth R.; Peters, Christopher L.; Anderson, Andrew E.
2013-10-01
In this study, statistical shape modeling (SSM) was used to quantify three-dimensional (3D) variation and morphologic differences between femurs with and without cam femoroacetabular impingement (FAI). 3D surfaces were generated from CT scans of femurs from 41 controls and 30 cam FAI patients. SSM correspondence particles were optimally positioned on each surface using a gradient descent energy function. Mean shapes for control and patient groups were defined from the resulting particle configurations. Morphological differences between group mean shapes and between the control mean and individual patients were calculated. Principal component analysis was used to describe anatomical variation present in both groups. The first 6 modes (or principal components) captured statistically significant shape variations, which comprised 84% of cumulative variation among the femurs. Shape variation was greatest in femoral offset, greater trochanter height, and the head-neck junction. The mean cam femur shape protruded above the control mean by a maximum of 3.3 mm with sustained protrusions of 2.5-3.0 mm along the anterolateral head-neck junction and distally along the anterior neck, corresponding well with reported cam lesion locations and soft-tissue damage. This study provides initial evidence that SSM can describe variations in femoral morphology in both controls and cam FAI patients and may be useful for developing new measurements of pathological anatomy. SSM may also be applied to characterize cam FAI severity and provide templates to guide patient-specific surgical resection of bone.
Jet fragmentation via recombination of parton showers
NASA Astrophysics Data System (ADS)
Han, Kyong Chol; Fries, Rainer J.; Ko, Che Ming
2016-04-01
We propose to model hadronization of parton showers in QCD jets through a hybrid approach involving quark recombination and string fragmentation. This is achieved by allowing gluons at the end of the perturbative shower evolution to undergo a nonperturbative splitting into quark and antiquark pairs, then applying a Monte Carlo version of instantaneous quark recombination, and finally subjecting remnant quarks (those which have not found a recombination partner) to Lund string fragmentation. When applied to parton showers from the pythia Monte Carlo event generator, the final hadron spectra from our calculation compare quite well to pythia jets that have been hadronized with the default Lund string fragmentation. Our new approach opens up the possibility to generalize hadronization to jets embedded in a quark gluon plasma.
A statistical model for predicting muscle performance
NASA Astrophysics Data System (ADS)
Byerly, Diane Leslie De Caix
The objective of these studies was to develop a capability for predicting muscle performance and fatigue to be utilized for both space- and ground-based applications. To develop this predictive model, healthy test subjects performed a defined, repetitive dynamic exercise to failure using a Lordex spinal machine. Throughout the exercise, surface electromyography (SEMG) data were collected from the erector spinae using a Mega Electronics ME3000 muscle tester and surface electrodes placed on both sides of the back muscle. These data were analyzed using a 5th order Autoregressive (AR) model and statistical regression analysis. It was determined that an AR derived parameter, the mean average magnitude of AR poles, significantly correlated with the maximum number of repetitions (designated Rmax) that a test subject was able to perform. Using the mean average magnitude of AR poles, a test subject's performance to failure could be predicted as early as the sixth repetition of the exercise. This predictive model has the potential to provide a basis for improving post-space flight recovery, monitoring muscle atrophy in astronauts and assessing the effectiveness of countermeasures, monitoring astronaut performance and fatigue during Extravehicular Activity (EVA) operations, providing pre-flight assessment of the ability of an EVA crewmember to perform a given task, improving the design of training protocols and simulations for strenuous International Space Station assembly EVA, and enabling EVA work task sequences to be planned enhancing astronaut performance and safety. Potential ground-based, medical applications of the predictive model include monitoring muscle deterioration and performance resulting from illness, establishing safety guidelines in the industry for repetitive tasks, monitoring the stages of rehabilitation for muscle-related injuries sustained in sports and accidents, and enhancing athletic performance through improved training protocols while reducing
Nonparametric statistical modeling of binary star separations
NASA Astrophysics Data System (ADS)
Heacox, William D.; Gathright, John
1994-09-01
We develop a comprehensive statistical model for the distribution of observed separations in binary star systems, in terms of distributions of orbital elements, projection effects, and distances to systems. We use this model to derive several diagnostics for estimating the completeness of imaging searches for stellar companions, and the underlying stellar multiplicities. In application to recent imaging searches for low-luminosity companions to nearby M dwarf stars, and for companions to young stars in nearby star-forming regions, our analyses reveal substantial uncertainty in estimates of stellar multiplicity. For binary stars with late-type dwarf companions, semimajor axes appear to be distributed approximately as a-1 for values ranging from about one to several thousand astronomical units. About one-quarter of the companions to field F and G dwarf stars have semimajor axes less than 1 AU, and about 15% lie beyond 1000 AU. The geometric efficiency (fraction of companions imaged onto the detector) of imaging searches is nearly independent of distances to program stars and orbital eccentricities, and varies only slowly with detector spatial limitations.
Nonparametric statistical modeling of binary star separations
NASA Technical Reports Server (NTRS)
Heacox, William D.; Gathright, John
1994-01-01
We develop a comprehensive statistical model for the distribution of observed separations in binary star systems, in terms of distributions of orbital elements, projection effects, and distances to systems. We use this model to derive several diagnostics for estimating the completeness of imaging searches for stellar companions, and the underlying stellar multiplicities. In application to recent imaging searches for low-luminosity companions to nearby M dwarf stars, and for companions to young stars in nearby star-forming regions, our analyses reveal substantial uncertainty in estimates of stellar multiplicity. For binary stars with late-type dwarf companions, semimajor axes appear to be distributed approximately as a(exp -1) for values ranging from about one to several thousand astronomical units. About one-quarter of the companions to field F and G dwarf stars have semimajor axes less than 1 AU, and about 15% lie beyond 1000 AU. The geometric efficiency (fraction of companions imaged onto the detector) of imaging searches is nearly independent of distances to program stars and orbital eccentricities, and varies only slowly with detector spatial limitations.
Fermi-Dirac distributions for quark partons
NASA Astrophysics Data System (ADS)
Bourrely, C.; Buccella, F.; Miele, G.; Migliore, G.; Soffer, J.; Tibullo, V.
1994-09-01
We propose to use Fermi-Dirac distributions for quark and antiquark partons. It allows a fair description of the x-dependence of the very recent NMC data on the proton and neutron structure functions F {2/ p } (x) and F {2/ n } (x) at Q 2=4 GeV2, as well as the CCFR antiquark distributionxbar q(x). We show that one can also use a corresponding Bose-Einstein expression to describe consistently the gluon distribution. The Pauli exclusion principle, which has been identified to explain the flavor asymmetry of the light-quark sea of the proton, is advocated to guide us for making a simple construction of the polarized parton distributions. We predict the spin dependent structure functions g {1/ p } (x) and g {1/ n } (x) in good agreement with EMC and SLAC data. The quark distributions involve some parameters whose values support well the hypothesis that the violation of the quark parton model sum rules is a consequence of the Pauli principle.
Pathway Model and Nonextensive Statistical Mechanics
NASA Astrophysics Data System (ADS)
Mathai, A. M.; Haubold, H. J.; Tsallis, C.
2015-12-01
The established technique of eliminating upper or lower parameters in a general hypergeometric series is profitably exploited to create pathways among confluent hypergeometric functions, binomial functions, Bessel functions, and exponential series. One such pathway, from the mathematical statistics point of view, results in distributions which naturally emerge within nonextensive statistical mechanics and Beck-Cohen superstatistics, as pursued in generalizations of Boltzmann-Gibbs statistics.
Statistical modeling of global soil NOx emissions
NASA Astrophysics Data System (ADS)
Yan, Xiaoyuan; Ohara, Toshimasa; Akimoto, Hajime
2005-09-01
On the basis of field measurements of NOx emissions from soils, we developed a statistical model to describe the influences of soil organic carbon (SOC) content, soil pH, land-cover type, climate, and nitrogen input on NOx emission. While also considering the effects of soil temperature, soil moisture change-induced pulse emission, and vegetation fire, we simulated NOx emissions from global soils at resolutions of 0.5° and 6 hours. Canopy reduction was included in both data processing and flux simulation. NOx emissions were positively correlated with SOC content and negatively correlated with soil pH. Soils in dry or temperate regions had higher NOx emission potentials than soils in cold or tropical regions. Needleleaf forest and agricultural soils had high NOx emissions. The annual NOx emission from global soils was calculated to be 7.43 Tg N, decreasing to 4.97 Tg N after canopy reduction. Global averages of nitrogen fertilizer-induced emission ratios were 1.16% above soil and 0.70% above canopy. Soil moisture change-induced pulse emission contributed about 4% to global annual NOx emission, and the effect of vegetation fire on soil NOx emission was negligible.
Tests of models for parton fragmentation in e e annihilation. [29 GeV center-of-mass energy
Gary, J.W.
1985-11-01
We examine the distribution of particles in the three jet events of e e annihilation. The data was collected with the PEP-4/Time Projection Chamber detector at 29 GeV center-of-mass energy at PEP. The experimental distributions are compared to the predictions of several fragmentation models which describe the transition of quarks and gluons into hadrons. In particular, our study emphasizes the three fragmentation models which are currently in widest use: the Lund string model, the Webber cluster model and the independent fragmentation model. These three models each possess different Lorentz frame structures for the distribution of hadron sources relative to the overall event c.m. in three jet events. The Lund string and independent fragmentation models are tuned to describe global event properties of our multihadronic annihilation event sample. This tuned Lund string model provides a good description of the distribution of particles between jet axes in three jet events, while the independent fragmentation model does not. We verify that the failure of the independent fragmentation model is not a consequence of parameter tuning or of model variant. The Webber cluster model, which is untuned, does not describe the absolute particle densities between jets but correctly predicts the ratios of those densities, which are less sensitive to the tuning. These results provide evidence that the sources of hadrons are boosted with respect to the overall center-of-mass in three jet events, with components of motion normal to the jet axes. The distribution of particles close to jet axes provides additional support for this conclusion. 94 refs.
Medium Effects in Parton Distributions
William Detmold, Huey-Wen Lin
2011-12-01
A defining experiment of high-energy physics in the 1980s was that of the EMC collaboration where it was first observed that parton distributions in nuclei are non-trivially related to those in the proton. This result implies that the presence of the nuclear medium plays an important role and an understanding of this from QCD has been an important goal ever since Here we investigate analogous, but technically simpler, effects in QCD and examine how the lowest moment of the pion parton distribution is modified by the presence of a Bose-condensed gas of pions or kaons.
Parton distributions from lattice QCD: an update
Detmold, W; Melnitchouk, W; Thomas, A W
2004-04-01
We review the extraction of parton distributions from their moments calculated in lattice QCD, focusing in particular on their extrapolation to the physical region. As examples, we consider both the unpolarized and polarized isovector parton distributions of the nucleon.
DOSE-RESPONSE ASSESSMENT FOR DEVELOPMENTAL TOXICITY: III. STATISTICAL MODELS
Although quantitative modeling has been central to cancer risk assessment for years, the concept of dose-response modeling for developmental effects is relatively new. Recently, statistical models appropriate for developmental toxicity testing have been developed and applied (Rai...
Modeling Human Performance in Statistical Word Segmentation
ERIC Educational Resources Information Center
Frank, Michael C.; Goldwater, Sharon; Griffiths, Thomas L.; Tenenbaum, Joshua B.
2010-01-01
The ability to discover groupings in continuous stimuli on the basis of distributional information is present across species and across perceptual modalities. We investigate the nature of the computations underlying this ability using statistical word segmentation experiments in which we vary the length of sentences, the amount of exposure, and…
Experimental consistency in parton distribution fitting
Pumplin, Jon
2010-04-01
The recently developed 'data set diagonalization' method is applied to measure compatibility of the data sets that are used to determine parton distribution functions. Discrepancies among the experiments are found to be somewhat larger than is predicted by propagating the published experimental errors according to Gaussian statistics. The results support a tolerance criterion of {Delta}{chi}{sup 2{approx_equal}}10 to estimate the 90% confidence range for parton distribution function uncertainties. No basis is found in the data sets for the larger {Delta}{chi}{sup 2} values that are in current use, though it may be necessary to retain those larger values until improved methods can be developed to take account of systematic errors in applying the theory, including the effect of parametrization dependence. The data set diagonalization method also measures how much influence each experiment has on the global fit and identifies experiments that show significant tension with respect to the others. The method is used to explore the contribution from muon scattering experiments, which are found to exhibit the largest discrepancies in the current fit.
Power Curve Modeling in Complex Terrain Using Statistical Models
NASA Astrophysics Data System (ADS)
Bulaevskaya, V.; Wharton, S.; Clifton, A.; Qualley, G.; Miller, W.
2014-12-01
Traditional power output curves typically model power only as a function of the wind speed at the turbine hub height. While the latter is an essential predictor of power output, wind speed information in other parts of the vertical profile, as well as additional atmospheric variables, are also important determinants of power. The goal of this work was to determine the gain in predictive ability afforded by adding wind speed information at other heights, as well as other atmospheric variables, to the power prediction model. Using data from a wind farm with a moderately complex terrain in the Altamont Pass region in California, we trained three statistical models, a neural network, a random forest and a Gaussian process model, to predict power output from various sets of aforementioned predictors. The comparison of these predictions to the observed power data revealed that considerable improvements in prediction accuracy can be achieved both through the addition of predictors other than the hub-height wind speed and the use of statistical models. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under contract DE-AC52-07NA27344 and was funded by Wind Uncertainty Quantification Laboratory Directed Research and Development Project at LLNL under project tracking code 12-ERD-069.
Fiori, Simone
2007-01-01
Bivariate statistical modeling from incomplete data is a useful statistical tool that allows to discover the model underlying two data sets when the data in the two sets do not correspond in size nor in ordering. Such situation may occur when the sizes of the two data sets do not match (i.e., there are “holes” in the data) or when the data sets have been acquired independently. Also, statistical modeling is useful when the amount of available data is enough to show relevant statistical features of the phenomenon underlying the data. We propose to tackle the problem of statistical modeling via a neural (nonlinear) system that is able to match its input-output statistic to the statistic of the available data sets. A key point of the new implementation proposed here is that it is based on look-up-table (LUT) neural systems, which guarantee a computationally advantageous way of implementing neural systems. A number of numerical experiments, performed on both synthetic and real-world data sets, illustrate the features of the proposed modeling procedure. PMID:18566641
First moments of nucleon generalized parton distributions
Wang, P.; Thomas, A. W.
2010-06-01
We extrapolate the first moments of the generalized parton distributions using heavy baryon chiral perturbation theory. The calculation is based on the one loop level with the finite range regularization. The description of the lattice data is satisfactory, and the extrapolated moments at physical pion mass are consistent with the results obtained with dimensional regularization, although the extrapolation in the momentum transfer to t=0 does show sensitivity to form factor effects, which lie outside the realm of chiral perturbation theory. We discuss the significance of the results in the light of modern experiments as well as QCD inspired models.
Electroweak boson production in double parton scattering
NASA Astrophysics Data System (ADS)
Golec-Biernat, Krzysztof; Lewandowska, Emilia
2014-11-01
We study the W+W- and Z0Z0 electroweak boson production in double parton scattering using QCD evolution equations for double parton distributions. In particular, we analyze the impact of splitting terms in the evolution equations on the double parton scattering cross sections. Unlike the standard terms, the splitting terms are not suppressed for large values of the relative momentum of two partons in the double parton scattering. Thus, they play an important role which we discuss in detail for the single splitting contribution to the cross sections under the study.
The Long Way to the Statistical Bootstrap Model: 1994
NASA Astrophysics Data System (ADS)
Hagedorn, Rolf
I describe the long way from the first theoretical ideas about multiple particle production up to the situation in which constructing of a statistical model of strong interactions seemed natural. I begin in 1936, and argue that the statistical method came to be from a large network of observations and theoretical ideas. I shall pick up only a few primary lines, chosen for their common end point: the statistical bootstrap model of 1964/65.
Infinite statistics condensate as a model of dark matter
Ebadi, Zahra; Mirza, Behrouz; Mohammadzadeh, Hosein E-mail: b.mirza@cc.iut.ac.ir
2013-11-01
In some models, dark matter is considered as a condensate bosonic system. In this paper, we prove that condensation is also possible for particles that obey infinite statistics and derive the critical condensation temperature. We argue that a condensed state of a gas of very weakly interacting particles obeying infinite statistics could be considered as a consistent model of dark matter.
A statistical model for landfill surface emissions.
Héroux, Martin; Guy, Christophe; Millette, Denis
2010-02-01
Landfill operators require a rapid, simple, low-cost, and accurate method for estimation of landfill methane surface emissions over time. Several methods have been developed to obtain instantaneous field measurements of landfill methane surface emissions. This paper provides a methodology for interpolating instantaneous measurements over time, taking variations in meteorological conditions into account. The goal of this study was to determine the effects of three factors on landfill methane surface emissions: air temperature, pressure gradient between waste and atmosphere, and soil moisture content of the cover material. On the basis of a statistical three-factor and two-level full factorial design, field measurements of methane emissions were conducted at the City of Montreal landfill site during the summer of 2004. Three areas were measured: test area 1 (4800 m2), test area 2 (1400 m2), and test area 3 (1000 m2). Analyses of variance were performed on the data. They showed a significant statistical effect of the three factors and the interaction between temperature and soil moisture content on methane emissions. Analysis also led to the development of a multifactor correlation, which can be explained by the underlying processes of diffusive and advective flow and biological oxidation. This correlation was used to estimate total emissions of the three test areas for July and August 2004. The approach was validated using a second dataset for another area adjacent to the landfill. PMID:20222535
Statistical Modeling of Large-Scale Scientific Simulation Data
Eliassi-Rad, T; Baldwin, C; Abdulla, G; Critchlow, T
2003-11-15
With the advent of massively parallel computer systems, scientists are now able to simulate complex phenomena (e.g., explosions of a stars). Such scientific simulations typically generate large-scale data sets over the spatio-temporal space. Unfortunately, the sheer sizes of the generated data sets make efficient exploration of them impossible. Constructing queriable statistical models is an essential step in helping scientists glean new insight from their computer simulations. We define queriable statistical models to be descriptive statistics that (1) summarize and describe the data within a user-defined modeling error, and (2) are able to answer complex range-based queries over the spatiotemporal dimensions. In this chapter, we describe systems that build queriable statistical models for large-scale scientific simulation data sets. In particular, we present our Ad-hoc Queries for Simulation (AQSim) infrastructure, which reduces the data storage requirements and query access times by (1) creating and storing queriable statistical models of the data at multiple resolutions, and (2) evaluating queries on these models of the data instead of the entire data set. Within AQSim, we focus on three simple but effective statistical modeling techniques. AQSim's first modeling technique (called univariate mean modeler) computes the ''true'' (unbiased) mean of systematic partitions of the data. AQSim's second statistical modeling technique (called univariate goodness-of-fit modeler) uses the Andersen-Darling goodness-of-fit method on systematic partitions of the data. Finally, AQSim's third statistical modeling technique (called multivariate clusterer) utilizes the cosine similarity measure to cluster the data into similar groups. Our experimental evaluations on several scientific simulation data sets illustrate the value of using these statistical models on large-scale simulation data sets.
Exponential order statistic models of software reliability growth
NASA Technical Reports Server (NTRS)
Miller, D. R.
1985-01-01
Failure times of a software reliabilty growth process are modeled as order statistics of independent, nonidentically distributed exponential random variables. The Jelinsky-Moranda, Goel-Okumoto, Littlewood, Musa-Okumoto Logarithmic, and Power Law models are all special cases of Exponential Order Statistic Models, but there are many additional examples also. Various characterizations, properties and examples of this class of models are developed and presented.
Exponential order statistic models of software reliability growth
NASA Technical Reports Server (NTRS)
Miller, D. R.
1986-01-01
Failure times of a software reliability growth process are modeled as order statistics of independent, nonidentically distributed exponential random variables. The Jelinsky-Moranda, Goel-Okumoto, Littlewood, Musa-Okumoto Logarithmic, and Power Law models are all special cases of Exponential Order Statistic Models, but there are many additional examples also. Various characterizations, properties and examples of this class of models are developed and presented.
Statistical Methods with Varying Coefficient Models
Fan, Jianqing; Zhang, Wenyang
2008-01-01
The varying coefficient models are very important tool to explore the dynamic pattern in many scientific areas, such as economics, finance, politics, epidemiology, medical science, ecology and so on. They are natural extensions of classical parametric models with good interpretability and are becoming more and more popular in data analysis. Thanks to their flexibility and interpretability, in the past ten years, the varying coefficient models have experienced deep and exciting developments on methodological, theoretical and applied sides. This paper gives a selective overview on the major methodological and theoretical developments on the varying coefficient models. PMID:18978950
Deviance statistics in model fit and selection in ROC studies
NASA Astrophysics Data System (ADS)
Lei, Tianhu; Bae, K. Ty
2013-03-01
A general non-linear regression model-based Bayesian inference approach is used in our ROC (Receiver Operating Characteristics) study. In the sampling of posterior distribution, two prior models - continuous Gaussian and discrete categorical - are used for the scale parameter. How to judge Goodness-of-Fit (GOF) of each model and how to criticize these two models, Deviance statistics and Deviance information criterion (DIC) are adopted to address these problems. Model fit and model selection focus on the adequacy of models. Judging model adequacy is essentially measuring agreement of model and observations. Deviance statistics and DIC provide overall measures on model fit and selection. In order to investigate model fit at each category of observations, we find that the cumulative, exponential contributions from individual observations to Deviance statistics are good estimates of FPF (false positive fraction) and TPF (true positive fraction) on which the ROC curve is based. This finding further leads to a new measure for model fit, called FPF-TPF distance, which is an Euclidean distance defined on FPF-TPF space. It combines both local and global fitting. Deviance statistics and FPFTPF distance are shown to be consistent and in good agreement. Theoretical derivation and numerical simulations for this new method for model fit and model selection of ROC data analysis are included. Keywords: General non-linear regression model, Bayesian Inference, Markov Chain Monte Carlo (MCMC) method, Goodness-of-Fit (GOF), Model selection, Deviance statistics, Deviance information criterion (DIC), Continuous conjugate prior, Discrete categorical prior. ∗
Generalized parton correlation functions for a spin-1/2 hadron
Stephan Meissner, Andreas Metz, Marc Schlegel
2009-08-01
The fully unintegrated, off-diagonal quark-quark correlator for a spin-1/2 hadron is parameterized in terms of so-called generalized parton correlation functions. Such objects, in particular, can be considered as mother distributions of generalized parton distributions on the one hand and transverse momentum dependent parton distributions on the other. Therefore, our study provides new, model-independent insights into the recently proposed nontrivial relations between generalized and transverse momentum dependent parton distributions. We find that none of these relations can be promoted to a model-independent status. As a by-product we obtain the first complete classification of generalized parton distributions beyond leading twist. The present paper is a natural extension of our previous corresponding analysis for spin-0 hadrons.
Structure functions and parton distributions
Martin, A.D.; Stirling, W.J.; Roberts, R.G.
1995-07-01
The MRS parton distribution analysis is described. The latest sets are shown to give an excellent description of a wide range of deep-inelastic and other hard scattering data. Two important theoretical issues-the behavior of the distributions at small x and the flavor structure of the quark sea-are discussed in detail. A comparison with the new structure function data from HERA is made, and the outlook for the future is discussed.
Generalized parton distributions in nuclei
Vadim Guzey
2009-12-01
Generalized parton distributions (GPDs) of nuclei describe the distribution of quarks and gluons in nuclei probed in hard exclusive reactions, such as e.g. deeply virtual Compton scattering (DVCS). Nuclear GPDs and nuclear DVCS allow us to study new aspects of many traditional nuclear effects (nuclear shadowing, EMC effect, medium modifications of the bound nucleons) as well as to access novel nuclear effects. In my talk, I review recent theoretical progress in the area of nuclear GPDs.
Statistical analysis of synaptic transmission: model discrimination and confidence limits.
Stricker, C; Redman, S; Daley, D
1994-01-01
Procedures for discriminating between competing statistical models of synaptic transmission, and for providing confidence limits on the parameters of these models, have been developed. These procedures were tested against simulated data and were used to analyze the fluctuations in synaptic currents evoked in hippocampal neurones. All models were fitted to data using the Expectation-Maximization algorithm and a maximum likelihood criterion. Competing models were evaluated using the log-likelihood ratio (Wilks statistic). When the competing models were not nested, Monte Carlo sampling of the model used as the null hypothesis (H0) provided density functions against which H0 and the alternate model (H1) were tested. The statistic for the log-likelihood ratio was determined from the fit of H0 and H1 to these probability densities. This statistic was used to determine the significance level at which H0 could be rejected for the original data. When the competing models were nested, log-likelihood ratios and the chi 2 statistic were used to determine the confidence level for rejection. Once the model that provided the best statistical fit to the data was identified, many estimates for the model parameters were calculated by resampling the original data. Bootstrap techniques were then used to obtain the confidence limits of these parameters. PMID:7948672
Links to sources of cancer-related statistics, including the Surveillance, Epidemiology and End Results (SEER) Program, SEER-Medicare datasets, cancer survivor prevalence data, and the Cancer Trends Progress Report.
Bivariate statistical modeling of color and range in natural scenes
NASA Astrophysics Data System (ADS)
Su, Che-Chun; Cormack, Lawrence K.; Bovik, Alan C.
2014-02-01
The statistical properties embedded in visual stimuli from the surrounding environment guide and affect the evolutionary processes of human vision systems. There are strong statistical relationships between co-located luminance/chrominance and disparity bandpass coefficients in natural scenes. However, these statistical rela- tionships have only been deeply developed to create point-wise statistical models, although there exist spatial dependencies between adjacent pixels in both 2D color images and range maps. Here we study the bivariate statistics of the joint and conditional distributions of spatially adjacent bandpass responses on both luminance/chrominance and range data of naturalistic scenes. We deploy bivariate generalized Gaussian distributions to model the underlying statistics. The analysis and modeling results show that there exist important and useful statistical properties of both joint and conditional distributions, which can be reliably described by the corresponding bivariate generalized Gaussian models. Furthermore, by utilizing these robust bivariate models, we are able to incorporate measurements of bivariate statistics between spatially adjacent luminance/chrominance and range information into various 3D image/video and computer vision applications, e.g., quality assessment, 2D-to-3D conversion, etc.
Statistical Contact Model for Confined Molecules
NASA Astrophysics Data System (ADS)
Santamaria, Ruben; de la Paz, Antonio Alvarez; Roskop, Luke; Adamowicz, Ludwik
2016-08-01
A theory that describes in a realistic form a system of atoms under the effects of temperature and confinement is presented. The theory departs from a Lagrangian of the Zwanzig type and contains the main ingredients for describing a system of atoms immersed in a heat bath that is also formed by atoms. The equations of motion are derived according to Lagrangian mechanics. The application of statistical mechanics to describe the bulk effects greatly reduces the complexity of the equations. The resultant equations of motion are of the Langevin type with the viscosity and the temperature of the heat reservoir able to influence the trajectories of the particles. The pressure effects are introduced mechanically by using a container with an atomic structure immersed in the heat bath. The relevant variables that determine the equation of state are included in the formulation. The theory is illustrated by the derivation of the equation of state for a system with 76 atoms confined inside of a 180-atom fullerene-like cage that is immersed in fluid forming the heat bath at a temperature of 350 K and with the friction coefficient of 3.0 {ps}^{-1}. The atoms are of the type believed to form the cores of the Uranus and Neptune planets. The dynamic and the static pressures of the confined system are varied in the 3-5 KBar and 2-30 MBar ranges, respectively. The formulation can be equally used to analyze chemical reactions under specific conditions of pressure and temperature, determine the structure of clusters with their corresponding equation of state, the conditions for hydrogen storage, etc. The theory is consistent with the principles of thermodynamics and it is intrinsically ergodic, of general use, and the first of this kind.
Statistical Contact Model for Confined Molecules
NASA Astrophysics Data System (ADS)
Santamaria, Ruben; de la Paz, Antonio Alvarez; Roskop, Luke; Adamowicz, Ludwik
2016-06-01
A theory that describes in a realistic form a system of atoms under the effects of temperature and confinement is presented. The theory departs from a Lagrangian of the Zwanzig type and contains the main ingredients for describing a system of atoms immersed in a heat bath that is also formed by atoms. The equations of motion are derived according to Lagrangian mechanics. The application of statistical mechanics to describe the bulk effects greatly reduces the complexity of the equations. The resultant equations of motion are of the Langevin type with the viscosity and the temperature of the heat reservoir able to influence the trajectories of the particles. The pressure effects are introduced mechanically by using a container with an atomic structure immersed in the heat bath. The relevant variables that determine the equation of state are included in the formulation. The theory is illustrated by the derivation of the equation of state for a system with 76 atoms confined inside of a 180-atom fullerene-like cage that is immersed in fluid forming the heat bath at a temperature of 350 K and with the friction coefficient of 3.0 {ps}^{-1} . The atoms are of the type believed to form the cores of the Uranus and Neptune planets. The dynamic and the static pressures of the confined system are varied in the 3-5 KBar and 2-30 MBar ranges, respectively. The formulation can be equally used to analyze chemical reactions under specific conditions of pressure and temperature, determine the structure of clusters with their corresponding equation of state, the conditions for hydrogen storage, etc. The theory is consistent with the principles of thermodynamics and it is intrinsically ergodic, of general use, and the first of this kind.
Statistical physics models for nacre fracture simulation
NASA Astrophysics Data System (ADS)
Nukala, Phani Kumar V. V.; Šimunović, Srđan
2005-10-01
Natural biological materials such as nacre (or mother-of-pearl), exhibit phenomenal fracture strength and toughness properties despite the brittle nature of their constituents. For example, nacre’s work of fracture is three orders of magnitude greater than that of a single crystal of its constituent mineral. This study investigates the fracture properties of nacre using a simple discrete lattice model based on continuous damage random thresholds fuse network. The discrete lattice topology of the proposed model is based on nacre’s unique brick and mortar microarchitecture, and the mechanical behavior of each of the bonds in the discrete lattice model is governed by the characteristic modular damage evolution of the organic matrix that includes the mineral bridges between the aragonite platelets. The analysis indicates that the excellent fracture properties of nacre are a result of their unique microarchitecture, repeated unfolding of protein molecules (modular damage evolution) in the organic polymer, and the presence of fiber bundle of mineral bridges between the aragonite platelets. The numerical results obtained using this simple discrete lattice model are in excellent agreement with the previously obtained experimental results, such as nacre’s stiffness, tensile strength, and work of fracture.
Structured Statistical Models of Inductive Reasoning
ERIC Educational Resources Information Center
Kemp, Charles; Tenenbaum, Joshua B.
2009-01-01
Everyday inductive inferences are often guided by rich background knowledge. Formal models of induction should aim to incorporate this knowledge and should explain how different kinds of knowledge lead to the distinctive patterns of reasoning found in different inductive contexts. This article presents a Bayesian framework that attempts to meet…
Statistical physics models for nacre fracture simulation.
Nukala, Phani Kumar V V; Simunović, Srdan
2005-10-01
Natural biological materials such as nacre (or mother-of-pearl), exhibit phenomenal fracture strength and toughness properties despite the brittle nature of their constituents. For example, nacre's work of fracture is three orders of magnitude greater than that of a single crystal of its constituent mineral. This study investigates the fracture properties of nacre using a simple discrete lattice model based on continuous damage random thresholds fuse network. The discrete lattice topology of the proposed model is based on nacre's unique brick and mortar microarchitecture, and the mechanical behavior of each of the bonds in the discrete lattice model is governed by the characteristic modular damage evolution of the organic matrix that includes the mineral bridges between the aragonite platelets. The analysis indicates that the excellent fracture properties of nacre are a result of their unique microarchitecture, repeated unfolding of protein molecules (modular damage evolution) in the organic polymer, and the presence of fiber bundle of mineral bridges between the aragonite platelets. The numerical results obtained using this simple discrete lattice model are in excellent agreement with the previously obtained experimental results, such as nacre's stiffness, tensile strength, and work of fracture. PMID:16383432
Statistical mechanical models of virus capsid assembly
NASA Astrophysics Data System (ADS)
Hicks, Stephen Daniel
Viruses have become an increasingly popular subject of physics investigation, particularly in the last decade. Advances in imaging of virus capsids---the protective protein shells---in a wide variety of stages of assembly have encouraged physical assembly models at a similarly wide variety of scales, while the apparent simplicity of the capsid system---typically, many identical units assembling spontaneously into an icosahedrally symmetric (rather than amorphous) shell---makes the problem particularly interesting. We take a look at the existing physical assembly models in light of the question of how a particular assembly target can be consistently achieved in the presence of so many possible incorrect results. This review leads us to pose our own model of fully irreversible virus assembly, which we study in depth using a large ensemble of simulated assembled capsids, generated under a variety of capsid shell elastic parameters. While this irreversible model (predictably) did not yield consistently symmetric results, we do glean some insight into the effect of elasticity on growth, as well as an understanding of common failure modes. In particular, we found that (i) capsid size depends strongly on the spontaneous curvature and weakly on the ratio of bending to stretching elastic stiffnesses, (ii) the probability of successful capsid completion decays exponentially with capsid size, and (iii) the degree of localization of Gaussian curvature depends heavily on the ratio of elastic stiffnesses. We then go on to consider more thoroughly the nature of the ensemble of symmetric and almost-symmetric capsids---ultimately computing a phase diagram of minimum-energy capsids as a function of the two above-mentioned elastic parameters---and also look at a number of modifications we can make to our irreversible model, finally putting forth a rather different type of model potentially appropriate for understanding immature HIV assembly, and concluding with a fit of this new
Parton-parton elastic scattering and rapidity gaps at SSC and LHC energies
Duca, V.D.
1993-08-01
The theory of the perturbative pomeron, due to Lipatov and collaborators, is used to compute the probability of observing parton-parton elastic scattering and rapidity gaps between jets in hadron collisions at SSC and LHC energies.
Parton-parton elastic scattering and rapidity gaps at Tevatron energies
Del Duca, V.; Tang, Wai-Keung
1993-08-01
The theory of the perturbative pomeron, due to Lipatov and collaborators, is used to compute the probability of observing parton-parton elastic scattering and rapidity gaps between jets in hadron collisions at Tevatron energies.
A Model of Statistics Performance Based on Achievement Goal Theory.
ERIC Educational Resources Information Center
Bandalos, Deborah L.; Finney, Sara J.; Geske, Jenenne A.
2003-01-01
Tests a model of statistics performance based on achievement goal theory. Both learning and performance goals affected achievement indirectly through study strategies, self-efficacy, and test anxiety. Implications of these findings for teaching and learning statistics are discussed. (Contains 47 references, 3 tables, 3 figures, and 1 appendix.)…
Petersson, K M; Nichols, T E; Poline, J B; Holmes, A P
1999-01-01
Functional neuroimaging (FNI) provides experimental access to the intact living brain making it possible to study higher cognitive functions in humans. In this review and in a companion paper in this issue, we discuss some common methods used to analyse FNI data. The emphasis in both papers is on assumptions and limitations of the methods reviewed. There are several methods available to analyse FNI data indicating that none is optimal for all purposes. In order to make optimal use of the methods available it is important to know the limits of applicability. For the interpretation of FNI results it is also important to take into account the assumptions, approximations and inherent limitations of the methods used. This paper gives a brief overview over some non-inferential descriptive methods and common statistical models used in FNI. Issues relating to the complex problem of model selection are discussed. In general, proper model selection is a necessary prerequisite for the validity of the subsequent statistical inference. The non-inferential section describes methods that, combined with inspection of parameter estimates and other simple measures, can aid in the process of model selection and verification of assumptions. The section on statistical models covers approaches to global normalization and some aspects of univariate, multivariate, and Bayesian models. Finally, approaches to functional connectivity and effective connectivity are discussed. In the companion paper we review issues related to signal detection and statistical inference. PMID:10466149
Partonic orbital angular momentum
NASA Astrophysics Data System (ADS)
Arash, Firooz; Taghavi-Shahri, Fatemeh; Shahveh, Abolfazl
2013-04-01
Ji's decomposition of nucleon spin is used and the orbital angular momentum of quarks and gluon are calculated. We have utilized the so called valon model description of the nucleon in the next to leading order. It is found that the average orbital angular momentum of quarks is positive, but small, whereas that of gluon is negative and large. Individual quark flavor contributions are also calculated. Some regularities on the total angular momentum of the quarks and gluon are observed.
Statistical mechanics model of angiogenic tumor growth.
Ferreira, António Luis; Lipowska, Dorota; Lipowski, Adam
2012-01-01
We examine a lattice model of tumor growth where the survival of tumor cells depends on the supplied nutrients. When such a supply is random, the extinction of tumors belongs to the directed percolation universality class. However, when the supply is correlated with the distribution of tumor cells, which as we suggest might mimic the angiogenic growth, the extinction shows different critical behavior. Such a correlation affects also the morphology of the growing tumors and drastically raises tumor-survival probability. PMID:22400505
Modeling single-molecule detection statistics
NASA Astrophysics Data System (ADS)
Enderlein, Joerg; Robbins, David L.; Ambrose, W. P.; Goodwin, Peter M.; Keller, Richard A.
1997-05-01
We present experimental results of single B-phycoerythrin molecule detection in a fluid flow at different sample introduction rates. A new mathematical approach is used for calculating the resulting burst size distributions. The calculations are based upon a complete physical model including absorption, fluorescence and photobleaching characteristics of the fluorophore; its diffusion; the sample stream hydrodynamics; the spatially dependent optical detection efficiency; and the excitation laser beam characteristics. Special attention is paid to the phenomenon of `molecular noise'--fluctuations in the number of overlapping crossings of molecules through the detection volume. The importance of this study and its connections to experimental applications are discussed.
Fusion yield: Guderley model and Tsallis statistics
NASA Astrophysics Data System (ADS)
Haubold, H. J.; Kumar, D.
2011-02-01
The reaction rate probability integral is extended from Maxwell-Boltzmann approach to a more general approach by using the pathway model introduced by Mathai in 2005 (A pathway to matrix-variate gamma and normal densities. Linear Algebr. Appl. 396, 317-328). The extended thermonuclear reaction rate is obtained in the closed form via a Meijer's G-function and the so-obtained G-function is represented as a solution of a homogeneous linear differential equation. A physical model for the hydrodynamical process in a fusion plasma-compressed and laser-driven spherical shock wave is used for evaluating the fusion energy integral by integrating the extended thermonuclear reaction rate integral over the temperature. The result obtained is compared with the standard fusion yield obtained by Haubold and John in 1981 (Analytical representation of the thermonuclear reaction rate and fusion energy production in a spherical plasma shock wave. Plasma Phys. 23, 399-411). An interpretation for the pathway parameter is also given.
Statistical physics model of an evolving population
NASA Astrophysics Data System (ADS)
Sznajd-Weron, K.; Pȩkalski, A.
1999-12-01
There are many possible approaches by a theoretical physicist to problems of biological evolution. Some focus on physically interesting features, like the self-organized criticality (P. Bak, K. Sneppen, Phys. Rev. Lett 71 (1993); N. Vadewalle, M. Ausloos, Physica D 90 (1996) 262). Others put on more effort taking into account factors considered by biologists to be important in determining one or another aspect of biological evolution (D. Derrida, P.G. Higgs, J. Phys. A 24 (1991) L985; I. Mróz, A. Pȩkalski, K. Sznajd-Weron, Phys. Rev. Lett. 76 (1996) 3025; A. Pȩkalski, Physica A 265 (1999) 255). The intrinsic complexity of the problem enforces nevertheless drastic simplifications. Certain consolation may come from the fact that the mathematical models used by biologists themselves are quite often even more “coarse grained”.
A no extensive statistical model for the nucleon structure function
Trevisan, Luis A.; Mirez, Carlos
2013-03-25
We studied an application of nonextensive thermodynamics to describe the structure function of nucleon, in a model where the usual Fermi-Dirac and Bose-Einstein energy distribution were replaced by the equivalent functions of the q-statistical. The parameters of the model are given by an effective temperature T, the q parameter (from Tsallis statistics), and two chemical potentials given by the corresponding up (u) and down (d) quark normalization in the nucleon.
a Nonextensive Statistical Model for the Nucleon Structure Function
NASA Astrophysics Data System (ADS)
Trevisan, Luis Augusto; Mirez, Carlos
2013-07-01
We studied an application of nonextensive thermodynamics to describe the structure function of nucleon, in a model where the usual Fermi-Dirac and Bose-Einstein energy distribution were replaced by the equivalent functions of the q-statistical. The parameters of the model are given by an effective temperature T, the q parameter (from Tsallis statistics), and two chemical potentials given by the corresponding up (u) and down (d) quark normalizations in the nucleon.
A no extensive statistical model for the nucleon structure function
NASA Astrophysics Data System (ADS)
Trevisan, Luis A.; Mirez, Carlos
2013-03-01
We studied an application of nonextensive thermodynamics to describe the structure function of nucleon, in a model where the usual Fermi-Dirac and Bose-Einstein energy distribution were replaced by the equivalent functions of the q-statistical. The parameters of the model are given by an effective temperature T, the q parameter (from Tsallis statistics), and two chemical potentials given by the corresponding up (u) and down (d) quark normalization in the nucleon.
Statistical modelling of mitochondrial power supply.
James, A T; Wiskich, J T; Conyers, R A
1989-01-01
By experiment and theory, formulae are derived to calculate the response of mitochondrial power supply, in flux and potential, to an ATP consuming enzyme load, incorporating effects of varying amounts of (i) enzyme, (ii) total circulating adenylate, and (iii) inhibition of the ATP/ADP translocase. The formulae, which apply between about 20% and 80% of maximum respiration, are the same as for the current and voltage of an electrical circuit in which a battery with potential, linear in the logarithm of the total adenylate, charges another battery whose opposing potential is also linear in the same logarithm, through three resistances. These resistances produce loss of potential due to dis-equilibrium of (i) intramitochondrial oxidative phosphorylation, (ii) the ATP/ADP translocase, and (iii) the ATP-consuming enzyme load. The model is represented geometrically by the following configuration: when potential is plotted against flux, the points lie on two pencils of lines each concurrent at zero respiration, the two pencils describing the respective characteristics of the mitochondrion and enzyme. Control coefficients and elasticities are calculated from the formulae. PMID:2708917
Nuclear modifications of Parton Distribution Functions
NASA Astrophysics Data System (ADS)
Adeluyi, Adeola Adeleke
-called shadowing region. We also investigate the effects of nuclear modifications on observed quantities in ultrarelativistic nucleus-nucleus collisions. Specifically, we consider deuteron-gold collisions and observables which are directly impacted by modifications, such as pseudorapidity asymmetry and nuclear modification factors. A good description of the shadowing region is afforded by Gribov Theory. Gribov related the shadowing correction to the differential diffractive hadron-nucleon cross section. We generalize Gribov theory to include both the real part of the diffractive scattering amplitude and higher order multiple scattering necessary for heavy nuclei. The diffractive dissociation inputs are taken from experiments. We calculate observables in deuteron-gold collisions. Utilizing the factorization theorem, we use the existing parameterizations of nuclear PDFs and fragmentation functions in a pQCD-improved parton model to calculate nuclear modification factors and pseudorapidity asymmetries. The nuclear modification factor is essentially the ratio of the deuteron-gold cross section to that of the proton-proton cross section scaled by the number of binary collisions. The pseudorapidity asymmetry is the ratio of the cross section in the negative rapidity region relative to that in the equivalent positive rapidity region. Both quantities are sensitive to the effects of nuclear modifications on PDFs. Results are compared to experimental data from the BRAHMS and STAR collaborations.
ERIC Educational Resources Information Center
Cziko, Gary A.
The development of an empirically based model of communicative competence is discussed in terms of conceptual, statistical, and methodological considerations. A distinction is made between descriptive and working models of communicative competence. Working models attempt to show how components of communicative competence are interrelated…
Relaxation in statistical many-agent economy models
NASA Astrophysics Data System (ADS)
Patriarca, M.; Chakraborti, A.; Heinsalu, E.; Germano, G.
2007-05-01
We review some statistical many-agent models of economic and social systems inspired by microscopic molecular models and discuss their stochastic interpretation. We apply these models to wealth exchange in economics and study how the relaxation process depends on the parameters of the system, in particular on the saving propensities that define and diversify the agent profiles.
Statistical models and NMR analysis of polymer microstructure
Technology Transfer Automated Retrieval System (TEKTRAN)
Statistical models can be used in conjunction with NMR spectroscopy to study polymer microstructure and polymerization mechanisms. Thus, Bernoullian, Markovian, and enantiomorphic-site models are well known. Many additional models have been formulated over the years for additional situations. Typica...
Statistical inference for exploratory data analysis and model diagnostics.
Buja, Andreas; Cook, Dianne; Hofmann, Heike; Lawrence, Michael; Lee, Eun-Kyung; Swayne, Deborah F; Wickham, Hadley
2009-11-13
We propose to furnish visual statistical methods with an inferential framework and protocol, modelled on confirmatory statistical testing. In this framework, plots take on the role of test statistics, and human cognition the role of statistical tests. Statistical significance of 'discoveries' is measured by having the human viewer compare the plot of the real dataset with collections of plots of simulated datasets. A simple but rigorous protocol that provides inferential validity is modelled after the 'lineup' popular from criminal legal procedures. Another protocol modelled after the 'Rorschach' inkblot test, well known from (pop-)psychology, will help analysts acclimatize to random variability before being exposed to the plot of the real data. The proposed protocols will be useful for exploratory data analysis, with reference datasets simulated by using a null assumption that structure is absent. The framework is also useful for model diagnostics in which case reference datasets are simulated from the model in question. This latter point follows up on previous proposals. Adopting the protocols will mean an adjustment in working procedures for data analysts, adding more rigour, and teachers might find that incorporating these protocols into the curriculum improves their students' statistical thinking. PMID:19805449
Statistical error model for a solar electric propulsion thrust subsystem
NASA Technical Reports Server (NTRS)
Bantell, M. H.
1973-01-01
The solar electric propulsion thrust subsystem statistical error model was developed as a tool for investigating the effects of thrust subsystem parameter uncertainties on navigation accuracy. The model is currently being used to evaluate the impact of electric engine parameter uncertainties on navigation system performance for a baseline mission to Encke's Comet in the 1980s. The data given represent the next generation in statistical error modeling for low-thrust applications. Principal improvements include the representation of thrust uncertainties and random process modeling in terms of random parametric variations in the thrust vector process for a multi-engine configuration.
The parton distribution function library
Plothow-Besch, H.
1995-07-01
This article describes an integrated package of Parton Density Functions called PDFLIB which has been added to the CERN Program Library Pool W999 and is labelled as W5051. In this package all the different sets of parton density functions of the Nucleon, Pion and the Photon which are available today have been put together. All these sets have been combined in a consistent way such that they all have similar calling sequences and no external data files have to be read in anymore. A default set has been prepared, although those preferring their own set or wanting to test a new one may do so within the package. The package also offers a program to calculate the strong coupling constant {alpha}, to first or second order. The correct {Lambda}{sub QCD} associated to the selected set of structure functions and the number of allowed flavours with respect to the given Q{sup 2} is automatically used in the calculation. The selection of sets, the program parameters as well as the possibilities to modify the defaults and to control errors occurred during execution are described.
Jet correlations from unintegrated parton distributions
Hautmann, F.; Jung, H.
2008-10-13
Transverse-momentum dependent parton distributions can be introduced gauge-invariantly in QCD from high-energy factorization. We discuss Monte Carlo applications of these distributions to parton showers and jet physics, with a view to the implications for the Monte Carlo description of complex hadronic final states with multiple hard scales at the LHC.
Introduction to Parton-Shower Event Generators
NASA Astrophysics Data System (ADS)
Höche, Stefan
This lecture discusses the physics implemented by Monte Carlo event generators for hadron colliders. It details the construction of parton showers and the matching of parton showers to fixed-order calculations at higher orders in perturbative QCD. It also discusses approaches to merge calculations for a varying number of jets, the interface to the underlying event and hadronization.
A Stochastic Fractional Dynamics Model of Rainfall Statistics
NASA Astrophysics Data System (ADS)
Kundu, Prasun; Travis, James
2013-04-01
Rainfall varies in space and time in a highly irregular manner and is described naturally in terms of a stochastic process. A characteristic feature of rainfall statistics is that they depend strongly on the space-time scales over which rain data are averaged. A spectral model of precipitation has been developed based on a stochastic differential equation of fractional order for the point rain rate, that allows a concise description of the second moment statistics of rain at any prescribed space-time averaging scale. The model is designed to faithfully reflect the scale dependence and is thus capable of providing a unified description of the statistics of both radar and rain gauge data. The underlying dynamical equation can be expressed in terms of space-time derivatives of fractional orders that are adjusted together with other model parameters to fit the data. The form of the resulting spectrum gives the model adequate flexibility to capture the subtle interplay between the spatial and temporal scales of variability of rain but strongly constrains the predicted statistical behavior as a function of the averaging length and times scales. The main restriction is the assumption that the statistics of the precipitation field is spatially homogeneous and isotropic and stationary in time. We test the model with radar and gauge data collected contemporaneously at the NASA TRMM ground validation sites located near Melbourne, Florida and in Kwajalein Atoll, Marshall Islands in the tropical Pacific. We estimate the parameters by tuning them to the second moment statistics of the radar data. The model predictions are then found to fit the second moment statistics of the gauge data reasonably well without any further adjustment. Some data sets containing periods of non-stationary behavior that involves occasional anomalously correlated rain events, present a challenge for the model.
Illuminating the 1/x Moment of Parton Distribution Functions
Brodsky, Stanley J.; Llanes-Estrada, Felipe J.; Szczepaniak, Adam P.; /Indiana U.
2007-10-15
The Weisberger relation, an exact statement of the parton model, elegantly relates a high-energy physics observable, the 1/x moment of parton distribution functions, to a nonperturbative low-energy observable: the dependence of the nucleon mass on the value of the quark mass or its corresponding quark condensate. We show that contemporary fits to nucleon structure functions fail to determine this 1/x moment; however, deeply virtual Compton scattering can be described in terms of a novel F1/x(t) form factor which illuminates this physics. An analysis of exclusive photon-induced processes in terms of the parton-nucleon scattering amplitude with Regge behavior reveals a failure of the high Q2 factorization of exclusive processes at low t in terms of the Generalized Parton-Distribution Functions which has been widely believed to hold in the past. We emphasize the need for more data for the DVCS process at large t in future or upgraded facilities.
Neutrino Production of Mesons in the Generalized Parton Picture
NASA Astrophysics Data System (ADS)
McAskill, Tracy
The handbag model and its usefulness in generating cross sections for light pseudoscalar mesons is investigated here. The soft part of the handbag model is first parametrized to fit well-known models of generalized parton distributions (GPD's), then cross sections are calculated directly from the GPD's. This is then directly extended to the calculation of neutrino cross sections for the production of the same type of light mesons.
Medium Modifications of Hadron Properties and Partonic Processes
Brooks, W. K.; Strauch, S.; Tsushima, K.
2011-06-01
Chiral symmetry is one of the most fundamental symmetries in QCD. It is closely connected to hadron properties in the nuclear medium via the reduction of the quark condensate
Statistical analysis of modeling error in structural dynamic systems
NASA Technical Reports Server (NTRS)
Hasselman, T. K.; Chrostowski, J. D.
1990-01-01
The paper presents a generic statistical model of the (total) modeling error for conventional space structures in their launch configuration. Modeling error is defined as the difference between analytical prediction and experimental measurement. It is represented by the differences between predicted and measured real eigenvalues and eigenvectors. Comparisons are made between pre-test and post-test models. Total modeling error is then subdivided into measurement error, experimental error and 'pure' modeling error, and comparisons made between measurement error and total modeling error. The generic statistical model presented in this paper is based on the first four global (primary structure) modes of four different structures belonging to the generic category of Conventional Space Structures (specifically excluding large truss-type space structures). As such, it may be used to evaluate the uncertainty of predicted mode shapes and frequencies, sinusoidal response, or the transient response of other structures belonging to the same generic category.
Parton and valon distributions in the nucleon
Hwa, R.C.; Sajjad Zahir, M.
1981-06-01
Structure functions of the nucleon are analyzed in the valon model in which a nucleon is assumed to be a bound state of three valence quark clusters (valons). At high Q/sup 2/ the structure of the valons is described by leading-order results in the perturbative quantum chromodynamics. From the experimental data on deep-inelastic scattering off protons and neutrons, the flavor-dependent valon distributions in the nucleon are determined. Predictions for the parton distributions are then made for high Q/sup 2/ without guesses concerning the quark and gluon distributions at low Q/sup 2/. The sea-quark and gluon distributions are found to have a sharp peak at very small x. Convenient parametrization is provided which interpolates between different numbers of flavors.
Adapting internal statistical models for interpreting visual cues to depth
Seydell, Anna; Knill, David C.; Trommershäuser, Julia
2010-01-01
The informativeness of sensory cues depends critically on statistical regularities in the environment. However, statistical regularities vary between different object categories and environments. We asked whether and how the brain changes the prior assumptions about scene statistics used to interpret visual depth cues when stimulus statistics change. Subjects judged the slants of stereoscopically presented figures by adjusting a virtual probe perpendicular to the surface. In addition to stereoscopic disparities, the aspect ratio of the stimulus in the image provided a “figural compression” cue to slant, whose reliability depends on the distribution of aspect ratios in the world. As we manipulated this distribution from regular to random and back again, subjects’ reliance on the compression cue relative to stereoscopic cues changed accordingly. When we randomly interleaved stimuli from shape categories (ellipses and diamonds) with different statistics, subjects gave less weight to the compression cue for figures from the category with more random aspect ratios. Our results demonstrate that relative cue weights vary rapidly as a function of recently experienced stimulus statistics, and that the brain can use different statistical models for different object categories. We show that subjects’ behavior is consistent with that of a broad class of Bayesian learning models. PMID:20465321
Right-sizing statistical models for longitudinal data.
Wood, Phillip K; Steinley, Douglas; Jackson, Kristina M
2015-12-01
Arguments are proposed that researchers using longitudinal data should consider more and less complex statistical model alternatives to their initially chosen techniques in an effort to "right-size" the model to the data at hand. Such model comparisons may alert researchers who use poorly fitting, overly parsimonious models to more complex, better-fitting alternatives and, alternatively, may identify more parsimonious alternatives to overly complex (and perhaps empirically underidentified and/or less powerful) statistical models. A general framework is proposed for considering (often nested) relationships between a variety of psychometric and growth curve models. A 3-step approach is proposed in which models are evaluated based on the number and patterning of variance components prior to selection of better-fitting growth models that explain both mean and variation-covariation patterns. The orthogonal free curve slope intercept (FCSI) growth model is considered a general model that includes, as special cases, many models, including the factor mean (FM) model (McArdle & Epstein, 1987), McDonald's (1967) linearly constrained factor model, hierarchical linear models (HLMs), repeated-measures multivariate analysis of variance (MANOVA), and the linear slope intercept (linearSI) growth model. The FCSI model, in turn, is nested within the Tuckerized factor model. The approach is illustrated by comparing alternative models in a longitudinal study of children's vocabulary and by comparing several candidate parametric growth and chronometric models in a Monte Carlo study. PMID:26237507
Simple Statistical Model for Branched Aggregates: Application to Cooee Bitumen.
Lemarchand, Claire A; Hansen, Jesper S
2015-11-01
We propose a statistical model that can reproduce the size distribution of any branched aggregate, including amylopectin, dendrimers, molecular clusters of monoalcohols, and asphaltene nanoaggregates. It is based on the conditional probability for one molecule to form a new bond with a molecule, given that it already has bonds with others. The model is applied here to asphaltene nanoaggregates observed in molecular dynamics simulations of Cooee bitumen. The variation with temperature of the probabilities deduced from this model is discussed in terms of statistical mechanics arguments. The relevance of the statistical model in the case of asphaltene nanoaggregates is checked by comparing the predicted value of the probability for one molecule to have exactly i bonds with the same probability directly measured in the molecular dynamics simulations. The agreement is satisfactory. PMID:26458140
DEVELOPMENT OF A STATISTICAL MODEL FOR METAL-HUMIC INTERACTIONS
A statistical model for describing the distribution of binding sites in humic substances was developed. he model was applied to study the spectral titration plot generated by the lanthanide ion probe spectroscopy (LIPS) technique. his titration plot is used as a basis for studyin...
Statistical Design Model (SDM) of satellite thermal control subsystem
NASA Astrophysics Data System (ADS)
Mirshams, Mehran; Zabihian, Ehsan; Aarabi Chamalishahi, Mahdi
2016-07-01
Satellites thermal control, is a satellite subsystem that its main task is keeping the satellite components at its own survival and activity temperatures. Ability of satellite thermal control plays a key role in satisfying satellite's operational requirements and designing this subsystem is a part of satellite design. In the other hand due to the lack of information provided by companies and designers still doesn't have a specific design process while it is one of the fundamental subsystems. The aim of this paper, is to identify and extract statistical design models of spacecraft thermal control subsystem by using SDM design method. This method analyses statistical data with a particular procedure. To implement SDM method, a complete database is required. Therefore, we first collect spacecraft data and create a database, and then we extract statistical graphs using Microsoft Excel, from which we further extract mathematical models. Inputs parameters of the method are mass, mission, and life time of the satellite. For this purpose at first thermal control subsystem has been introduced and hardware using in the this subsystem and its variants has been investigated. In the next part different statistical models has been mentioned and a brief compare will be between them. Finally, this paper particular statistical model is extracted from collected statistical data. Process of testing the accuracy and verifying the method use a case study. Which by the comparisons between the specifications of thermal control subsystem of a fabricated satellite and the analyses results, the methodology in this paper was proved to be effective. Key Words: Thermal control subsystem design, Statistical design model (SDM), Satellite conceptual design, Thermal hardware
Computationally efficient statistical differential equation modeling using homogenization
Hooten, Mevin B.; Garlick, Martha J.; Powell, James A.
2013-01-01
Statistical models using partial differential equations (PDEs) to describe dynamically evolving natural systems are appearing in the scientific literature with some regularity in recent years. Often such studies seek to characterize the dynamics of temporal or spatio-temporal phenomena such as invasive species, consumer-resource interactions, community evolution, and resource selection. Specifically, in the spatial setting, data are often available at varying spatial and temporal scales. Additionally, the necessary numerical integration of a PDE may be computationally infeasible over the spatial support of interest. We present an approach to impose computationally advantageous changes of support in statistical implementations of PDE models and demonstrate its utility through simulation using a form of PDE known as “ecological diffusion.” We also apply a statistical ecological diffusion model to a data set involving the spread of mountain pine beetle (Dendroctonus ponderosae) in Idaho, USA.
Statistical Modeling of Large-Scale Simulation Data
Eliassi-Rad, T; Critchlow, T; Abdulla, G
2002-02-22
With the advent of fast computer systems, Scientists are now able to generate terabytes of simulation data. Unfortunately, the shear size of these data sets has made efficient exploration of them impossible. To aid scientists in gathering knowledge from their simulation data, we have developed an ad-hoc query infrastructure. Our system, called AQSim (short for Ad-hoc Queries for Simulation) reduces the data storage requirements and access times in two stages. First, it creates and stores mathematical and statistical models of the data. Second, it evaluates queries on the models of the data instead of on the entire data set. In this paper, we present two simple but highly effective statistical modeling techniques for simulation data. Our first modeling technique computes the true mean of systematic partitions of the data. It makes no assumptions about the distribution of the data and uses a variant of the root mean square error to evaluate a model. In our second statistical modeling technique, we use the Andersen-Darling goodness-of-fit method on systematic partitions of the data. This second method evaluates a model by how well it passes the normality test on the data. Both of our statistical models summarize the data so as to answer range queries in the most effective way. We calculate precision on an answer to a query by scaling the one-sided Chebyshev Inequalities with the original mesh's topology. Our experimental evaluations on two scientific simulation data sets illustrate the value of using these statistical modeling techniques on large simulation data sets.
Modeling Ka-band low elevation angle propagation statistics
NASA Technical Reports Server (NTRS)
Russell, Thomas A.; Weinfield, John; Pearson, Chris; Ippolito, Louis J.
1995-01-01
The statistical variability of the secondary atmospheric propagation effects on satellite communications cannot be ignored at frequencies of 20 GHz or higher, particularly if the propagation margin allocation is such that link availability falls below 99 percent. The secondary effects considered in this paper are gaseous absorption, cloud absorption, and tropospheric scintillation; rain attenuation is the primary effect. Techniques and example results are presented for estimation of the overall combined impact of the atmosphere on satellite communications reliability. Statistical methods are employed throughout and the most widely accepted models for the individual effects are used wherever possible. The degree of correlation between the effects is addressed and some bounds on the expected variability in the combined effects statistics are derived from the expected variability in correlation. Example estimates are presented of combined effects statistics in the Washington D.C. area of 20 GHz and 5 deg elevation angle. The statistics of water vapor are shown to be sufficient for estimation of the statistics of gaseous absorption at 20 GHz. A computer model based on monthly surface weather is described and tested. Significant improvement in prediction of absorption extremes is demonstrated with the use of path weather data instead of surface data.
The beta distribution: A statistical model for world cloud cover
NASA Technical Reports Server (NTRS)
Falls, L. W.
1973-01-01
Much work has been performed in developing empirical global cloud cover models. This investigation was made to determine an underlying theoretical statistical distribution to represent worldwide cloud cover. The beta distribution with probability density function is given to represent the variability of this random variable. It is shown that the beta distribution possesses the versatile statistical characteristics necessary to assume the wide variety of shapes exhibited by cloud cover. A total of 160 representative empirical cloud cover distributions were investigated and the conclusion was reached that this study provides sufficient statical evidence to accept the beta probability distribution as the underlying model for world cloud cover.
Quantitative statistical assessment of conditional models for synthetic aperture radar.
DeVore, Michael D; O'Sullivan, Joseph A
2004-02-01
Many applications of object recognition in the presence of pose uncertainty rely on statistical models-conditioned on pose-for observations. The image statistics of three-dimensional (3-D) objects are often assumed to belong to a family of distributions with unknown model parameters that vary with one or more continuous-valued pose parameters. Many methods for statistical model assessment, for example the tests of Kolmogorov-Smirnov and K. Pearson, require that all model parameters be fully specified or that sample sizes be large. Assessing pose-dependent models from a finite number of observations over a variety of poses can violate these requirements. However, a large number of small samples, corresponding to unique combinations of object, pose, and pixel location, are often available. We develop methods for model testing which assume a large number of small samples and apply them to the comparison of three models for synthetic aperture radar images of 3-D objects with varying pose. Each model is directly related to the Gaussian distribution and is assessed both in terms of goodness-of-fit and underlying model assumptions, such as independence, known mean, and homoscedasticity. Test results are presented in terms of the functional relationship between a given significance level and the percentage of samples that wold fail a test at that level. PMID:15376934
Disentangling correlations in multiple parton interactions
Calucci, G.; Treleani, D.
2011-01-01
Multiple Parton Interactions are the tool to obtain information on the correlations between partons in the hadron structure. Partons may be correlated in all degrees of freedom and all different correlation terms contribute to the cross section. The contributions due to the different parton flavors can be isolated, at least to some extent, by selecting properly the final state. In the case of high energy proton-proton collisions, the effects of correlations in the transverse coordinates and in fractional momenta are, on the contrary, unavoidably mixed in the final observables. The standard way to quantify the strength of double parton interactions is by the value of the effective cross section and a small value of the effective cross section may be originated both by the relatively short transverse distance between the pairs of partons undergoing the double interaction and by a large dispersion of the distribution in multiplicity of the multiparton distributions. The aim of the present paper is to show how the effects of longitudinal and transverse correlations may be disentangled by taking into account the additional information provided by double parton interactions in high energy proton-deuteron collisions.
Working Group I: Parton distributions: Summary report for the HERA LHC Workshop Proceedings
Dittmar, M.; Forte, S.; Glazov, A.; Moch, S.; Alekhin, S.; Altarelli, G.; Andersen, Jeppe R.; Ball, R.D.; Blumlein, J.; Bottcher, H.; Carli, T.; Ciafaloni, M.; Colferai, D.; Cooper-Sarkar, A.; Corcella, G.; Del Debbio, L.; Dissertori, G.; Feltesse, J.; Guffanti, A.; Gwenlan, C.; Huston, J.; /Zurich, ETH /DESY, Zeuthen /Serpukhov, IHEP /CERN /Rome III U. /INFN, Rome3 /Cambridge U. /Edinburgh U. /Florence U. /INFN, Florence /Oxford U. /DSM, DAPNIA, Saclay /Michigan State U. /Uppsala U. /Barcelona U., ECM /Podgorica U. /Turin U. /INFN, Turin /Harish-Chandra Res. Inst. /Fermilab /Hamburg U., Inst. Theor. Phys. II
2005-11-01
We provide an assessment of the impact of parton distributions on the determination of LHC processes, and of the accuracy with which parton distributions (PDFs) can be extracted from data, in particular from current and forthcoming HERA experiments. We give an overview of reference LHC processes and their associated PDF uncertainties, and study in detail W and Z production at the LHC.We discuss the precision which may be obtained from the analysis of existing HERA data, tests of consistency of HERA data from different experiments, and the combination of these data. We determine further improvements on PDFs which may be obtained from future HERA data (including measurements of F{sub L}), and from combining present and future HERA data with present and future hadron collider data. We review the current status of knowledge of higher (NNLO) QCD corrections to perturbative evolution and deep-inelastic scattering, and provide reference results for their impact on parton evolution, and we briefly examine non-perturbative models for parton distributions. We discuss the state-of-the art in global parton fits, we assess the impact on them of various kinds of data and of theoretical corrections, by providing benchmarks of Alekhin and MRST parton distributions and a CTEQ analysis of parton fit stability, and we briefly presents proposals for alternative approaches to parton fitting. We summarize the status of large and small x resummation, by providing estimates of the impact of large x resummation on parton fits, and a comparison of different approaches to small x resummation, for which we also discuss numerical techniques.
Propagating uncertainties in statistical model based shape prediction
NASA Astrophysics Data System (ADS)
Syrkina, Ekaterina; Blanc, Rémi; Székely, Gàbor
2011-03-01
This paper addresses the question of accuracy assessment and confidence regions estimation in statistical model based shape prediction. Shape prediction consists in estimating the shape of an organ based on a partial observation, due e.g. to a limited field of view or poorly contrasted images, and generally requires a statistical model. However, such predictions can be impaired by several sources of uncertainty, in particular the presence of noise in the observation, limited correlations between the predictors and the shape to predict, as well as limitations of the statistical shape model - in particular the number of training samples. We propose a framework which takes these into account and derives confidence regions around the predicted shape. Our method relies on the construction of two separate statistical shape models, for the predictors and for the unseen parts, and exploits the correlations between them assuming a joint Gaussian distribution. Limitations of the models are taken into account by jointly optimizing the prediction and minimizing the shape reconstruction error through cross-validation. An application to the prediction of the shape of the proximal part of the human tibia given the shape of the distal femur is proposed, as well as the evaluation of the reliability of the estimated confidence regions, using a database of 184 samples. Potential applications are reconstructive surgery, e.g. to assess whether an implant fits in a range of acceptable shapes, or functional neurosurgery when the target's position is not directly visible and needs to be inferred from nearby visible structures.
Statistical approaches to pharmacodynamic modeling: motivations, methods, and misperceptions.
Mick, R; Ratain, M J
1993-01-01
We have attempted to outline the fundamental statistical aspects of pharmacodynamic modeling. Unexpected yet substantial variability in effect in a group of similarly treated patients is the key motivation for pharmacodynamic investigations. Pharmacokinetic and/or pharmacodynamic factors may influence this variability. Residual variability in effect that persists after accounting for drug exposure indicates that further statistical modeling with pharmacodynamic factors is warranted. Factors that significantly predict interpatient variability in effect may then be employed to individualize the drug dose. In this paper we have emphasized the need to understand the properties of the effect measure and explanatory variables in terms of scale, distribution, and statistical relationship. The assumptions that underlie many types of statistical models have been discussed. The role of residual analysis has been stressed as a useful method to verify assumptions. We have described transformations and alternative regression methods that are employed when these assumptions are found to be in violation. Sequential selection procedures for the construction of multivariate models have been presented. The importance of assessing model performance has been underscored, most notably in terms of bias and precision. In summary, pharmacodynamic analyses are now commonly performed and reported in the oncologic literature. The content and format of these analyses has been variable. The goals of such analyses are to identify and describe pharmacodynamic relationships and, in many cases, to propose a statistical model. However, the appropriateness and performance of the proposed model are often difficult to judge. Table 1 displays suggestions (in a checklist format) for structuring the presentation of pharmacodynamic analyses, which reflect the topics reviewed in this paper. PMID:8269582
Multiple parton interaction studies at DØ
Lincoln, D.
2016-04-01
Here, we present the results of studies of multiparton interactions done by the DØ collaboration using the Fermilab Tevatron at a center of mass energy of 1.96 TeV. We also present three analyses, involving three distinct final signatures: (a) a photon with at least 3 jets ( γ + 3jets), (b) a photon with a bottom or charm quark tagged jet and at least 2 other jets ( γ + b/c + 2jets), and (c) two J/ ψ mesons. The fraction of photon + jet events initiated by double parton scattering is about 20%, while the fraction for events inmore » which two J/ ψ mesons were produced is 30 ± 10. While the two measurements are statistically compatible, the difference might indicate differences in the quark and gluon distribution within a nucleon. Finally, this speculation originates from the fact that photon + jet events are created by collisions with quarks in the initial states, while J/ ψ events are produced preferentially by a gluonic initial state.« less
Predicting lettuce canopy photosynthesis with statistical and neural network models.
Frick, J; Precetti, C; Mitchell, C A
1998-11-01
An artificial neural network (NN) and a statistical regression model were developed to predict canopy photosynthetic rates (Pn) for 'Waldman's Green' leaf lettuce (Latuca sativa L.). All data used to develop and test the models were collected for crop stands grown hydroponically and under controlled-environment conditions. In the NN and regression models, canopy Pn was predicted as a function of three independent variables: shootzone CO2 concentration (600 to 1500 micromoles mol-1), photosynthetic photon flux (PPF) (600 to 1100 micromoles m-2 s-1), and canopy age (10 to 20 days after planting). The models were used to determine the combinations of CO2 and PPF setpoints required each day to maintain maximum canopy Pn. The statistical model (a third-order polynomial) predicted Pn more accurately than the simple NN (a three-layer, fully connected net). Over an 11-day validation period, average percent difference between predicted and actual Pn was 12.3% and 24.6% for the statistical and NN models, respectively. Both models lost considerable accuracy when used to determine relatively long-range Pn predictions (> or = 6 days into the future). PMID:11542672
STATISTICAL BASED NON-LINEAR MODEL UPDATING USING FEATURE EXTRACTION
Schultz, J.F.; Hemez, F.M.
2000-10-01
This research presents a new method to improve analytical model fidelity for non-linear systems. The approach investigates several mechanisms to assist the analyst in updating an analytical model based on experimental data and statistical analysis of parameter effects. The first is a new approach at data reduction called feature extraction. This is an expansion of the update metrics to include specific phenomena or character of the response that is critical to model application. This is an extension of the classical linear updating paradigm of utilizing the eigen-parameters or FRFs to include such devices as peak acceleration, time of arrival or standard deviation of model error. The next expansion of the updating process is the inclusion of statistical based parameter analysis to quantify the effects of uncertain or significant effect parameters in the construction of a meta-model. This provides indicators of the statistical variation associated with parameters as well as confidence intervals on the coefficients of the resulting meta-model, Also included in this method is the investigation of linear parameter effect screening using a partial factorial variable array for simulation. This is intended to aid the analyst in eliminating from the investigation the parameters that do not have a significant variation effect on the feature metric, Finally an investigation of the model to replicate the measured response variation is examined.
Predicting lettuce canopy photosynthesis with statistical and neural network models
NASA Technical Reports Server (NTRS)
Frick, J.; Precetti, C.; Mitchell, C. A.
1998-01-01
An artificial neural network (NN) and a statistical regression model were developed to predict canopy photosynthetic rates (Pn) for 'Waldman's Green' leaf lettuce (Latuca sativa L.). All data used to develop and test the models were collected for crop stands grown hydroponically and under controlled-environment conditions. In the NN and regression models, canopy Pn was predicted as a function of three independent variables: shootzone CO2 concentration (600 to 1500 micromoles mol-1), photosynthetic photon flux (PPF) (600 to 1100 micromoles m-2 s-1), and canopy age (10 to 20 days after planting). The models were used to determine the combinations of CO2 and PPF setpoints required each day to maintain maximum canopy Pn. The statistical model (a third-order polynomial) predicted Pn more accurately than the simple NN (a three-layer, fully connected net). Over an 11-day validation period, average percent difference between predicted and actual Pn was 12.3% and 24.6% for the statistical and NN models, respectively. Both models lost considerable accuracy when used to determine relatively long-range Pn predictions (> or = 6 days into the future).
Statistical Properties of Downscaled CMIP3 Global Climate Model Simulations
NASA Astrophysics Data System (ADS)
Duffy, P.; Tyan, S.; Thrasher, B.; Maurer, E. P.; Tebaldi, C.
2009-12-01
Spatial downscaling of global climate model projections adds physically meaningful spatial detail, and brings the results down to a scale that is more relevant to human and ecological systems. Statistical/empirical downscaling methods are computationally inexpensive, and thus can be applied to large ensembles of global climate model projections. Here we examine some of the statistical properties of a large ensemble of empirically downscale global climate projections. The projections are the CMIP3 global climate model projections that were performed by modeling groups around the world and archived by the Program for Climate Model Diagnosis and Intercomparison at Lawrence Livermore National Laboratory. Downscaled versions of 112 of these simulations were created on 2007 and are archived at http://gdo-dcp.ucllnl.org/downscaled_cmip3_projections/dcpInterface.html. The downscaling methodology employed, “Bias Correction/Spatial Downscaling” (BCSD), includes a correction of GCM biases relative to observations during a historical reference period, as well as empirical downscaling to grid scale of ~12 km. We analyzed these downscaled projections and some of the original global model results to assess effects of the bias correction and downscaling on the statistical properties of the ensemble. We also assessed uncertainty in the climate response to increased greenhouse gases from initial conditions relative to the uncertainty introduced by choice of global climate model.
Simple classical model for Fano statistics in radiation detectors
NASA Astrophysics Data System (ADS)
Jordan, David V.; Renholds, Andrea S.; Jaffe, John E.; Anderson, Kevin K.; René Corrales, L.; Peurrung, Anthony J.
2008-02-01
A simple classical model that captures the essential statistics of energy partitioning processes involved in the creation of information carriers (ICs) in radiation detectors is presented. The model pictures IC formation from a fixed amount of deposited energy in terms of the statistically analogous process of successively sampling water from a large, finite-volume container ("bathtub") with a small dipping implement ("shot or whiskey glass"). The model exhibits sub-Poisson variance in the distribution of the number of ICs generated (the "Fano effect"). Elementary statistical analysis of the model clarifies the role of energy conservation in producing the Fano effect and yields Fano's prescription for computing the relative variance of the IC number distribution in terms of the mean and variance of the underlying, single-IC energy distribution. The partitioning model is applied to the development of the impact ionization cascade in semiconductor radiation detectors. It is shown that, in tandem with simple assumptions regarding the distribution of energies required to create an (electron, hole) pair, the model yields an energy-independent Fano factor of 0.083, in accord with the lower end of the range of literature values reported for silicon and high-purity germanium. The utility of this simple picture as a diagnostic tool for guiding or constraining more detailed, "microscopic" physical models of detector material response to ionizing radiation is discussed.
Estimating Preferential Flow in Karstic Aquifers Using Statistical Mixed Models
Anaya, Angel A.; Padilla, Ingrid; Macchiavelli, Raul; Vesper, Dorothy J.; Meeker, John D.; Alshawabkeh, Akram N.
2013-01-01
Karst aquifers are highly productive groundwater systems often associated with conduit flow. These systems can be highly vulnerable to contamination, resulting in a high potential for contaminant exposure to humans and ecosystems. This work develops statistical models to spatially characterize flow and transport patterns in karstified limestone and determines the effect of aquifer flow rates on these patterns. A laboratory-scale Geo-HydroBed model is used to simulate flow and transport processes in a karstic limestone unit. The model consists of stainless-steel tanks containing a karstified limestone block collected from a karst aquifer formation in northern Puerto Rico. Experimental work involves making a series of flow and tracer injections, while monitoring hydraulic and tracer response spatially and temporally. Statistical mixed models are applied to hydraulic data to determine likely pathways of preferential flow in the limestone units. The models indicate a highly heterogeneous system with dominant, flow-dependent preferential flow regions. Results indicate that regions of preferential flow tend to expand at higher groundwater flow rates, suggesting a greater volume of the system being flushed by flowing water at higher rates. Spatial and temporal distribution of tracer concentrations indicates the presence of conduit-like and diffuse flow transport in the system, supporting the notion of both combined transport mechanisms in the limestone unit. The temporal response of tracer concentrations at different locations in the model coincide with, and confirms the preferential flow distribution generated with the statistical mixed models used in the study. PMID:23802921
DATA MANAGEMENT, STATISTICS AND COMMUNITY IMPACT MODELING CORE
EPA GRANT NUMBER: R832141C007
Title: Data Management, Statistics and Community Impact Modeling Core
Investigator: Frederica P Perera
Institution: Columbia University
EPA Project Officer: Nigel Fields
Project Period: No...
A very simple statistical model to the quarks asymmetries
NASA Astrophysics Data System (ADS)
Trevisan, Luis Augusto; Mirez, Carlos
2016-04-01
A simple statistical model is developed with the Fock states being the meson-hadron fluctuations. As expected, a insight about the violation of the Gottfried sum rule is obtained, and also a small difference between the strangeness amount in proton and neutron is explained.
Reply to Discussants: Statistical Models for Behavioral Observations.
ERIC Educational Resources Information Center
Rogosa, David; Ghandour, Ghassan
1991-01-01
Issues raised with the statistical models developed are discussed point by point, restating the emphasis on finite observation time, and reiterating the criticism of traditional psychometric methods. It is noted that the language and technical formulation of psychometrics can be extremely awkward in dealing with biased estimates. (SLD)
Environmental Concern and Sociodemographic Variables: A Study of Statistical Models
ERIC Educational Resources Information Center
Xiao, Chenyang; McCright, Aaron M.
2007-01-01
Studies of the social bases of environmental concern over the past 30 years have produced somewhat inconsistent results regarding the effects of sociodemographic variables, such as gender, income, and place of residence. The authors argue that model specification errors resulting from violation of two statistical assumptions (interval-level…
Applying the luminosity function statistics in the fireshell model
NASA Astrophysics Data System (ADS)
Rangel Lemos, L. J.; Bianco, C. L.; Ruffini, R.
2015-12-01
The luminosity function (LF) statistics applied to the data of BATSE, GBM/Fermi and BAT/Swift is the theme approached in this work. The LF is a strong statistical tool to extract useful information from astrophysical samples, and the key point of this statistical analysis is in the detector sensitivity, where we have performed careful analysis. We applied the tool of the LF statistics to three GRB classes predicted by the Fireshell model. We produced, by LF statistics, predicted distributions of: peak ux N(Fph pk), redshift N(z) and peak luminosity N(Lpk) for the three GRB classes predicted by Fireshell model; we also used three GRB rates. We looked for differences among the distributions, and in fact we found. We performed a comparison between the distributions predicted and observed (with and without redshifts), where we had to build a list with 217 GRBs with known redshifts. Our goal is transform the GRBs in a standard candle, where a alternative is find a correlation between the isotropic luminosity and the Band peak spectral energy (Liso - Epk).
A review of the kinetic statistical strength model
Attia, A.V.
1996-03-11
This is a review of the Kinetic-Statistical Strength (KSS) model described in the report ``Models of Material Strength, Fracture and Failure`` by V. Kuropatenko and V. Bychenkov. The models for metals subjected to high strain rates (explosions) are focussed on. Model implementation appears possible in a hydrocode. Applying the model to the shock response of metals will require a data source for the Weibull parameter {alpha}{sub u}, short of measuing the strength of specimens of various sizes. Model validation will require more detail on the experiments successfully calculated by SPRUT. Evaluation of the KSS model is needed against other existing rate-dependent models for metals such as the Steinberg-Lund or MTS model on other shock experiments.
Statistical modelling of collocation uncertainty in atmospheric thermodynamic profiles
NASA Astrophysics Data System (ADS)
Fassò, A.; Ignaccolo, R.; Madonna, F.; Demoz, B. B.
2013-08-01
The uncertainty of important atmospheric parameters is a key factor for assessing the uncertainty of global change estimates given by numerical prediction models. One of the critical points of the uncertainty budget is related to the collocation mismatch in space and time among different observations. This is particularly important for vertical atmospheric profiles obtained by radiosondes or LIDAR. In this paper we consider a statistical modelling approach to understand at which extent collocation uncertainty is related to environmental factors, height and distance between the trajectories. To do this we introduce a new statistical approach, based on the heteroskedastic functional regression (HFR) model which extends the standard functional regression approach and allows us a natural definition of uncertainty profiles. Moreover, using this modelling approach, a five-folded uncertainty decomposition is proposed. Eventually, the HFR approach is illustrated by the collocation uncertainty analysis of relative humidity from two stations involved in GCOS reference upper-air network (GRUAN).
Statistical, Morphometric, Anatomical Shape Model (Atlas) of Calcaneus
Melinska, Aleksandra U.; Romaszkiewicz, Patryk; Wagel, Justyna; Sasiadek, Marek; Iskander, D. Robert
2015-01-01
The aim was to develop a morphometric and anatomically accurate atlas (statistical shape model) of calcaneus. The model is based on 18 left foot and 18 right foot computed tomography studies of 28 male individuals aged from 17 to 62 years, with no known foot pathology. A procedure for automatic atlas included extraction and identification of common features, averaging feature position, obtaining mean geometry, mathematical shape description and variability analysis. Expert manual assistance was included for the model to fulfil the accuracy sought by medical professionals. The proposed for the first time statistical shape model of the calcaneus could be of value in many orthopaedic applications including providing support in diagnosing pathological lesions, pre-operative planning, classification and treatment of calcaneus fractures as well as for the development of future implant procedures. PMID:26270812
Statistical mechanics models for motion and force planning
NASA Technical Reports Server (NTRS)
Rodriguez, G.
1990-01-01
The models of statistical mechanics provide an alternative to the methods of classical mechanics more traditionally used in robotics. They have a potential to: improve analysis of object collisions; handle kinematic and dynamic contact interactions within the same frmework; and reduce the need for perfect deterministic world model information. The statistical mechanics models characterize the state of the system as a probability density function (p.d.f.) whose time evolution is governed by a partial differential equation subject to boundary and initial conditions. The boundary conditions when rigid objects collide reflect the conservation of momentum. The models are being developed to embedd in remote semi-autonomous systems with a need to reason and interact with a multiobject environment.
Statistical-mechanical aids to calculating term-structure models
NASA Astrophysics Data System (ADS)
Ingber, Lester
1990-12-01
Recent work in statistical mechanics has developed new analytical and numerical techniques to solve coupled stochastic equations. This paper describes application of the very fast simulated reannealing and path-integral methodologies to the estimation of the Brennan and Schwartz two-factor term-structure (time-dependent) model of bond prices. It is shown that these methodologies can be utilized to estimate more complicated n-factor nonlinear models. Applications to other systems are stressed.
Bilingual Cluster Based Models for Statistical Machine Translation
NASA Astrophysics Data System (ADS)
Yamamoto, Hirofumi; Sumita, Eiichiro
We propose a domain specific model for statistical machine translation. It is well-known that domain specific language models perform well in automatic speech recognition. We show that domain specific language and translation models also benefit statistical machine translation. However, there are two problems with using domain specific models. The first is the data sparseness problem. We employ an adaptation technique to overcome this problem. The second issue is domain prediction. In order to perform adaptation, the domain must be provided, however in many cases, the domain is not known or changes dynamically. For these cases, not only the translation target sentence but also the domain must be predicted. This paper focuses on the domain prediction problem for statistical machine translation. In the proposed method, a bilingual training corpus, is automatically clustered into sub-corpora. Each sub-corpus is deemed to be a domain. The domain of a source sentence is predicted by using its similarity to the sub-corpora. The predicted domain (sub-corpus) specific language and translation models are then used for the translation decoding. This approach gave an improvement of 2.7 in BLEU score on the IWSLT05 Japanese to English evaluation corpus (improving the score from 52.4 to 55.1). This is a substantial gain and indicates the validity of the proposed bilingual cluster based models.
Predictive data modeling of human type II diabetes related statistics
NASA Astrophysics Data System (ADS)
Jaenisch, Kristina L.; Jaenisch, Holger M.; Handley, James W.; Albritton, Nathaniel G.
2009-04-01
During the course of routine Type II treatment of one of the authors, it was decided to derive predictive analytical Data Models of the daily sampled vital statistics: namely weight, blood pressure, and blood sugar, to determine if the covariance among the observed variables could yield a descriptive equation based model, or better still, a predictive analytical model that could forecast the expected future trend of the variables and possibly eliminate the number of finger stickings required to montior blood sugar levels. The personal history and analysis with resulting models are presented.
Statistical analysis of brain sulci based on active ribbon modeling
NASA Astrophysics Data System (ADS)
Barillot, Christian; Le Goualher, Georges; Hellier, Pierre; Gibaud, Bernard
1999-05-01
This paper presents a general statistical framework for modeling deformable object. This model is devoted being used in digital brain atlases. We first present a numerical modeling of brain sulci. We present also a method to characterize the high inter-individual variability of basic cortical structures on which the description of the cerebral cortex is based. The aimed applications use numerical modeling of brain sulci to assist non-linear registration of human brains by inter-individual anatomical matching or to better compare neuro-functional recordings performed on a series of individuals. The utilization of these methods is illustrated using a few examples.
Computer modelling of statistical properties of SASE FEL radiation
NASA Astrophysics Data System (ADS)
Saldin, E. L.; Schneidmiller, E. A.; Yurkov, M. V.
1997-06-01
The paper describes an approach to computer modelling of statistical properties of the radiation from self amplified spontaneous emission free electron laser (SASE FEL). The present approach allows one to calculate the following statistical properties of the SASE FEL radiation: time and spectral field correlation functions, distribution of the fluctuations of the instantaneous radiation power, distribution of the energy in the electron bunch, distribution of the radiation energy after monochromator installed at the FEL amplifier exit and the radiation spectrum. All numerical results presented in the paper have been calculated for the 70 nm SASE FEL at the TESLA Test Facility being under construction at DESY.
Occupation time statistics in the quenched trap model.
Burov, S; Barkai, E
2007-06-22
We investigate the distribution of the occupation time of a particle undergoing a random walk among random energy traps and in the presence of a deterministic potential field. When the distribution of energy traps is exponential with a width T(g), we find in thermal equilibrium a transition between Boltzmann statistics when T>T(g) to Lamperti statistics when T < T(g). We explain why our main results are valid for other models of quenched disorder, and discuss briefly implications on single particle experiments. PMID:17678005
Pre-equilibrium parton dynamics: Proceedings
Wang, Xin-Nian
1993-12-31
This report contains papers on the following topics: parton production and evolution; QCD transport theory; interference in the medium; QCD and phase transition; and future heavy ion experiments. This papers have been indexed separately elsewhere on the data base.
Statistical mechanics of network models of macroevolution and extinction
NASA Astrophysics Data System (ADS)
Solé, Ricard V.
The fossil record of life has been shown to provide evidence for scaling laws in both time series and in some statistical features. This evidence was suggested to be linked with a self-organized critical phenomenon by several authors. In this paper we review some of these models and their specific predictions. It is shown that most of the observed statistical properties of the evolutionary process on the long time scale can be reproduced by means of a simple model involving a network of interactions among species. The model is able to capture the essential features of the extinction and diversification process and gives power law distributions for (i) extinction events, (ii) taxonomy of species-genera data, (iii) lifetime distribution of genus close to those reported from paleontological databases. It also provides a natural decoupling between micro- and macroevolutionary processes.
Experimental, statistical, and biological models of radon carcinogenesis
Cross, F.T.
1991-09-01
Risk models developed for underground miners have not been consistently validated in studies of populations exposed to indoor radon. Imprecision in risk estimates results principally from differences between exposures in mines as compared to domestic environments and from uncertainties about the interaction between cigarette-smoking and exposure to radon decay products. Uncertainties in extrapolating miner data to domestic exposures can be reduced by means of a broad-based health effects research program that addresses the interrelated issues of exposure, respiratory tract dose, carcinogenesis (molecular/cellular and animal studies, plus developing biological and statistical models), and the relationship of radon to smoking and other copollutant exposures. This article reviews experimental animal data on radon carcinogenesis observed primarily in rats at Pacific Northwest Laboratory. Recent experimental and mechanistic carcinogenesis models of exposures to radon, uranium ore dust, and cigarette smoke are presented with statistical analyses of animal data. 20 refs., 1 fig.
Random matrix theory and classical statistical mechanics: Spin models
NASA Astrophysics Data System (ADS)
Meyer, H.; Angles D'Auriac, J.-C.
1997-06-01
We present a statistical analysis of spectra of transfer matrices of classical lattice spin models; this continues the work on the eight-vertex model of the preceding paper [H. Meyer, J.-C. Anglès d'Auriac, and J.-M. Maillard, Phys. Rev. E 55, 5261 (1997)]. We show that the statistical properties of these spectra can serve as a criterion of integrability. It also provides an operational numerical method to locate integrable varieties. In particular, we distinguish the notions of integrability and criticality, considering the two examples of the three-dimensional Ising critical point and the two-dimensional three-state Potts critical point. For complex spectra, which appear frequently in the context of transfer matrices, we show that the notion of independence of eigenvalues for integrable models still holds.
Normalized Texture Motifs and Their Application to Statistical Object Modeling
Newsam, S D
2004-03-09
A fundamental challenge in applying texture features to statistical object modeling is recognizing differently oriented spatial patterns. Rows of moored boats in remote sensed images of harbors should be consistently labeled regardless of the orientation of the harbors, or of the boats within the harbors. This is not straightforward to do, however, when using anisotropic texture features to characterize the spatial patterns. We here propose an elegant solution, termed normalized texture motifs, that uses a parametric statistical model to characterize the patterns regardless of their orientation. The models are learned in an unsupervised fashion from arbitrarily orientated training samples. The proposed approach is general enough to be used with a large category of orientation-selective texture features.
Parton physics on a Euclidean lattice.
Ji, Xiangdong
2013-06-28
I show that the parton physics related to correlations of quarks and gluons on the light cone can be studied through the matrix elements of frame-dependent, equal-time correlators in the large momentum limit. This observation allows practical calculations of parton properties on a Euclidean lattice. As an example, I demonstrate how to recover the leading-twist quark distribution by boosting an equal-time correlator to a large momentum. PMID:23848864
The midpoint between dipole and parton showers
NASA Astrophysics Data System (ADS)
Höche, Stefan; Prestel, Stefan
2015-09-01
We present a new parton-shower algorithm. Borrowing from the basic ideas of dipole cascades, the evolution variable is judiciously chosen as the transverse momentum in the soft limit. This leads to a very simple analytic structure of the evolution. A weighting algorithm is implemented that allows one to consistently treat potentially negative values of the splitting functions and the parton distributions. We provide two independent, publicly available implementations for the two event generators P ythia and S herpa.
The midpoint between dipole and parton showers
Höche, Stefan; Prestel, Stefan
2015-09-28
We present a new parton-shower algorithm. Borrowing from the basic ideas of dipole cascades, the evolution variable is judiciously chosen as the transverse momentum in the soft limit. This leads to a very simple analytic structure of the evolution. A weighting algorithm is implemented that allows one to consistently treat potentially negative values of the splitting functions and the parton distributions. Thus, we provide two independent, publicly available implementations for the two event generators PYTHIA and SHERPA.
Multi-region statistical shape model for cochlear implantation
NASA Astrophysics Data System (ADS)
Romera, Jordi; Kjer, H. Martin; Piella, Gemma; Ceresa, Mario; González Ballester, Miguel A.
2016-03-01
Statistical shape models are commonly used to analyze the variability between similar anatomical structures and their use is established as a tool for analysis and segmentation of medical images. However, using a global model to capture the variability of complex structures is not enough to achieve the best results. The complexity of a proper global model increases even more when the amount of data available is limited to a small number of datasets. Typically, the anatomical variability between structures is associated to the variability of their physiological regions. In this paper, a complete pipeline is proposed for building a multi-region statistical shape model to study the entire variability from locally identified physiological regions of the inner ear. The proposed model, which is based on an extension of the Point Distribution Model (PDM), is built for a training set of 17 high-resolution images (24.5 μm voxels) of the inner ear. The model is evaluated according to its generalization ability and specificity. The results are compared with the ones of a global model built directly using the standard PDM approach. The evaluation results suggest that better accuracy can be achieved using a regional modeling of the inner ear.
Statistical procedures for evaluating daily and monthly hydrologic model predictions
Coffey, M.E.; Workman, S.R.; Taraba, J.L.; Fogle, A.W.
2004-01-01
The overall study objective was to evaluate the applicability of different qualitative and quantitative methods for comparing daily and monthly SWAT computer model hydrologic streamflow predictions to observed data, and to recommend statistical methods for use in future model evaluations. Statistical methods were tested using daily streamflows and monthly equivalent runoff depths. The statistical techniques included linear regression, Nash-Sutcliffe efficiency, nonparametric tests, t-test, objective functions, autocorrelation, and cross-correlation. None of the methods specifically applied to the non-normal distribution and dependence between data points for the daily predicted and observed data. Of the tested methods, median objective functions, sign test, autocorrelation, and cross-correlation were most applicable for the daily data. The robust coefficient of determination (CD*) and robust modeling efficiency (EF*) objective functions were the preferred methods for daily model results due to the ease of comparing these values with a fixed ideal reference value of one. Predicted and observed monthly totals were more normally distributed, and there was less dependence between individual monthly totals than was observed for the corresponding predicted and observed daily values. More statistical methods were available for comparing SWAT model-predicted and observed monthly totals. The 1995 monthly SWAT model predictions and observed data had a regression Rr2 of 0.70, a Nash-Sutcliffe efficiency of 0.41, and the t-test failed to reject the equal data means hypothesis. The Nash-Sutcliffe coefficient and the R r2 coefficient were the preferred methods for monthly results due to the ability to compare these coefficients to a set ideal value of one.
Statistical analysis of target acquisition sensor modeling experiments
NASA Astrophysics Data System (ADS)
Deaver, Dawne M.; Moyer, Steve
2015-05-01
The U.S. Army RDECOM CERDEC NVESD Modeling and Simulation Division is charged with the development and advancement of military target acquisition models to estimate expected soldier performance when using all types of imaging sensors. Two elements of sensor modeling are (1) laboratory-based psychophysical experiments used to measure task performance and calibrate the various models and (2) field-based experiments used to verify the model estimates for specific sensors. In both types of experiments, it is common practice to control or measure environmental, sensor, and target physical parameters in order to minimize uncertainty of the physics based modeling. Predicting the minimum number of test subjects required to calibrate or validate the model should be, but is not always, done during test planning. The objective of this analysis is to develop guidelines for test planners which recommend the number and types of test samples required to yield a statistically significant result.
Organism-level models: When mechanisms and statistics fail us
NASA Astrophysics Data System (ADS)
Phillips, M. H.; Meyer, J.; Smith, W. P.; Rockhill, J. K.
2014-03-01
Purpose: To describe the unique characteristics of models that represent the entire course of radiation therapy at the organism level and to highlight the uses to which such models can be put. Methods: At the level of an organism, traditional model-building runs into severe difficulties. We do not have sufficient knowledge to devise a complete biochemistry-based model. Statistical model-building fails due to the vast number of variables and the inability to control many of them in any meaningful way. Finally, building surrogate models, such as animal-based models, can result in excluding some of the most critical variables. Bayesian probabilistic models (Bayesian networks) provide a useful alternative that have the advantages of being mathematically rigorous, incorporating the knowledge that we do have, and being practical. Results: Bayesian networks representing radiation therapy pathways for prostate cancer and head & neck cancer were used to highlight the important aspects of such models and some techniques of model-building. A more specific model representing the treatment of occult lymph nodes in head & neck cancer were provided as an example of how such a model can inform clinical decisions. A model of the possible role of PET imaging in brain cancer was used to illustrate the means by which clinical trials can be modelled in order to come up with a trial design that will have meaningful outcomes. Conclusions: Probabilistic models are currently the most useful approach to representing the entire therapy outcome process.
Unifying wildfire models from ecology and statistical physics.
Zinck, Richard D; Grimm, Volker
2009-11-01
Understanding the dynamics of wildfire regimes is crucial for both regional forest management and predicting global interactions between fire regimes and climate. Accordingly, spatially explicit modeling of forest fire ecosystems is a very active field of research, including both generic and highly specific models. There is, however, a second field in which wildfire has served as a metaphor for more than 20 years: statistical physics. So far, there has been only limited interaction between these two fields of wildfire modeling. Here we show that two typical generic wildfire models from ecology are structurally equivalent to the most commonly used model from statistical physics. All three models can be unified to a single model in which they appear as special cases of regrowth-dependent flammability. This local "ecological memory" of former fire events is key to self-organization in wildfire ecosystems. The unified model is able to reproduce three different patterns observed in real boreal forests: fire size distributions, fire shapes, and a hump-shaped relationship between disturbance intensity (average annual area burned) and diversity of succession stages. The unification enables us to bring together insights from both disciplines in a novel way and to identify limitations that provide starting points for further research. PMID:19799499
Statistics of a neuron model driven by asymmetric colored noise
NASA Astrophysics Data System (ADS)
Müller-Hansen, Finn; Droste, Felix; Lindner, Benjamin
2015-02-01
Irregular firing of neurons can be modeled as a stochastic process. Here we study the perfect integrate-and-fire neuron driven by dichotomous noise, a Markovian process that jumps between two states (i.e., possesses a non-Gaussian statistics) and exhibits nonvanishing temporal correlations (i.e., represents a colored noise). Specifically, we consider asymmetric dichotomous noise with two different transition rates. Using a first-passage-time formulation, we derive exact expressions for the probability density and the serial correlation coefficient of the interspike interval (time interval between two subsequent neural action potentials) and the power spectrum of the spike train. Furthermore, we extend the model by including additional Gaussian white noise, and we give approximations for the interspike interval (ISI) statistics in this case. Numerical simulations are used to validate the exact analytical results for pure dichotomous noise, and to test the approximations of the ISI statistics when Gaussian white noise is included. The results may help to understand how correlations and asymmetry of noise and signals in nerve cells shape neuronal firing statistics.
Spatio-temporal statistical models with applications to atmospheric processes
Wikle, C.K.
1996-12-31
This doctoral dissertation is presented as three self-contained papers. An introductory chapter considers traditional spatio-temporal statistical methods used in the atmospheric sciences from a statistical perspective. Although this section is primarily a review, many of the statistical issues considered have not been considered in the context of these methods and several open questions are posed. The first paper attempts to determine a means of characterizing the semiannual oscillation (SAO) spatial variation in the northern hemisphere extratropical height field. It was discovered that the midlatitude SAO in 500hPa geopotential height could be explained almost entirely as a result of spatial and temporal asymmetries in the annual variation of stationary eddies. It was concluded that the mechanism for the SAO in the northern hemisphere is a result of land-sea contrasts. The second paper examines the seasonal variability of mixed Rossby-gravity waves (MRGW) in lower stratospheric over the equatorial Pacific. Advanced cyclostationary time series techniques were used for analysis. It was found that there are significant twice-yearly peaks in MRGW activity. Analyses also suggested a convergence of horizontal momentum flux associated with these waves. In the third paper, a new spatio-temporal statistical model is proposed that attempts to consider the influence of both temporal and spatial variability. This method is mainly concerned with prediction in space and time, and provides a spatially descriptive and temporally dynamic model.
Contribution towards statistical intercomparison of general circulation models
Sengupta, S.; Boyle, J.
1995-06-01
The Atmospheric Model Intercomparison Project (AMIP) of the World Climate Research Programme`s Working Group on Numerical Experimentation (WGNE) is an ambitious attempt to comprehensively intercompare atmospheric General Circulation Models (GCMs). The participants in AMIP simulate the global atmosphere for the decade 1979 to 1988 using, a common solar constant and Carbon Dioxide(CO{sub 2}) concentration and a common monthly averaged sea surface temperature (SST) and sea ice data set. In this work we attempt to present a statistical framework to address the difficult task of model intercomparison and verification.
Multiple photon production in double parton scattering at the LHC
NASA Astrophysics Data System (ADS)
Palota da Silva, R.; Brenner Mariotto, C.; Goncalves, V. P.
2016-04-01
The high density of gluons in the initial state of hadronic collisions at LHC implies that the probability of multiple parton interactions within one proton-proton collision increases. In particular, the probability of having two or more hard interactions in a collision is not significantly suppressed with respect to the single interaction probability. In this contribution we study for the first time the production of prompt photons in double parton scattering processes. In particular, we estimate the rapidity distribution for the double Compton process, which leads to two photons plus two jets in the final state. Besides, we study the production of three and four photons in the final state, which are backgrounds to physics beyond the Standard Model.
Double Parton Fragmentation Function and its Evolution in Quarkonium Production
NASA Astrophysics Data System (ADS)
Kang, Zhong-Bo
2014-01-01
We summarize the results of a recent study on a new perturbative QCD factorization formalism for the production of heavy quarkonia of large transverse momentum pT at collider energies. Such a new factorization formalism includes both the leading power (LP) and next-to-leading power (NLP) contributions to the cross section in the mQ2/p_T^2 expansion for heavy quark mass mQ. For the NLP contribution, the so-called double parton fragmentation functions are involved, whose evolution equations have been derived. We estimate fragmentation functions in the non-relativistic QCD formalism, and found that their contribution reproduce the bulk of the large enhancement found in explicit NLO calculations in the color singlet model. Heavy quarkonia produced from NLP channels prefer longitudinal polarization, in contrast to the single parton fragmentation function. This might shed some light on the heavy quarkonium polarization puzzle.