Free-field representations and geometry of some Gepner models
Parkhomenko, S. E.
2010-09-15
The geometry of the k{sup K} Gepner model, where k + 2 = 2K, is investigated by a free-field representation known as the 'bc{beta}{gamma}' system. Using this representation, we directly show that the internal sector of the model is given by Landau-Ginzburg C{sup K}/Z{sub 2K} orbifold. Then we consider the deformation of the orbifold by a marginal antichiral-chiral operator. Analyzing the chiral de Rham complex structure in the holomorphic sector, we show that it coincides with chiral de Rham complex of some toric manifold, where toric data are given by certain fermionic screening currents. This allows relating the Gepner model deformed by the marginal operator to a {sigma}-model on the CY manifold realized as a double cover of p{sup K-1} with ramification along a certain submanifold.
N=4 characters in Gepner models, orbits and elliptic genera
NASA Astrophysics Data System (ADS)
Grünberg, Daniel B.
2003-12-01
We review the properties of characters of the N=4 superconformal algebra in the context of a nonlinear sigma model on K3, how they are used to span the orbits, and how the orbits produce topological invariants like the elliptic genus. We derive the same expression for the K3 elliptic genus using three different Gepner models (16, 24, and 43 theories), detailing the orbits and verifying that their coefficients Fi are given by elementary modular functions. We also reveal the orbits for the 1322, 144, and 1242 theories. We derive relations for cubes of theta functions and study the function (1/?)?n?Z(-1)n(6n+1)k q(6n+1)2/24 for k=1,2,3,4.
On gauge couplings and thresholds in Type I Gepner models and otherwise
NASA Astrophysics Data System (ADS)
Anastasopoulos, Pascal; Bianchi, Massimo; Sarkissian, Gor; Stanev, Yassen S.
2007-03-01
We derive general formulae for tree level gauge couplings and their one-loop thresholds in Type I models based on genuinely interacting internal Script N = 2 SCFT's, such as Gepner models. We illustrate our procedure in the simple yet non-trivial instance of the Quintic. We briefly address the phenomenologically more relevant issue of determining the Weinberg angle in this class of models. Finally we initiate the study of the correspondence between `magnetized' or `coisotropic' D-branes in Gepner models and twisted representations of the underlying Script N = 2 SCA.
Semiholographic model revisited
NASA Astrophysics Data System (ADS)
CÃ¡rdenas, VÃctor H.; MagaÃ±a, Juan; Villanueva, J. R.
2014-03-01
In a recent work, Zhang, Li and Noh proposed a model for dark energy assuming that this component strictly obeys the holographic principle. They performed a dynamical system analysis, finding a scaling solution which is helpful to solve the coincidence problem. However, they need explicitly a cosmological constant. In this paper, we derive an explicit analytical solution, without Î›, that shows agreement with the supernovae data. However, this solution is not physical because it violates all the energy conditions.
Bayesian Constrained Local Models Revisited.
Martins, Pedro; Henriques, Jo Ao F; Caseiro, Rui; Batista, Jorge
2016-04-01
This paper presents a novel Bayesian formulation for aligning faces in unseen images. Our approach revisits the Constrained Local Models (CLM) formulation where an ensemble of local feature detectors are constrained to lie within the subspace spanned by a Point Distribution Model (PDM). Fitting such a model to an image typically involves two main steps: a local search using a detector, obtaining response maps for each landmark (likelihood term) and a global optimization that finds the PDM parameters that jointly maximize all the detections at once. The so-called global optimization can be posed as a Bayesian inference problem, where the posterior distribution of the shape (and pose) parameters can be inferred in a maximum a posteriori (MAP) sense. This work introduces an extended Bayesian global optimization strategy that includes two novel additions: (1) to perform second order updates of the PDM parameters (accounting for their covariance) and (2) to model the underlying dynamics of the shape variations, encoded in the prior term, by using recursive Bayesian estimation. Extensive evaluations were performed against state-of-the-art methods on several standard datasets (IMM, BioID, XM2VTS, LFW and FGNET Talking Face). Results show that the proposed approach significantly increases the fitting performance. PMID:26959675
Revisiting the standard solar model
Turck-Chieze, S.; Cahen, S.; Casse, M.; Doom, C.
1988-12-01
The mutual consistency between standard solar models is studied based on the recent Los Alamos opacity tables. Satisfactory agreement is found among these models concerning the helium content and the neutrino capture rates. The reference model leads to a solar helium content of 0.276 + or - 0.012 by mass fraction. 75 references.
Galeta, Leonardo; Pirjol, Dan; Schat, Carlos
2009-12-01
We show how to match the Isgur-Karl model to the spin-flavor quark operator expansion used in the 1/N{sub c} studies of the nonstrange negative parity L=1 excited baryons. Using the transformation properties of states and interactions under the permutation group S{sub 3} we are able to express the operator coefficients as overlap integrals, without making any assumption on the spatial dependence of the quark wave functions. The general mass operator leads to parameter free mass relations and constraints on the mixing angles that are valid beyond the usual harmonic oscillator approximation. The Isgur-Karl model with harmonic oscillator wave functions provides a simple counterexample that demonstrates explicitly that the alternative operator basis for the 1/N{sub c} expansion for excited baryons recently proposed by Matagne and Stancu is incomplete.
The sausage sigma model revisited
NASA Astrophysics Data System (ADS)
Suneeta, Vardarajan
2015-06-01
Fateevâ€™s sausage sigma models in two and three dimensions are known to be integrable. We study their stability under renormalization group (RG) flow in the target space by using results from the mathematics of Ricci flow. We show that the three-dimensional sausage is unstable, whereas the two-dimensional sausage appears to be stable at least at leading order as it approaches the sphere. We speculate that the stability results obtained are linked to the classification of ancient solutions to Ricci flow (i.e., sigma models that are nonperturbative in the infrared regime) in two and three dimensions. We also describe a class of perturbations of the three-dimensional sausage (with the same continuous symmetries) which remarkably decouple. This indicates that there could be a new solution to RG flow, which is described at least perturbatively as a deformation of the sausage.
Fusion by Diffusion Model Revisited
NASA Astrophysics Data System (ADS)
Cap, T.; Siwek-WilczyÅ„ska, K.; WilczyÅ„ski, J.
A complete set of 27 excitation functions for synthesis of superheavy nuclei produced in cold fusion reactions was analyzed in terms of the "Fusion by Diffusion Model" of Åšwiatecki et al., modified to account for the angular momentum dependence of the fusion hindrance factor. The data on cold fusion reactions originate from experiments carried out at GSI Darmstadt, RIKEN Tokyo and LBNL Berkeley in which 208Pb and 209Bi targets were bombarded with the variety of projectiles ranging from 48,50Ti to 70Zn.
A Multi-Level Model of Moral Functioning Revisited
ERIC Educational Resources Information Center
Reed, Don Collins
2009-01-01
The model of moral functioning scaffolded in the 2008 "JME" Special Issue is here revisited in response to three papers criticising that volume. As guest editor of that Special Issue I have formulated the main body of this response, concerning the dynamic systems approach to moral development, the problem of moral relativism and the role of…
Topological Twisted Sigma Model with H-flux Revisited
Chuang, Wu-yen
2006-08-18
In this paper we revisit the topological twisted sigma model with H-flux. We explicitly expand and then twist the worldsheet Lagrangian for bi-Hermitian geometry. we show that the resulting action consists of a BRST exact term and pullback terms, which only depend on one of the two generalized complex structures and the B-field. We then discuss the topological feature of the model.
Extended primitive models of water revisited
NASA Astrophysics Data System (ADS)
Strnad, Martin; Nezbeda, Ivo
A 4-site extended primitive model of water descending from the TIP4 potential has been reexamined over a range of the model parameters. It has been found, in contrast to recently reported results (Nezbeda, I., and Slovak, J., 1997, Molec. Phys. , 90, 353), that the model clearly is superior to 5-site models descending from the ST2 potential and yields the liquid structure in very good agreement with that of real water.
Morse's Markov Model of Book Use Revisited.
ERIC Educational Resources Information Center
Beheshti, Jamshid; Tague, Jean M.
1984-01-01
Utilizing 11 years of University of Saskatchewan circulation transactions, it is shown that P. M. Morse's Markov model of book use fits approximately 99 percent of data for whole collection and for three subject areas; contrary to his assumptions, one of the model's parameters is time dependent. Thirteen references are included. (EJS)
Sverdlovsk revisited: modeling human inhalation anthrax.
Wilkening, Dean A
2006-05-16
Several models have been proposed for the dose-response function and the incubation period distribution for human inhalation anthrax. These models give very different predictions for the severity of a hypothetical bioterror attack, when an attack might be detected from clinical cases, the efficacy of medical intervention and the requirements for decontamination. Using data from the 1979 accidental atmospheric release of anthrax in Sverdlovsk, Russia, and limited nonhuman primate data, this paper eliminates two of the contending models and derives parameters for the other two, thereby narrowing the range of models that accurately predict the effects of human inhalation anthrax. Dose-response functions that exhibit a threshold for infectivity are contraindicated by the Sverdlovsk data. Dose-dependent incubation period distributions explain the 10-day median incubation period observed at Sverdlovsk and the 1- to 5-day incubation period observed in nonhuman primate experiments. PMID:16679412
Complex trait architecture: the pleiotropic model revisited
North, T.-L.; Beaumont, M. A.
2015-01-01
There is currently much debate about how much the genetic heritability of complex traits is due to very rare alleles. This issue is important because it determines sampling strategies for genetic association studies. Several recent theoretical papers based on a pleiotropic model for trait evolution suggest that it is possible that a large proportion of the genetic variance could be explained by rare alleles. This model assumes that mutations with a large effect on fitness also tend to have large positive or negative effects on phenotypic traits. We show that conclusions based on standard diffusion results are generally applicable to simulations of whole genomes with overlapping generations in a finite population, although the variance contribution of rare alleles is somewhat smaller than theoretical predictions. We show that under many scenarios the pleiotropic model predicts trait distributions that are unrealistically leptokurtic. We argue that this imposes a limit on the relationship between fitness and trait effects. PMID:25792462
Potts-model critical manifolds revisited
NASA Astrophysics Data System (ADS)
Scullard, Christian R.; Lykke Jacobsen, Jesper
2016-03-01
We compute critical polynomials for the q-state Potts model on the Archimedean lattices, using a parallel implementation of the algorithm of Jacobsen (2014 J. Phys. A: Math. Theor 47 135001) that gives us access to larger sizes than previously possible. The exact polynomials are computed for bases of size 6 Ã— 6 unit cells, and the root in the temperature variable v={{{e}}}K-1 is determined numerically at q = 1 for bases of size 8 Ã— 8. This leads to improved results for bond percolation thresholds, and for the Potts-model critical manifolds in the real (q, v) plane. In the two most favourable cases, we find now the kagome-lattice threshold to eleven digits and that of the (3,{12}2) lattice to thirteen. Our critical manifolds reveal many interesting features in the antiferromagnetic region of the Potts model, and determine accurately the extent of the Berkerâ€“Kadanoff phase for the lattices studied.
The Bonn nuclear quark model revisited
Providencia, Constanca; Providencia, Joao da Cordeiro, Flavio; Yamamura, Masatoshi; Tsue, Yasuhiko; Nishiyama, Seiya
2009-08-15
We present the exact solutions to the equations of the lowest energy states of the colored and color-symmetric sectors of the Bonn quark model, which is SU(3) symmetric and is defined in terms of an effective pairing force with su(4) algebraic structure. We show that the groundstate of the model is not color symmetrical except for a narrow interval in the range of possible quark numbers. We also study the performance of the Glauber coherent state, as well as of superconducting states of the BCS type, with respect to the description, not only of the absolute (colored) groundstate, but also of the minimum energy state of the color-symmetrical sector, finding that it is remarkably good. We use the model to discuss, in a schematic context, some controversial aspects of the conventional treatment of color superconductivity.
Diffusion approximation of neuronal models revisited.
Cupera, Jakub
2014-02-01
Leaky integrate-and-fire neuronal models with reversal potentials have a number of different diffusion approximations, each depending on the form of the amplitudes of the postsynaptic potentials. Probability distributions of the first-passage times of the membrane potential in the original model and its diffusion approximations are numerically compared in order to find which of the approximations is the most suitable one. The properties of the random amplitudes of postsynaptic potentials are discussed. It is shown on a simple example that the quality of the approximation depends directly on them. PMID:24245676
Crime, criminals, and cures: medical model revisited.
Sampson, R J
2000-06-01
David Lykken's target article assesses the causes of crime and advocates a controversial "cure"--parental licensure. Although Lykken gets many of the facts about criminals right, ultimately the disease metaphor breaks down. Crime requires three things--motivated offenders ("criminals"), suitable targets or victims, and the absence of capable guardians to prevent the act. Typical of medical model approaches, failure to consider the convergence in time and space of the three necessary elements for crime results in a misdiagnosis. In this invited commentary, I briefly note three reasons why Lykken's cure, along with the medical model in general, is unlikely to bear fruit. PMID:10831313
The Hopfield model revisited: covariance and quantization
NASA Astrophysics Data System (ADS)
Belgiorno, F.; Cacciatori, S. L.; Dalla Piazza, F.
2016-01-01
There are several possible applications of quantum electrodynamics in dielectric media which require a quantum description for the electromagnetic field interacting with matter fields. The associated quantum models can refer to macroscopic electromagnetic fields or, alternatively, to mesoscopic fields (polarization fields) describing an effective interaction between electromagnetic field and matter fields. We adopt the latter approach, and focus on the Hopfield model for the electromagnetic field in a dielectric dispersive medium in a framework in which space–time dependent mesoscopic parameters occur, like susceptibility, matter resonance frequency, and also coupling between electromagnetic field and polarization field. Our most direct goal is to describe in a phenomenological way a space–time varying dielectric perturbation induced by means of the Kerr effect in nonlinear dielectric media. This extension of the model is implemented by means of a Lorentz-invariant Lagrangian which, for constant microscopic parameters, and in the rest frame, coincides with the standard one. Moreover, we deduce a covariant scalar product and provide a canonical quantization scheme which takes into account the constraints implicit in the model. Examples of viable applications are indicated.
Density waves in the Calogero model - revisited
Bardek, V. Feinberg, J. Meljanac, S.
2010-03-15
The Calogero model bears, in the continuum limit, collective excitations in the form of density waves and solitary modulations of the density of particles. This sector of the spectrum of the model was investigated, mostly within the framework of collective-field theory, by several authors, over the past 15 years or so. In this work we shall concentrate on periodic solutions of the collective BPS-equation (also known as 'finite amplitude density waves'), as well as on periodic solutions of the full static variational equations which vanish periodically (also known as 'large amplitude density waves'). While these solutions are not new, we feel that our analysis and presentation add to the existing literature, as we explain in the text. In addition, we show that these solutions also occur in a certain two-family generalization of the Calogero model, at special points in parameter space. A compendium of useful identities associated with Hilbert transforms, including our own proofs of these identities, appears in Appendix A. In Appendix B we also elucidate in the present paper some fine points having to do with manipulating Hilbert-transforms, which appear ubiquitously in the collective field formalism. Finally, in order to make this paper self-contained, we briefly summarize in Appendix C basic facts about the collective field formulation of the Calogero model.
Revisiting the vanishing refuge model of diversification.
Damasceno, Roberta; Strangas, Maria L; Carnaval, Ana C; Rodrigues, Miguel T; Moritz, Craig
2014-01-01
Much of the debate around speciation and historical biogeography has focused on the role of stabilizing selection on the physiological (abiotic) niche, emphasizing how isolation and vicariance, when associated with niche conservatism, may drive tropical speciation. Yet, recent re-emphasis on the ecological dimensions of speciation points to a more prominent role of divergent selection in driving genetic, phenotypic, and niche divergence. The vanishing refuge model (VRM), first described by Vanzolini and Williams (1981), describes a process of diversification through climate-driven habitat fragmentation and exposure to new environments, integrating both vicariance and divergent selection. This model suggests that dynamic climates and peripheral isolates can lead to genetic and functional (i.e., ecological and phenotypic) diversity, resulting in sister taxa that occupy contrasting habitats with abutting distributions. Here, we provide predictions for populations undergoing divergence according to the VRM that encompass habitat dynamics, phylogeography, and phenotypic differentiation across populations. Such integrative analyses can, in principle, differentiate the operation of the VRM from other speciation models. We applied these principles to a lizard species, Coleodactylus meridionalis, which was used to illustrate the model in the original paper. We incorporate data on inferred historic habitat dynamics, phylogeography and thermal physiology to test for divergence between coastal and inland populations in the Atlantic Forest of Brazil. Environmental and genetic analyses are concordant with divergence through the VRM, yet physiological data are not. We emphasize the importance of multidisciplinary approaches to test this and alternative speciation models while seeking to explain the extraordinarily high genetic and phenotypic diversity of tropical biomes. PMID:25374581
Revisiting the vanishing refuge model of diversification
Damasceno, Roberta; Strangas, Maria L.; Carnaval, Ana C.; Rodrigues, Miguel T.; Moritz, Craig
2014-01-01
Much of the debate around speciation and historical biogeography has focused on the role of stabilizing selection on the physiological (abiotic) niche, emphasizing how isolation and vicariance, when associated with niche conservatism, may drive tropical speciation. Yet, recent re-emphasis on the ecological dimensions of speciation points to a more prominent role of divergent selection in driving genetic, phenotypic, and niche divergence. The vanishing refuge model (VRM), first described by Vanzolini and Williams (1981), describes a process of diversification through climate-driven habitat fragmentation and exposure to new environments, integrating both vicariance and divergent selection. This model suggests that dynamic climates and peripheral isolates can lead to genetic and functional (i.e., ecological and phenotypic) diversity, resulting in sister taxa that occupy contrasting habitats with abutting distributions. Here, we provide predictions for populations undergoing divergence according to the VRM that encompass habitat dynamics, phylogeography, and phenotypic differentiation across populations. Such integrative analyses can, in principle, differentiate the operation of the VRM from other speciation models. We applied these principles to a lizard species, Coleodactylus meridionalis, which was used to illustrate the model in the original paper. We incorporate data on inferred historic habitat dynamics, phylogeography and thermal physiology to test for divergence between coastal and inland populations in the Atlantic Forest of Brazil. Environmental and genetic analyses are concordant with divergence through the VRM, yet physiological data are not. We emphasize the importance of multidisciplinary approaches to test this and alternative speciation models while seeking to explain the extraordinarily high genetic and phenotypic diversity of tropical biomes. PMID:25374581
'Andesite Model' of Continental Crust Formation Revisited
NASA Astrophysics Data System (ADS)
Gill, J. B.
2014-12-01
Fifty years ago, Ross Taylor observed that the chemical composition of average continental crust and orogenic andesites were similar for many major and trace elements. However, almost immediately it was recognized that this generalization does not apply to many oceanic island arcs because they are too mafic and too depleted in the most incompatible elements, including K, Th, and U. Although intra-crustal differentiation and delamination might produce sufficiently evolved middle and lower crust, it is an inefficient way to increase incompatible element concentrations in intermediate magma. There are alternative processes. First, because the vast majority of arc rocks are volcaniclastic sediments rather than igneous rocks, mechanical mixing is a dominant process during arc evolution. Although its effects are similar to those of magma mixing and assimilation, it has fewer physical constraints and is more likely. The finer-grained the sediment, the more mixing is likely. Drilling in the Izu arc shows that the majority of the sediment is andesitic mud even though the majority of lavas are basalt and rhyolite. Volcaniclastic sediments may be a better match to continental crust than the lavas that were the original basis of the andesite model. Second, some reararc lavas and sediments are more like continental crust than those from volcanic fronts, and reararc crust is at least as thick. Third, aeolian dust in arc muds may be a larger mass fraction of recycled continental crust in oceanic arcs than the subducted pelagic sediment component in arc magmas. The net effect of these features adds nuance to the andesite model but is not enough to overcome the deficit of incompatible elements in Cenozoic oceanic arcs. Accreting them still makes the continents less continental in composition.
Neural network modelling of CIMIS-ET0 (revisited)
NASA Astrophysics Data System (ADS)
Abrahart, R. J.; Ghani, N. Ab
2009-04-01
This paper will revisit the use of four independent hydrometeorological variables to predict 'reference crop evapotranspiration' in a neural network model - calculated as CIMIS-ET0 (Kisi, 2006; Aytek et al., 2008). The two earlier studies are coalesced and their published findings positioned in a broader environmental modelling context. Four models developed on similar datasets are compared and contrasted in the current exercise: a multiple linear regression model (MLIN: Pearson, 1896), a piecewise multiple linear regression model (M5 Model Tree; M5MT: Quinlan, 1992; Wang & Witten, 1997) and two neural network models developed on different optimisation algorithms - Conjugate Gradient (CGNN: Hestenes & Stiefel, 1952) and Levenberg-Marquet (LMNN: Levenberg, 1944; Marquardt, 1963). The results are presented using residual scatterplots so that the exact nature of the each individual modelling solution can be determined: permitting outputs to be interpreted in terms of structures, symmetries, orientations, local features and outliers. The reported inspection and interpretation of plots is matched against a selection of traditional numerical modelling statistics that were computed on HydroTest (http://www.hydrotest.org.uk; Dawson et al., 2007). The reported closeness of earlier neurocomputing outputs to predicted values estimated using a counterpart multiple linear regression model is explained in detail.
The maternal deprivation animal model revisited.
Marco, Eva M; Llorente, Ricardo; López-Gallardo, Meritxell; Mela, Virginia; Llorente-Berzal, Álvaro; Prada, Carmen; Viveros, María-Paz
2015-04-01
Early life stress, in the form of MD (24h at pnd 9), interferes with brain developmental trajectories modifying both behavioral and neurobiochemical parameters. MD has been reported to enhance neuroendocrine responses to stress, to affect emotional behavior and to impair cognitive function. More recently, changes in body weight gain, metabolic parameters and immunological responding have also been described. Present data give support to the fact that neuronal degeneration and/or astrocyte proliferation are present in specific brain regions, mainly hippocampus, prefrontal cortex and hypothalamus, which are particularly vulnerable to the effects of neonatal stress. The MD animal model arises as a valuable tool for the investigation of the brain processes occurring at the narrow time window comprised between pnd 9 and 10 that are critical for the establishment of brain circuitries critical for the regulation of behavior, metabolism and energy homeostasis. In the present review we will discuss three possible mechanisms that might be crucial for the effects of MD, namely, the rapid increase in glucocorticoids, the lack of the neonatal leptin surge, and the enhanced endocannabinoid signaling during the specific critical period of MD. A better understanding of the mechanisms underlying the detrimental consequences of MD is a concern for public health and may provide new insights into mental health prevention strategies and into novel therapeutic approaches in neuropsychiatry. PMID:25616179
How clonal are Neisseria species? The epidemic clonality model revisited
Tibayrenc, Michel; Ayala, Francisco J.
2015-01-01
The three species Neisseria meningitidis, Neisseria gonorrheae, and Neisseria lactamica are often regarded as highly recombining bacteria. N. meningitidis has been considered a paradigmatic case of the â€œsemiclonal modelâ€ or of â€œepidemic clonality,â€ demonstrating occasional bouts of clonal propagation in an otherwise recombining species. In this model, occasional clonality generates linkage disequilibrium in the short term. In the long run, however, the effects of clonality are countered by recombination. We show that many data are at odds with this proposal and that N. meningitidis fits the criteria that we have proposed for predominant clonal evolution (PCE). We point out that (i) the proposed way to distinguish epidemic clonality from PCE may be faulty and (ii) the evidence of deep phylogenies by microarrays and whole-genome sequencing is at odds with the predictions of the semiclonal model. Last, we revisit the species status of N. meningitidis, N. gonorrheae, and N. lactamica in the light of the PCE model. PMID:26195766
Richards model revisited: validation by and application to infection dynamics.
Wang, Xiang-Sheng; Wu, Jianhong; Yang, Yong
2012-11-21
Ever since Richards proposed his flexible growth function more than half a century ago, it has been a mystery that this empirical function has made many incredible coincidences with real ecological or epidemic data even though one of its parameters (i.e., the exponential term) does not seem to have clear biological meaning. It is therefore a natural challenge to mathematical biologists to provide an explanation of the interesting coincidences and a biological interpretation of the parameter. Here we start from a simple epidemic SIR model to revisit Richards model via an intrinsic relation between both models. Especially, we prove that the exponential term in the Richards model has a one-to-one nonlinear correspondence to the basic reproduction number of the SIR model. This one-to-one relation provides us an explicit formula in calculating the basic reproduction number. Another biological significance of our study is the observation that the peak time is approximately just a serial interval after the turning point. Moreover, we provide an explicit relation between final outbreak size, basic reproduction number and the peak epidemic size which means that we can predict the final outbreak size shortly after the peak time. Finally, we introduce a constraint in Richards model to address over fitting problem observed in the existing studies and then apply our method with constraint to conduct some validation analysis using the data of recent outbreaks of prototype infectious diseases such as Canada 2009 H1N1 outbreak, GTA 2003 SARS outbreak, Singapore 2005 dengue outbreak, and Taiwan 2003 SARS outbreak. Our new formula gives much more stable and precise estimate of model parameters and key epidemic characteristics such as the final outbreak size, the basic reproduction number, and the turning point, compared with earlier simulations without constraints. PMID:22889641
Effective-Medium Models for Marine Gas Hydrates, Mallik Revisited
NASA Astrophysics Data System (ADS)
Terry, D. A.; Knapp, C. C.; Knapp, J. H.
2011-12-01
Hertz-Mindlin type effective-medium dry-rock elastic models have been commonly used for more than three decades in rock physics analysis, and recently have been applied to assessment of marine gas hydrate resources. Comparisons of several effective-medium models with derivative well-log data from the Mackenzie River Valley, Northwest Territories, Canada (i.e. Mallik 2L-38 and 5L-38) were made several years ago as part of a marine gas hydrate joint industry project in the Gulf of Mexico. The matrix/grain supporting model (one of the five models compared) was clearly a better representation of the Mallik data than the other four models (2 cemented sand models; a pore-filling model; and an inclusion model). Even though the matrix/grain supporting model was clearly better, reservations were noted that the compressional velocity of the model was higher than the compressional velocity measured via the sonic logs, and that the shear velocities showed an even greater discrepancy. Over more than thirty years, variations of Hertz-Mindlin type effective medium models have evolved for unconsolidated sediments and here, we briefly review their development. In the past few years, the perfectly smooth grain version of the Hertz-Mindlin type effective-medium model has been favored over the infinitely rough grain version compared in the Gulf of Mexico study. We revisit the data from the Mallik wells to review assertions that effective-medium models with perfectly smooth grains are a better predictor than models with infinitely rough grains. We briefly review three Hertz-Mindlin type effective-medium models, and standardize nomenclature and notation. To calibrate the extended effective-medium model in gas hydrates, we use a well accepted framework for unconsolidated sediments through Hashin-Shtrikman bounds. We implement the previously discussed effective-medium models for saturated sediments with gas hydrates and compute theoretical curves of seismic velocities versus gas hydrate saturation to compare with well log data available from the Canadian gas hydrates research site. By directly comparing the infinitely rough and perfectly smooth grain versions of the Hertz-Mindlin type effective-medium model, we provide additional insight to the discrepancies noted in the Gulf of Mexico study.
On kaonic hydrogen. Phenomenological quantum field theoretic model revisited
NASA Astrophysics Data System (ADS)
Ivanov, A. N.; Cargnelli, M.; Faber, M.; Fuhrmann, H.; Ivanova, V. A.; Marton, J.; Troitskaya, N. I.; Zmeskal, J.
2005-09-01
We argue that due to isospin and U-spin invariance of strong low-energy interactions the S-wave scattering lengths a 0 0 and a 1 0 of ¯N scattering with isospin I = 0 and I = 1 satisfy the low-energy theorem a 0 0 +3a 1 0 = 0 valid to leading order in chiral expansion. In the model of strong low-energy ¯N interactions at threshold (Eur. Phys. J. A 21, 11 (2004)) we revisit the contribution of the ?(1750) resonance, which does not saturate the low-energy theorem a 0 0 +3a 1 0 = 0, and replace it by the baryon background with properties of an SU(3) octet. We calculate the S-wave scattering amplitudes of K-N and K-d scattering at threshold. We calculate the energy level displacements of the ground states of kaonic hydrogen and deuterium. The result obtained for kaonic hydrogen agrees well with recent experimental data by the DEAR Collaboration. We analyse the cross-sections for elastic and inelastic K-p scattering for laboratory momenta 70MeV/c < p K < 150MeV/c of the incident K--meson. The theoretical results agree with the available experimental data within two standard deviations.
Revisiting the use of hyperdiffusivities in numerical dynamo models
NASA Astrophysics Data System (ADS)
Fournier, A.; Aubert, J.
2012-04-01
The groundbreaking numerical dynamo models of Glatzmaier & Roberts (1995) and Kuang & Bloxham (1997) received some criticism due to their use of hyperdiffusivities, whereby small scale processes artificially experience much stronger dissipation than large scale processes. This stronger dissipation they chose was anisotropic, in that it was only effective in the horizontal direction, and parameterized in spectral space using the following generic formula for any diffusive parameter Î½ Î½(l) = Î½0 ifl â‰¤ l0, Î½(l) = Î½0[1 + a(l- l0)n] ifl > l0, in which l is the spherical harmonic degree, Î½0 is a reference value, l0 is the degree above which hyperdiffusivities start operating, and a and n are real numbers. Following the same choice as the studies mentioned above (which had most notably l0 = 0), Grote & Busse (2000) showed in a fully nonlinear context that the usage of hyperdiffusivities could lead to substantially different dynamics and magnetic field generation mechanisms. Without questioning the physical relevance of this parameterization of subgrid scale processes, we wish here to revisit the use of hyperdiffusivities (as defined mathematically above), on the account of the observation that today's models are run with a truncation at much larger spherical harmonic degree than early models. Consequently, they do not require hyperdiffusivities to kick in at the largest scales (l0 can be set to several tens). An exploration of those regions of parameter space less accessible to numerical models could therefore benefit from their use, provided they do not alter noticeably the largest scales of the dynamo (which are the ones expressing themselves in the record of the geomagnetic secular variation). We compare the statistics of a direct numerical simulation with the statistics of several hyperdiffusive simulations. In the prospect of exploring the parameter space and constructing statistics for their subsequent use for geomagnetic data assimilation practice, we conclude that a sensible use of hyperdiffusivities can lead to a much wanted decrease in computational cost, while not altering the nature of the solution.
Friction versus dilation revisited: insights from theoretical and numerical models
NASA Astrophysics Data System (ADS)
Makedonska, N.; Sparks, D. W.; Aharonov, E.; Goren, L.
2009-12-01
The intimate relation between apparent friction of shearing granular layers and their dilation was already discussed by Mead in 1925. Motivated by the importance of this connection to the frictional strength of geological faults and to earthquake generation, many laboratory and numerical experiments on sheared granular layers investigated the relation between the apparent friction, ?a, and the dilation (under most situations equivalent to the change in porosity). Apparent friction is defined as the ratio of the externally-applied shear stress to the stress applied normal to the layer, measured during constant shear strain rate. Although the nature of the connection is not very well established, ?a is often cited to be the sum of two contributions: 1. The surface friction coefficient, ?s, of the grains and 2. The dilation rate. The contribution of the dilation rate to ?a arises since dilation is required to allow grain rearrangement during shear, yet dilation requires input of work against the normal stress. We revisit the connection between apparent friction and dilation using theoretical treatment of two-dimensional sheared uniform granular layers and complementary Discrete Element simulations, both for gouge layers and for a rough surface without gouge. Our theoretical calculation shows that fluctuations in both ?a and dilation rate that occur during a particular type grain-scale shear motion follow a relationship that is non-linear, although in practice appears close to linear. Results show that dilation (and hence ?a) is connected to shear localization. In numerical simulations of mono-sized gouge layers (without grain breaking or chemical processes) shear localization occurs but does not persist; instead the systems fluctuate between a state of distributed shear and dilation and a localized motion on short-lived shear planes, with overall compaction. The transition between these two types of shear involves a large change in dilation rate, and leads to large deviations from the sublinear friction-dilation rate relationship. Models with non-uniform grains also show significant scatter about the linear relationship, and we attribute this to short-lived temporal and spatial variations in the extent of shear localization. We discuss the physical origin of these fluctuations.
Revisiting the model predicting maximal 2-3 mixing and CP violation for neutrinos
NASA Astrophysics Data System (ADS)
Takasugi, Eiichi
2015-11-01
The model of the neutrino mass matrix that we proposed in 2000 is revisited in the light of the recent T2K experiments. This model has the special property that it predicts maximal 2-3 mixing and CP violation under some simple condition. In this model, if the condition is relaxed, the 2-3 angle and the CP violation deviate from their maximal values and are related. We present such relations for typical cases.
Revisiting the Mode-Beating Model of AC Helicity Injection
NASA Astrophysics Data System (ADS)
Sauppe, J. P.; Sovinec, C. R.
2010-11-01
Oscillating field current drive (OFCD), or AC helicity injection, is an important candidate for current sustainment in reversed-field pinch devices. Bellan examined AC helicity injection in a slab geometry and described it as a beating between two plasma modes that produces a mean current parallel to the equilibrium magnetic field [P. M. Bellan. Phys. Rev. Lett. 54, 1381 (1985)]. This mean current is confined to within a classical resistive skin depth of the plasma surface, and plasma relaxation is responsible for transporting this current to the core. We revisit this analytical work and examine how this wave-beating effect is represented in zero-beta MHD simulations, including consideration of the choice of boundary conditions. In addition to the expected parallel current, numerical simulations show a pinch effect from a cycle-averaged current that is perpendicular to the mean magnetic field, which is not described in Bellan's original work. Our results are discussed with respect to Boozer's general anti-dynamo theorem [A. H. Boozer. Phys. Fluids B Vol. 5, 2271 (1993)].
Revisiting the Boundary Layer Leaf Water Isotopic Model
NASA Astrophysics Data System (ADS)
Feng, X.; Shu, Y.; Posmentier, E. S.; Sonder, L. J.; Yakir, D.
2007-12-01
The boundary layer (BL) model for oxygen or hydrogen isotopic composition of leaf water has been widely used in the past four decades, and has been incorporated into models that require information about leaf water isotopic variations. However, since its introduction, model predictions of the bulk leaf water have often exceeded observed isotopic enrichments. There are also cases in which the model yielded lower than observed isotopic enrichments of bulk leaf water. In general, underpredictions occur under relatively high humidity. In order to explain why the BL model overpredicts the isotopic composition, several modifications of the model have been proposed. However, no explanation exists for why the BL model underestimates observed isotopic enrichments. We recently developed a 2D model that successfully simulates the observed along-leaf 18O enrichment of pine needles, and can explain why the BL model could have over- or under-predicted the bulk leaf water ?18O values. In the BL model, bulk leaf water is isotopically equivalent to water at the evaporation site, fed directly by stem water. In a real leaf, however, stem water enters the base of the leaf and becomes progressively enriched in 18O towards the tip due to fractionation by transpiration, consistent with both our observations and behavior of the 2D model. Therefore, at least part of the leaf water, that near the base, would have isotopic values lower than water at transpiration sites predicted by the BL model, which might thus overestimate bulk isotope values. On the other hand, as water moves through a leaf, it becomes increasingly enriched in 18O, and the leaf water near the tip may have ?18O values well above the BL model prediction. Therefore, it is also possible for the BL model to underestimate the bulk leaf water ?18O. The actual isotopic composition of the bulk leaf water is a combination of these two effects. It is clear then that the BL model may not accurately predict the ?18O value of the bulk leaf water, because to do so would require the volumetric average of the ?18O in the depleted and enriched parts of the leaf to exactly equal the BL model prediction. Furthermore, if leaves are assumed to transpire fast under low humidity, our 2D model can also reproduce the humidity dependence of the discrepancy between observation and the BL model prediction. This suggests an interaction between environmental conditions and the physiological behavior of plants. If the simplicity of the BL model justifies its continued use, then it is important to investigate its accuracy further under different environmental conditions and for leaves with different morphologies and water transport pathways.
Revisiting "Discrepancy Analysis in Continuing Medical Education: A Conceptual Model"
ERIC Educational Resources Information Center
Fox, Robert D.
2011-01-01
Based upon a review and analysis of selected literature, the author presents a conceptual model of discrepancy analysis evaluation for planning, implementing, and assessing the impact of continuing medical education (CME). The model is described in terms of its value as a means of diagnosing errors in the development and implementation of CME. The…
Revisiting the domain model for lithium intercalated graphite
Krishnan, Sridevi; Brenet, Gilles; Caliste, Damien; Genovese, Luigi; Deutsch, Thierry; Pochet, Pascal
2013-12-16
In this Letter, we study the stability of the domain model for lithium intercalated graphite in stages III and II by means of Density Functional Theory and Kinetic Lattice Monte Carlo simulations. We find that the domain model is either thermodynamically or kinetically stable when compared to the standard model in stages III and II. The existence of domains in the intercalation sequence is well supported by recent high resolution transmission electron microscope observations in lithiated graphite. Moreover, we predict that such domain staging sequences leads to a wide range of diffusivity as reported in experiments.
Running of radiative neutrino masses: the scotogenic model — revisited
NASA Astrophysics Data System (ADS)
Merle, Alexander; Platscher, Moritz
2015-11-01
A few years ago, it had been shown that effects stemming from renormalisation group running can be quite large in the scotogenic model, where neutrinos obtain their mass only via a 1-loop diagram (or, more generally, in many models in which the light neutrino mass is generated via quantum corrections at loop-level). We present a new computation of the renormalisation group equations (RGEs) for the scotogenic model, thereby updating previous results. We discuss the matching in detail, in particular in what regards the different mass spectra possible for the new particles involved. We furthermore develop approximate analytical solutions to the RGEs for an extensive list of illustrative cases, covering all general tendencies that can appear in the model. Comparing them with fully numerical solutions, we give a comprehensive discussion of the running in the scotogenic model. Our approach is mainly top-down, but we also discuss an attempt to get information on the values of the fundamental parameters when inputting the low-energy measured quantities in a bottom-up manner. This work serves the basis for a full parameter scan of the model, thereby relating its low- and high-energy phenomenology, to fully exploit the available information.
Zwislocki's model of the middle ear re-visited
NASA Astrophysics Data System (ADS)
Withnell, Robert H.; Fields, Taylor N.
2015-12-01
Zwislocki's circuit model of the middle ear [11] has been used, in original or modified form, in subsequent studies modeling the ear [4, 6]. The model includes two eardrum modes of vibration, a shunt for flexible coupling between the incus and stapes, and a single tuned oscillator for ossicular vibration. The contribution of each of these mechanisms was examined by fitting a model of the ear to acoustic input impedance data from healthy human ears. The circuit elements for a non-ossicular eardrum vibration and a flexible coupling between the incus and stapes were found to be detrimental or non-essential for the model-fit-to-data. A single mode of eardrum vibration for sound transmission to the middle ear is consistent with the eardrum acting as an impedance-matching device, with pars-tensa eardrum vibration coupled to the ossicles [1]. A single-tuned oscillator was insufficient to account for the bandwidth of the ear. The frequency response of the ear suggests multiple resonant modes of ossicular vibration.
The Dualistic Model of Ovarian Carcinogenesis: Revisited, Revised, and Expanded.
Kurman, Robert J; Shih, Ie-Ming
2016-04-01
Since our proposal of a dualistic model of epithelial ovarian carcinogenesis more than a decade ago, a large number of molecular and histopathologic studies were published that have provided important insights into the origin and molecular pathogenesis of this disease. This has required that the original model be revised and expanded to incorporate these findings. The new model divides type I tumors into three groups: i)Â endometriosis-related tumors that include endometrioid, clear cell, and seromucinous carcinomas; ii) low-grade serous carcinomas; and iii) mucinous carcinomas and malignant Brenner tumors. As in the previous model, type II tumors are composed, for the most part, of high-grade serous carcinomas that can be further subdivided into morphologic and molecular subtypes. Type I tumors develop from benign extraovarian lesions that implant on the ovary and which can subsequently undergo malignant transformation, whereas many type II carcinomas develop from intraepithelial carcinomas in the fallopian tube and, as a result, disseminate as carcinomas that involve the ovary and extraovarian sites, which probably accounts for their clinically aggressive behavior. The new molecular genetic data, especially those derived from next-generation sequencing, further underline the heterogeneity of ovarian cancer and identify actionable mutations. The dualistic model highlights these differences between type I and type II tumors which, it can be argued, describe entirely different groups of diseases. PMID:27012190
The hydrogeologic-geochemical model of Cerro Prieto revisited
Lippmann, M.J.; Halfman, S.E.; Truesdell, A.H.; Manon M., A.
1989-01-01
As the exploitation of the Cerro Prieto, Mexico, geothermal field continues, there is increasing evidence that the hydrogeologic model developed by Halfman et al. (1984, 1986) presents the basic features controlling the movement of geothermal fluids in the system. At the present time the total installed capacity at Cerro Prieto is 620 MWe requiring the production of more than 10,500 tonnes/hr of a brine-steam mixture. This significant rate of fluid production has resulted in changes in reservoir thermodynamic conditions and in the chemistry of the produced fluids. After reviewing the hydrogeologic-geochemical model of Cerro Prieto, some of the changes observed in the field due to its exploitation are discussed and interpreted on the basis of the model. 21 refs., 11 figs., 1 tab.
Energy-economy interactions revisited within a comprehensive sectoral model
Hanson, D. A.; Laitner, J. A.
2000-07-24
This paper describes a computable general equilibrium (CGE) model with considerable sector and technology detail, the ``All Modular Industry Growth Assessment'' Model (AMIGA). It is argued that a detailed model is important to capture and understand the several rolls that energy plays within the economy. Fundamental consumer and industrial demands are for the services from energy; hence, energy demand is a derived demand based on the need for heating, cooling mechanical, electrical, and transportation services. Technologies that provide energy-services more efficiently (on a life cycle basis), when adopted, result in increased future output of the economy and higher paths of household consumption. The AMIGA model can examine the effects on energy use and economic output of increases in energy prices (e.g., a carbon charge) and other incentive-based policies or energy-efficiency programs. Energy sectors and sub-sector activities included in the model involve energy extraction conversion and transportation. There are business opportunities to produce energy-efficient goods (i.e., appliances, control systems, buildings, automobiles, clean electricity). These activities are represented in the model by characterizing their likely production processes (e.g., lighter weight motor vehicles). Also, multiple industrial processes can produce the same output but with different technologies and inputs. Secondary recovery, i.e., recycling processes, are examples of these multiple processes. Combined heat and power (CHP) is also represented for energy-intensive industries. Other modules represent residential and commercial building technologies to supply energy services. All sectors of the economy command real resources (capital services and labor).
The Thirring-Wess model revisited: a functional integral approach
Belvedere, L.V. . E-mail: armflavio@if.uff.br
2005-06-01
We consider the Wess-Zumino-Witten theory to obtain the functional integral bosonization of the Thirring-Wess model with an arbitrary regularization parameter. Proceeding a systematic of decomposing the Bose field algebra into gauge-invariant- and gauge-non-invariant field subalgebras, we obtain the local decoupled quantum action. The generalized operator solutions for the equations of motion are reconstructed from the functional integral formalism. The isomorphism between the QED {sub 2} (QCD {sub 2}) with broken gauge symmetry by a regularization prescription and the Abelian (non-Abelian) Thirring-Wess model with a fixed bare mass for the meson field is established.
What drives health care expenditure?--Baumol's model of 'unbalanced growth' revisited.
Hartwig, Jochen
2008-05-01
The share of health care expenditure in GDP rises rapidly in virtually all OECD countries, causing increasing concern among politicians and the general public. Yet, economists have to date failed to reach an agreement on what the main determinants of this development are. This paper revisits Baumol's [Baumol, W.J., 1967. Macroeconomics of unbalanced growth: the anatomy of urban crisis. American Economic Review 57 (3), 415-426] model of 'unbalanced growth', showing that the latter offers a ready explanation for the observed inexorable rise in health care expenditure. The main implication of Baumol's model in this context is that health care expenditure is driven by wage increases in excess of productivity growth. This hypothesis is tested empirically using data from a panel of 19 OECD countries. Our tests yield robust evidence in favor of Baumol's theory. PMID:18164773
Revisiting competition in a classic model system using formal links between theory and data.
Hart, Simon P; Burgin, Jacqueline R; Marshall, Dustin J
2012-09-01
Formal links between theory and data are a critical goal for ecology. However, while our current understanding of competition provides the foundation for solving many derived ecological problems, this understanding is fractured because competition theory and data are rarely unified. Conclusions from seminal studies in space-limited benthic marine systems, in particular, have been very influential for our general understanding of competition, but rely on traditional empirical methods with limited inferential power and compatibility with theory. Here we explicitly link mathematical theory with experimental field data to provide a more sophisticated understanding of competition in this classic model system. In contrast to predictions from conceptual models, our estimates of competition coefficients show that a dominant space competitor can be equally affected by interspecific competition with a poor competitor (traditionally defined) as it is by intraspecific competition. More generally, the often-invoked competitive hierarchies and intransitivities in this system might be usefully revisited using more sophisticated empirical and analytical approaches. PMID:23094373
MRAC Revisited: Guaranteed Performance with Reference Model Modification
NASA Technical Reports Server (NTRS)
Stepanyan, Vahram; Krishnakumar, Kalmaje
2010-01-01
This paper presents modification of the conventional model reference adaptive control (MRAC) architecture in order to achieve guaranteed transient performance both in the output and input signals of an uncertain system. The proposed modification is based on the tracking error feedback to the reference model. It is shown that approach guarantees tracking of a given command and the ideal control signal (one that would be designed if the system were known) not only asymptotically but also in transient by a proper selection of the error feedback gain. The method prevents generation of high frequency oscillations that are unavoidable in conventional MRAC systems for large adaptation rates. The provided design guideline makes it possible to track a reference command of any magnitude form any initial position without re-tuning. The benefits of the method are demonstrated in simulations.
Laws of reflection and Snell's law revisited by video modeling
NASA Astrophysics Data System (ADS)
Rodrigues, M.; Simeão Carvalho, P.
2014-07-01
Video modelling is being used, nowadays, as a tool for teaching and learning several topics in Physics. Most of these topics are related to kinematics. In this work we show how video modelling can be used for demonstrations and experimental teaching in optics, namely the laws of reflection and the well-known Snell's Law of light. Videos were recorded with a photo camera at 30 frames/s, and analysed with the open source software Tracker. Data collected from several frames was treated with the Data Tool module, and graphs were built to obtain relations between incident, reflected and refraction angles, as well as to determine the refractive index of Perspex. These videos can be freely distributed in the web and explored with students within the classroom, or as a homework assignment to improve student's understanding on specific contents. They present a large didactic potential for teaching basic optics in high school with an interactive methodology.
NEW SOLAR COMPOSITION: THE PROBLEM WITH SOLAR MODELS REVISITED
Serenelli, Aldo M.; Asplund, Martin; Basu, Sarbani; Ferguson, Jason W.
2009-11-10
We construct updated solar models with different sets of solar abundances, including the most recent determinations by Asplund et al. The latter work predicts a larger (approx10%) solar metallicity compared to previous measurements by the same authors but significantly lower (approx25%) than the recommended value from a decade ago by Grevesse and Sauval. We compare the results of our models with determinations of the solar structure inferred through helioseismology measurements. The model that uses the most recent solar abundance determinations predicts the base of the solar convective envelope to be located at R {sub CZ} = 0.724 R{sub sun} and a surface helium mass fraction of Y{sub surf} = 0.231. These results are in conflict with helioseismology data (R{sub CZ} = 0.713 +- 0.001 R{sub sun} and Y{sub surf} = 0.2485 +- 0.0035) at 5sigma and 11sigma levels, respectively. Using the new solar abundances, we calculate the magnitude by which radiative opacities should be modified in order to restore agreement with helioseismology. We find that a maximum change of approx15% at the base of the convective zone is required with a smooth decrease toward the core, where the change needed is approx5%. The required change at the base of the convective envelope is about half the value estimated previously. We also present the solar neutrino fluxes predicted by the new models. The most important changes brought about by the new solar abundances are the increase by approx10% in the predicted {sup 13}N and {sup 15}O fluxes that arise mostly due to the increase in the C and N abundances in the newly determined solar composition.
A revisit to the one form kinetic model of prothrombinase.
Lee, Chang Jun; Wu, Sangwook; Eun, Changsun; Pedersen, Lee G
2010-06-01
Thrombin is generated enzymatically from prothrombin by two pathways with the intermediates of meizothrombin and prethrombin-2. Experimental concentration profiles from two independent groups for these two pathways have been re-analyzed. By rationally combining the independent data sets, a simple mechanism can be established and rate constants determined. A structural model is consistent with the data-derived finding that mechanisms that feature channeling or ratcheting are not necessary to describe thrombin production. PMID:20435402
The comfortable driving model revisited: traffic phases and phase transitions
NASA Astrophysics Data System (ADS)
Knorr, Florian; Schreckenberg, Michael
2013-07-01
We study the spatiotemporal patterns resulting from different boundary conditions for a microscopic traffic model and contrast them with empirical results. By evaluating the time series of local measurements, the local traffic states are assigned to the different traffic phases of Kerner’s three-phase traffic theory. For this classification we use the rule-based FOTO-method, which provides ‘hard’ rules for this assignment. Using this approach, our analysis shows that the model is indeed able to reproduce three qualitatively different traffic phases: free flow (F), synchronized traffic (S), and wide moving jams (J). In addition, we investigate the likelihood of transitions between the three traffic phases. We show that a transition from free flow to a wide moving jam often involves an intermediate transition: first from free flow to synchronized flow and then from synchronized flow to a wide moving jam. This is supported by the fact that the so-called F ? S transition (from free flow to synchronized traffic) is much more likely than a direct F ? J transition. The model under consideration has a functional relationship between traffic flow and traffic density. The fundamental hypothesis of the three-phase traffic theory, however, postulates that the steady states of synchronized flow occupy a two-dimensional region in the flow-density plane. Due to the obvious discrepancy between the model investigated here and the postulate of the three-phase traffic theory, the good agreement that we found could not be expected. For a more detailed analysis, we also studied vehicle dynamics at a microscopic level and provide a comparison of real detector data with simulated data of the identical highway segment.
Casimir force in noncommutative Randall-Sundrum models revisited
Teo, L. P.
2010-07-15
We propose another method to compute the Casimir force in noncommutative Randall-Sundrum braneworld model considered by K. Nouicer and Y. Sabri, Phys. Rev. D 80, 086013 (2009). recently. Our method can be used to compute the Casimir force to any order in the noncommutative parameter. Contrary to the claim made by K. Nouicer and Y. Sabri that repulsive Casimir force can appear in the first order approximation, we show that the Casimir force is always attractive at any order of approximation.
Allele frequencies at microsatellite loci: The stepwise mutation model revisited
Valdes, A.M.; Slatkin, M. ); Freimer, N.B. )
1993-03-01
The authors summarize available data on the frequencies of alleles at microsatellite loci in human populations and compare observed distributions of allele frequencies to those generated by a simulation of the stepwise mutation model. They show that observed frequency distributions at 108 loci are consistent with the results of the model under the assumption that mutations cause an increase or decrease in repeat number by one and under the condition that the product Nu, where N is the effective population size and u is the mutation rate, is larger than one. It is also shown that the variance of the distribution of allele sizes is a useful estimator of Nu and performs much better than previously suggested estimators for the stepwise mutation model. In the data, there is no correlation between the mean and variance in allele size at a locus or between the number of alleles and mean allele size, which suggests that the mutation rate at these loci is independent of allele size. 39 refs., 6 figs., 4 tabs.
The Distributed D-Clean Model Revisited by Templates
NASA Astrophysics Data System (ADS)
Zsók, Viktória; Porkoláb, Zoltán
2011-09-01
D-Clean is a functional coordination language for distributed computation. The language was designed for the need of high-level process description and communication coordination of functional programs distributed over a cluster. The pure functional computational nodes required language primitives to control the dataflow in a distributed process-network. Therefore, in order to achieve parallel features, we created an extension for the lazy functional programming language Clean using new language elements. D-Clean is compiled to an intermediate level language called D-Box, which is designed for the description of the computational nodes. Every D-Clean construct generates a D-Box expression. The D-Box expressions hide the low level implementation details and enable direct control over the process-network. The asynchronous communication is based on language-independent middleware services. Earlier we have presented the syntax and the semantics of both coordination languages. Practical experiences showed the difficulties of distributed program development, especially in testing and debugging. This paper aims to provide software comprehension application for a better understanding and usage of the D-Clean distributed system. We model the elements and the behaviour of the D-Clean system using C++ templates. The strong type system of C++ templates guarantees the correctness of the model. Using templates we can avoid run-time overhead achieving impressive efficiency.
The Zipf Law revisited: An evolutionary model of emerging classification
Levitin, L.B.; Schapiro, B.; Perlovsky, L.
1996-12-31
Zipf`s Law is a remarkable rank-frequency relationship observed in linguistics (the frequencies of the use of words are approximately inversely proportional to their ranks in the decreasing frequency order) as well as in the behavior of many complex systems of surprisingly different nature. We suggest an evolutionary model of emerging classification of objects into classes corresponding to concepts and denoted by words. The evolution of the system is derived from two basic assumptions: first, the probability to recognize an object as belonging to a known class is proportional to the number of objects in this class already recognized, and, second, there exists a small probability to observe an object that requires creation of a new class ({open_quotes}mutation{close_quotes} that gives birth to a new {open_quotes}species{close_quotes}). It is shown that the populations of classes in such a system obey the Zipf Law provided that the rate of emergence of new classes is small. The model leads also to the emergence of a second-tier structure of {open_quotes}super-classes{close_quotes} - groups of classes with almost equal populations.
Tail Lobe Revisited: Magnetic Field Modeling Based on Plasma Data
NASA Technical Reports Server (NTRS)
Karlsson, S. B. P.; Tsyganenko, N. A.
1999-01-01
Plasma data from the ISEE-1 and -2 spacecraft during 1977-1980 have been used to determine the distribution of data points in the magnetotail in the range of distances -20 < XGSM < --15, i.e. which of the records that were located in the current sheet, in the tail lobe, in the magnetosheath and in the boundary layers respectively. The ISEE-1 and -2 magnetic field data for the records in the tail lobe were then used to model the tail lobe magnetic field dependence on the solar wind dynamic pressure, on the Interplanetary Magnetic Field (IMF) and on the Dst index. The tail lobe magnetic field was assumed to be dependent on the square root of the dynamic pressure based on the balance between the total magnetic pressure in the tail lobes and the dynamic pressure of the solar wind. The IMF dependent terms, added to the pressure term, were sought in many different forms while the Dst dependence of the tail lobe magnetic field was assumed to be linear. The field shows a strong dependence on the square root of the dynamic pressure and the different IMF dependent terms all constitute a significant contribution to the total field. However, the dependence on the Dst index turned out to be very weak at those down-tail distances. The results of this study are intended to be used for parameterizing future versions of the data-based models of the global magnetospheric magnetic field.
Murray's law revisited: Quémada's fluid model and fractal trees
NASA Astrophysics Data System (ADS)
Moreau, Baptiste; Mauroy, Benjamin
2015-11-01
In 1926, Murray proposed the first law for the optimal design of blood vessels. He minimized the power dissipation arising from the trade-off between fluid circulation and blood maintenance. The law, based on a constant fluid viscosity, states that in the optimal configuration the fluid flow rate inside the vessel is proportional to the cube of the vessel radius, implying that wall shear stress is not dependent on the vessel radius. Murray's law has been found to be true in blood macrocirculation, but not in microcirculation. In 2005, Alarc\\'on et al took into account the non monotonous dependence of viscosity on vessel radius - F{\\aa}hr{\\ae}us-Lindqvist effect - due to phase separation effect of blood. They were able to predict correctly the behavior of wall shear stresses in microcirculation. One last crucial step remains however: to account for the dependence of blood viscosity on shear rates. In this work, we investigate how viscosity dependence on shear rate affects Murray's law. We extended Murray's optimal design to the whole range of Qu\\'emada's fluids, that models pseudo-plastic fluids such as blood. Our study shows that Murray's original law is not restricted to Newtonian fluids, it is actually universal for all Qu\\'emada's fluid as long as there is no phase separation effect. When phase separation effect occurs, then we derive an extended version of Murray's law. Our analyses are very general and apply to most of fluids with shear dependent rheology. Finally, we study how these extended laws affect the optimal geometries of fractal trees to mimic an idealized arterial network.
Yukawa Meson, Sakata Model and Baryon-Lepton Symmetry Revisited
NASA Astrophysics Data System (ADS)
Marshak, R. E.
It is difficult for me to grasp that this symposium is celebrating the jubilee of meson theory since I was a junior at Columbia College in 1935. I recall hearing a colloquium by Paul Dirac that year telling an enraptured audience about the infinite sea of negative energy states but I do not recall any special note being taken of the birth of an equally revolutionary concept, the Yukawa meson. Perhaps the reason was the publication of Hideki Yukawa's paper in an inaccessible Japanese journal, perhaps Dirac's electron theory was dealing with the well-known electromagnetic force whereas Yukawa' meson theory was put forth to understand the nature of two new forces - the nuclear and the weak. Whatever the reason, the situation changed drastically when I migrated to Cornell (to do my thesis under Hans Bethe during the years 1937sim39) and found a deep interest in meson theory. Thus, my own scientific career has almost spanned the period since the birth of meson theory but, what is more to the point, it has been strongly influenced by the work of Yukawa and his collaborators. It therefore gives me great pleasure to be able to talk at this MESON 50 symposium. As one of the oldest speakers, I shall respond in a loose way to Professor Maki's invitation to cover ``topics concerning the historical developments of hadron physics''. I shall select several major themes from the Japanese work that have had special interest for me. My remarks will fall under the four headings: (A) Yukawa Meson; (B) Sakata Model; (C) Baryon-Lepton Symmetry; and (D) Extensions of Baryon-Lepton Symmetry.
Revisiting food-based models of territoriality in solitary predators.
López-Bao, José V; Rodríguez, Alejandro; Delibes, Miguel; Fedriani, José M; Calzada, Javier; Ferreras, Pablo; Palomares, Francisco
2014-07-01
Food availability is considered a major factor determining spacing behaviour in territorial species, especially for females. Theoretically, spatial overlap (considered the opposite of territoriality) and food availability are related in a nonlinear manner (hypothesized inverted-U function), with high overlap levels at the extremes of a food availability gradient and low overlap at intermediate levels of this gradient. Similar patterns are expected for encounter frequencies owing to its expected correlation with spatial overlap. However, these predictions have rarely been tested in highly structured social systems on a broad gradient of food availability, which implicitly requires experimental manipulation. We test these predictions in a solitary, territorial and trophic specialist, the Iberian lynx Lynx pardinus, taking advantage of a three-decade data set of spatial behaviour in different scenarios of food availability (i.e. rabbit density). In contrast with expectations, home range overlap among resident females was low (median overlap index = 0.08, range 0-0.57) and core area overlap was nearly nil (median overlap index = 0, range 0-0.22) throughout the entire gradient of prey availability. Furthermore, spatial associations between pairs of females were negligible regardless marked variation in prey availability. Therefore, we did not find support for a model of flexible lynx territoriality driven by food availability. Our results suggest that the exclusive use of space in the Iberian lynx was not related to food. Lack of influence of prey availability on lynx territoriality may be adaptive to cope with the consequences of frequent drought-induced periods of prey scarcity or other disturbance typically affecting wild rabbit populations in Mediterranean environments. Thus, lynx would adopt an obstinate strategy of territoriality that consists in defending exclusive areas across a broad range of resource availability ensuring an exclusive access to the minimum amount of prey necessary for survival and eventually reproduction even during periods of prey scarcity. However, we found signs that territoriality was influenced by lynx density in a nonlinear fashion. Our results suggest the occurrence of population regulation through territoriality in this species. PMID:24720673
The Eating Attitudes Test-26 revisited using exploratory structural equation modeling.
MaÃ¯ano, Christophe; Morin, Alexandre J S; Lanfranchi, Marie-Christine; Therme, Pierre
2013-07-01
Most previous studies have failed to replicate the original factor structure of the 26-item version of the Eating Attitudes Test (EAT-26) among community samples of adolescents. The main objective of the present series of four studies (nâ€‰=â€‰2178) was to revisit the factor structure of this instrument among mixed gender community samples of adolescents using both exploratory structural equation modeling (ESEM) and confirmatory factor analysis (CFA). First, results from the ESEM analyses provided satisfactory goodness-of-fit statistics and reliability coefficients for a six-factor model of the EAT with 18 items (EAT-18) closely corresponding to the original seven-factor structure proposed for the 40-item version of the EAT. Second, these analyses were satisfactorily replicated among a new sample of community adolescents using CFA. The results confirmed the factor loading and intercept invariance of this model across gender and age groups (i.e., early and late adolescence), as well as the complete invariance of the EAT-18 measurement model between ethnicities (i.e., European versus African origins) and across weight categories (i.e., underweight, normal weight and overweight). Finally, the last study provided support for convergent validity of the EAT-18 with the Eating Disorder Inventory and with instruments measuring global self-esteem, physical appearance, social physique anxiety and fear of negative appearance evaluation. PMID:23344702
NASA Astrophysics Data System (ADS)
Vattré, A.; Devincre, B.; Feyel, F.; Gatti, R.; Groh, S.; Jamond, O.; Roos, A.
2014-02-01
A unified model coupling 3D dislocation dynamics (DD) simulations with the finite element (FE) method is revisited. The so-called Discrete-Continuous Model (DCM) aims to predict plastic flow at the (sub-)micron length scale of materials with complex boundary conditions. The evolution of the dislocation microstructure and the short-range dislocation-dislocation interactions are calculated with a DD code. The long-range mechanical fields due to the dislocations are calculated by a FE code, taking into account the boundary conditions. The coupling procedure is based on eigenstrain theory, and the precise manner in which the plastic slip, i.e. the dislocation glide as calculated by the DD code, is transferred to the integration points of the FE mesh is described in full detail. Several test cases are presented, and the DCM is applied to plastic flow in a single-crystal Nickel-based superalloy.
Revisiting the Slichter mode of Mercury in the context of differentiated models
NASA Astrophysics Data System (ADS)
Escapa, A.; Fukushima, T.
2015-11-01
We revisit the Slichter mode of a set of differentiated Mercury models that were considered previously. We apply a different and independent formalism, based on variational principles of mechanics, to analytically model the internal translations of a body with a global internal perfect fluid layer limited by two differentiated rigid solid constituents that have a spherical symmetric mass distribution. The Slichter mode is specified by providing the time evolution of the barycenters of the solid constituents, the amplitude ratio of the mantle to the inner core, and the period of the oscillatory motion. All these parameters only depend on the mass of the body, the mass of the inner core, the density of the fluid, and the mean density of the inner core. For previously developed Mercury models we find, in contrast to other results, that there are no discontinuities in the Slichter period when passing from undifferentiated to differentiated inner cores. Hence, in a general situation, a potential detection of the Slichter mode cannot determine the differentiation of the inner core. We also find that the BepiColombo mission may be able to detect the Slichter mode caused by an impactor meteroid with a mass of about 1012 kg in the most favorable of circumstances. This shows that a measurable excitation of the mode by this mechanism is even more unlikely than was previously established, where a mass about 1010 kg was found. Appendices are available in electronic form at http://www.aanda.org
NASA Astrophysics Data System (ADS)
Guterl, Jerome; Smirnov, R. D.; Krasheninnikov, S. I.
2015-07-01
Desorption phase of thermal desorption spectroscopy (TDS) experiments performed on tungsten samples exposed to flux of hydrogen isotopes in fusion relevant conditions is analyzed using a reaction-diffusion model describing hydrogen retention in material bulk. Two regimes of hydrogen desorption are identified depending on whether hydrogen trapping rate is faster than hydrogen diffusion rate in material during TDS experiments. In both regimes, a majority of hydrogen released from material defects is immediately outgassed instead of diffusing deeply in material bulk when the evolution of hydrogen concentration in material is quasi-static, which is the case during TDS experiments performed with tungsten samples exposed to flux of hydrogen isotopes in fusion related conditions. In this context, analytical expressions of the hydrogen outgassing flux as a function of the material temperature are obtained with sufficient accuracy to describe main features of thermal desorption spectra (TDSP). These expressions are then used to highlight how characteristic temperatures of TDSP depend on hydrogen retention parameters, such as trap concentration or activation energy of detrapping processes. The use of Arrhenius plots to characterize retention processes is then revisited when hydrogen trapping takes place during TDS experiments. Retention processes are also characterized using the shape of desorption peaks in TDSP, and it is shown that diffusion of hydrogen in material during TDS experiment can induce long desorption tails visible aside desorption peaks at high temperature in TDSP. These desorption tails can be used to estimate activation energy of diffusion of hydrogen in material.
One-zone synchrotron self-Compton model for the core emission of Centaurus A revisited
NASA Astrophysics Data System (ADS)
Petropoulou, M.; Lefa, E.; Dimitrakoudis, S.; Mastichiadis, A.
2014-02-01
Aims: We investigate the role of the second synchrotron self-Compton (SSC) photon generation to the multiwavelength emission from the compact regions of sources that are characterized as misaligned blazars. For this, we focus on the nearest high-energy emitting radio galaxy Centaurus A and we revisit the one-zone SSC model for its core emission. Methods: We have calculated analytically the peak luminosities of the first and second SSC components by first deriving the steady-state electron distribution in the presence of synchrotron and SSC cooling, and then by using appropriate expressions for the positions of the spectral peaks. We have also tested our analytical results against those derived from a numerical code where the full emissivities and cross-sections were used. Results: We show that the one-zone SSC model cannot account for the core emission of Centaurus A above a few GeV, where the peak of the second SSC component appears. We thus propose an alternative explanation for the origin of the high-energy (?0.4 GeV) and TeV emission, where these are attributed to the radiation emitted by a relativistic proton component through photohadronic interactions with the photons produced by the primary leptonic component. We show that the required proton luminosities are not extremely high, i.e. ~1043 erg/s, provided that the injection spectra are modelled by a power law with a high value of the lower energy cutoff. Finally, we find that the contribution of the core emitting region of Cen A to the observed neutrino and ultra-high-energy cosmic-ray fluxes is negligible.
Revisiting the monopole components of effective interactions for the shell model
NASA Astrophysics Data System (ADS)
Wang, X. B.; Dong, G. X.
2015-12-01
In this paper, we revisit the monopole components of effective interactions for the shell model. Without going through specific nuclei or shell gaps, universal roles of central, tensor, and spin–orbit forces can be proved, reflecting the intrinsic features of shell model effective interactions. For monopole matrix elements, even and odd channels of central force often have a canceling effect. However, for the contributions to the shell evolution, its even and odd channels could have both positive or negative contributions, enhancing the role of central force on the shell structure. Tensor force is generally weaker than central force. However, for the effect on shell evolutions, tensor force can dominate or play a competitive role. A different systematics has been discovered between T = 1 and 0 channels. For example, tensor force, well established in the T = 0 channel, becomes uncertain in the T = 1 channel. We calculate the properties of neutron-rich oxygen and calcium isotopes in order to study T = 1 channel interactions further. It is learned that the main improvements of empirical interactions are traced to the central force. For non-central forces, antisymmetric spin–orbit (ALS) force, originated from many-body perturbations or three-body force, could also play an explicit role. T = 1 tensor forces are less constrained so their effect can differ in different empirical interactions. The influence of tensor force may sometimes be canceled by many-body effects. For T = 0 channels of effective interactions, which is the main source of neutron–proton correlations, central and tensor forces are the leading components. For T = 1 channels, which can act between like-particles, the request for many-body correlations could be more demanding, so that the monopole anomaly of the T = 1 channel might be more serious.
Making Productive Use of Four Models of School English: A Case Study Revisited
ERIC Educational Resources Information Center
Macken-Horarik, Mary
2014-01-01
At a time when political leaders and media pundits seek to narrow the English curriculum and reduce its knowledge structure to the "basics," it is helpful to revisit the potential of different approaches to learning in English that have evolved over time. In this paper I reflect on the semantic features of personal growth, cultural…
Goh, Chin Heng; Kathiresan, Purushothaman; Németh, Sándor; Jeney, Zsigmond; Bercsényi, Miklós; Orbán, László
2013-01-01
The body of most fishes is fully covered by scales that typically form tight, partially overlapping rows. While some of the genes controlling the formation and growth of fish scales have been studied, very little is known about the genetic mechanisms regulating scale pattern formation. Although the existence of two genes with two pairs of alleles (S&s and N&n) regulating scale coverage in cyprinids has been predicted by Kirpichnikov and colleagues nearly eighty years ago, their identity was unknown until recently. In 2009, the ‘S’ gene was found to be a paralog of fibroblast growth factor receptor 1, fgfr1a1, while the second gene called ‘N’ has not yet been identified. We re-visited the original model of Kirpichnikov that proposed four major scale pattern types and observed a high degree of variation within the so-called scattered phenotype due to which this group was divided into two sub-types: classical mirror and irregular. We also analyzed the survival rates of offspring groups and found a distinct difference between Asian and European crosses. Whereas nude × nude crosses involving at least one parent of Asian origin or hybrid with Asian parent(s) showed the 25% early lethality predicted by Kirpichnikov (due to the lethality of the NN genotype), those with two Hungarian nude parents did not. We further extended Kirpichnikov's work by correlating changes in phenotype (scale-pattern) to the deformations of fins and losses of pharyngeal teeth. We observed phenotypic changes which were not restricted to nudes, as described by Kirpichnikov, but were also present in mirrors (and presumably in linears as well; not analyzed in detail here). We propose that the gradation of phenotypes observed within the scattered group is caused by a gradually decreasing level of signaling (a dose-dependent effect) probably due to a concerted action of multiple pathways involved in scale formation. PMID:24386179
Casas, Laura; Sz?cs, Réka; Vij, Shubha; Goh, Chin Heng; Kathiresan, Purushothaman; Németh, Sándor; Jeney, Zsigmond; Bercsényi, Miklós; Orbán, László
2013-01-01
The body of most fishes is fully covered by scales that typically form tight, partially overlapping rows. While some of the genes controlling the formation and growth of fish scales have been studied, very little is known about the genetic mechanisms regulating scale pattern formation. Although the existence of two genes with two pairs of alleles (S&s and N&n) regulating scale coverage in cyprinids has been predicted by Kirpichnikov and colleagues nearly eighty years ago, their identity was unknown until recently. In 2009, the 'S' gene was found to be a paralog of fibroblast growth factor receptor 1, fgfr1a1, while the second gene called 'N' has not yet been identified. We re-visited the original model of Kirpichnikov that proposed four major scale pattern types and observed a high degree of variation within the so-called scattered phenotype due to which this group was divided into two sub-types: classical mirror and irregular. We also analyzed the survival rates of offspring groups and found a distinct difference between Asian and European crosses. Whereas nude × nude crosses involving at least one parent of Asian origin or hybrid with Asian parent(s) showed the 25% early lethality predicted by Kirpichnikov (due to the lethality of the NN genotype), those with two Hungarian nude parents did not. We further extended Kirpichnikov's work by correlating changes in phenotype (scale-pattern) to the deformations of fins and losses of pharyngeal teeth. We observed phenotypic changes which were not restricted to nudes, as described by Kirpichnikov, but were also present in mirrors (and presumably in linears as well; not analyzed in detail here). We propose that the gradation of phenotypes observed within the scattered group is caused by a gradually decreasing level of signaling (a dose-dependent effect) probably due to a concerted action of multiple pathways involved in scale formation. PMID:24386179
ERIC Educational Resources Information Center
Court, Deborah
1999-01-01
Revisits and reviews Imre Lakatos' ideas on "Falsification and the Methodology of Scientific Research Programmes." Suggests that Lakatos' framework offers an insightful way of looking at the relationship between theory and research that is relevant not only for evaluating research programs in theoretical physics, but in the social sciences as…
Basic, Ivan; Nadramija, Damir; Flajslik, Mario; Amic, Dragan; Lucic, Bono
2007-12-26
Several quantitative structure-activity studies for this data set containing 107 HEPT derivatives have been performed since 1997, using the same set of molecules by (more or less) different classes of molecular descriptors. Multivariate Regression (MR) and Artificial Neural Network (ANN) models were developed and in each study the authors concluded that ANN models are superior to MR ones. We re-calculated multivariate regression models for this set of molecules using the same set of descriptors, and compared our results with the previous ones. Two main reasons for overestimation of the quality of the ANN models in previous studies comparing with MR models are: (1) wrong calculation of leave-one-out (LOO) cross-validated (CV) correlation coefficient for MR models in Luco et al., J. Chem. Inf. Comput. Sci. 37 392-401 (1997), and (2) incorrect estimation/interpretation of leave-one-out (LOO) cross-validated and predictive performance and power of ANN models. More precise and fairer comparison of fit and LOO CV statistical parameters shows that MR models are more stable. In addition, MR models are much simpler than ANN ones. For real testing the predictive performance of both classes of models we need more HEPT derivatives, because all ANN models that presented results for external set of molecules used experimental values in optimization of modeling procedure and model parameters.
Borza, Liana Rada; Gavrilovici, Cristina; Stockman, René
2015-01-01
The present paper revisits the ethical models of patient--physician relationship from the perspective of patient autonomy and values. It seems that the four traditional models of physician--patient relationship proposed by Emanuel & Emanuel in 1992 closely link patient values and patient autonomy. On the other hand, their reinterpretation provided by Agarwal & Murinson twenty years later emphasizes the independent expression of values and autonomy in individual patients. Additionally, patient education has been assumed to join patient values and patient autonomy. Moreover, several authors have noted that, over the past few decades, patient autonomy has gradually replaced the paternalistic approach based on the premise that the physician knows what is best for the patient. Neither the paternalistic model of physician-patient relationship, nor the informative model is considered to be satisfactory, as the paternalistic model excludes patient values from decision making, while the informative model excludes physician values from decision making. However, the deliberative model of patient-physician interaction represents an adequate alternative to the two unsatisfactory approaches by promoting shared decision making between the physician and the patient. It has also been suggested that the deliberative model would be ideal for exercising patient autonomy in chronic care and that the ethical role of patient education would be to make the deliberative model applicable to chronic care. In this regard, studies have indicated that the use of decision support interventions might increase the deliberative capacity of chronic patients. PMID:26204658
NASA Astrophysics Data System (ADS)
McGurk, B. J.; Painter, T. H.
2014-12-01
Deterministic snow accumulation and ablation simulation models are widely used by runoff managers throughout the world to predict runoff quantities and timing. Model fitting is typically based on matching modeled runoff volumes and timing with observed flow time series at a few points in the basin. In recent decades, sparse networks of point measurements of the mountain snowpacks have been available to compare with modeled snowpack, but the comparability of results from a snow sensor or course to model polygons of 5 to 50 sq. km is suspect. However, snowpack extent, depth, and derived snow water equivalent have been produced by the NASA/JPL Airborne Snow Observatory (ASO) mission for spring of 20013 and 2014 in the Tuolumne River basin above Hetch Hetchy Reservoir. These high-resolution snowpack data have exposed the weakness in a model calibration based on runoff alone. The U.S. Geological Survey's Precipitation Runoff Modeling System (PRMS) calibration that was based on 30-years of inflow to Hetch Hetchy produces reasonable inflow results, but modeled spatial snowpack location and water quantity diverged significantly from the weekly measurements made by ASO during the two ablation seasons. The reason is that the PRMS model has many flow paths, storages, and water transfer equations, and a calibrated outflow time series can be right for many wrong reasons. The addition of a detailed knowledge of snow extent and water content constrains the model so that it is a better representation of the actual watershed hydrology. The mechanics of recalibrating PRMS to the ASO measurements will be described, and comparisons in observed versus modeled flow for both a small subbasin and the entire Hetch Hetchy basin will be shown. The recalibrated model provided a bitter fit to the snowmelt recession, a key factor for water managers as they balance declining inflows with demand for power generation and ecosystem releases during the final months of snow melt runoff.
NASA Astrophysics Data System (ADS)
Saito, F.; Abe-Ouchi, A.; Takahashi, K.; Blatter, H.
2016-01-01
The present paper revisits the future surface-climate experiments on the Greenland ice sheet proposed by the Sea-level Response to Ice Sheet Evolution (SeaRISE; Bindschadler et al., 2013) study. The projections of the different SeaRISE participants show dispersion, which has not been examined in detail to date. A series of sensitivity experiments are conducted and analyzed using the ice-sheet model for integrated Earth-system studies (IcIES) by replacing one or more formulations of the model parameters with those adopted in other model(s). The results show that large potential sources of the dispersion among the projections of the different SeaRISE participants are differences in the initialization methods and in the surface mass balance methods, and both aspects have almost equal impact on the results. The treatment of ice-sheet margins in the simulation has a secondary impact on the dispersion. We conclude that spinning up the model using fixed topography through the spin-up period while the temperature is allowed to evolve according to the surface temperature history is the preferred representation, at least for the experiment configuration examined in the present paper. A benchmark model experimental setup that most of the numerical models can perform is proposed for future intercomparison projects, in order to evaluate the uncertainties relating to pure ice-sheet model flow characteristics.
Benzophenone Ultrafast Triplet Population: Revisiting the Kinetic Model by Surface-Hopping Dynamics
2016-01-01
The photochemistry of benzophenone, a paradigmatic organic molecule for photosensitization, was investigated by means of surface-hopping ab initio molecular dynamics. Different mechanisms were found to be relevant within the first 600 fs after excitation; the long-debated direct (S1 â†’ T1) and indirect (S1 â†’ T2 â†’ T1) mechanisms for population of the low-lying triplet state are both possible, with the latter being prevalent. Moreover, we established the existence of a kinetic equilibrium between the two triplet states, never observed before. This fact implies that a significant fraction of the overall population resides in T2, eventually allowing one to revisit the usual spectroscopic assignment proposed by transient absorption spectroscopy. This finding is of particular interest for photocatalysis as well as for DNA damages studies because both T1 and T2 channels are, in principle, available for benzophenone-mediated photoinduced energy transfer toward DNA. PMID:26821061
Benzophenone Ultrafast Triplet Population: Revisiting the Kinetic Model by Surface-Hopping Dynamics.
Marazzi, Marco; Mai, Sebastian; Roca-Sanjuán, Daniel; Delcey, Mickaël G; Lindh, Roland; González, Leticia; Monari, Antonio
2016-02-18
The photochemistry of benzophenone, a paradigmatic organic molecule for photosensitization, was investigated by means of surface-hopping ab initio molecular dynamics. Different mechanisms were found to be relevant within the first 600 fs after excitation; the long-debated direct (S1 ? T1) and indirect (S1 ? T2 ? T1) mechanisms for population of the low-lying triplet state are both possible, with the latter being prevalent. Moreover, we established the existence of a kinetic equilibrium between the two triplet states, never observed before. This fact implies that a significant fraction of the overall population resides in T2, eventually allowing one to revisit the usual spectroscopic assignment proposed by transient absorption spectroscopy. This finding is of particular interest for photocatalysis as well as for DNA damages studies because both T1 and T2 channels are, in principle, available for benzophenone-mediated photoinduced energy transfer toward DNA. PMID:26821061
Ries crater and suevite revisited—Observations and modeling Part II: Modeling
NASA Astrophysics Data System (ADS)
Artemieva, N. A.; Wünnemann, K.; Krien, F.; Reimold, W. U.; Stöffler, D.
2013-04-01
We present the results of numerical modeling of the formation of the Ries crater utilizing the two hydrocodes SOVA and iSALE. These standard models allow us to reproduce crater shape, size, and morphology, and composition and extension of the continuous ejecta blanket. Some of these results cannot, however, be readily reconciled with observations: the impact plume above the crater consists mainly of molten and vaporized sedimentary rocks, containing very little material in comparison with the ejecta curtain; at the end of the modification stage, the crater floor is covered by a thick layer of impact melt with a total volume of 6-11 km3; the thickness of true fallback material from the plume inside the crater does not exceed a couple of meters; ejecta from all stratigraphic units of the target are transported ballistically; no separation of sedimentary and crystalline rocks—as observed between suevites and Bunte Breccia at Ries—is noted. We also present numerical results quantifying the existing geological hypotheses of Ries ejecta emplacement from an impact plume, by melt flow, or by a pyroclastic density current. The results show that none of these mechanisms is consistent with physical constraints and/or observations. Finally, we suggest a new hypothesis of suevite formation and emplacement by postimpact interaction of hot impact melt with water or volatile-rich sedimentary rocks.
Revisiting a model of ontogenetic growth: estimating model parameters from theory and data.
Moses, Melanie E; Hou, Chen; Woodruff, William H; West, Geoffrey B; Nekola, Jeffery C; Zuo, Wenyun; Brown, James H
2008-05-01
The ontogenetic growth model (OGM) of West et al. provides a general description of how metabolic energy is allocated between production of new biomass and maintenance of existing biomass during ontogeny. Here, we reexamine the OGM, make some minor modifications and corrections, and further evaluate its ability to account for empirical variation on rates of metabolism and biomass in vertebrates both during ontogeny and across species of varying adult body size. We show that the updated version of the model is internally consistent and is consistent with other predictions of metabolic scaling theory and empirical data. The OGM predicts not only the near universal sigmoidal form of growth curves but also the M(1/4) scaling of the characteristic times of ontogenetic stages in addition to the curvilinear decline in growth efficiency described by Brody. Additionally, the OGM relates the M(3/4) scaling across adults of different species to the scaling of metabolic rate across ontogeny within species. In providing a simple, quantitative description of how energy is allocated to growth, the OGM calls attention to unexplained variation, unanswered questions, and opportunities for future research. PMID:18419571
Tibayrenc, Michel; Ayala, Francisco J
2015-11-01
Comparing the population structure of Trypanosoma cruzi with that of other pathogens, including parasitic protozoa, fungi, bacteria and viruses, shows that the agent of Chagas disease shares typical traits with many other species, related to a predominant clonal evolution (PCE) pattern: statistically significant linkage disequilibrium, overrepresented multilocus genotypes, near-clades (genetic subdivisions somewhat blurred by occasional genetic exchange/hybridization) and "Russian doll" patterns (PCE is observed, not only at the level of the whole species, but also, within the near-clades). Moreover, T. cruzi population structure exhibits linkage with the diversity of several strongly selected genes, with gene expression profiles, and with some major phenotypic traits. We discuss the evolutionary significance of these results, and their implications in terms of applied research (molecular epidemiology/strain typing, analysis of genes of interest, vaccine and drug design, immunological diagnosis) and of experimental evolution. Lastly, we revisit the long-term debate of describing new species within the T. cruzi taxon. PMID:26188332
Ups and Downs of Viagra: Revisiting Ototoxicity in the Mouse Model
Au, Adrian; Stuyt, John Gerka; Chen, Daniel; Alagramam, Kumar
2013-01-01
Sildenafil citrate (Viagra), a phosphodiesterase 5 inhibitor (PDE5i), is a commonly prescribed drug for erectile dysfunction. Since the introduction of Viagra in 1997, several case reports have linked Viagra to sudden sensorineural hearing loss. However, these studies are not well controlled for confounding factors, such as age and noise-induced hearing loss and none of these reports are based on prospective double-blind studies. Further, animal studies report contradictory data. For example, one study (2008) reported hearing loss in rats after long-term and high-dose exposure to sildenafil citrate. The other study (2012) showed vardenafil, another formulation of PDE5i, to be protective against noise-induced hearing loss in mice and rats. Whether or not clinically relevant doses of sildenafil citrate cause hearing loss in normal subjects (animals or humans) is controversial. One possibility is that PDE5i exacerbates age-related susceptibility to hearing loss in adults. Therefore, we tested sildenafil citrate in C57BL/6J, a strain of mice that displays increased susceptibility to age-related hearing loss, and compared the results to those obtained from the FVB/N, a strain of mice with no predisposition to hearing loss. Six-week-old mice were injected with the maximum tolerated dose of sildenafil citrate (10 mg/kg/day) or saline for 30 days. Auditory brainstem responses (ABRs) were recorded pre- and post injection time points to assess hearing loss. Entry of sildenafil citrate in the mouse cochlea was confirmed by qRT-PCR analysis of a downstream target of the cGMP-PKG cascade. ABR data indicated no statistically significant difference in hearing between treated and untreated mice in both backgrounds. Results show that the maximum tolerated dose of sildenafil citrate administered daily for 4 weeks does not affect hearing in the mouse. Our study gives no indication that Viagra will negatively impact hearing and it emphasizes the need to revisit the issue of Viagra related ototoxicity in humans. PMID:24244454
Revisiting a Statistical Shortcoming When Fitting the Langmuir Model to Sorption Data
Technology Transfer Automated Retrieval System (TEKTRAN)
The Langmuir model is commonly used for describing sorption behavior of reactive solutes to surfaces. Fitting the Langmuir model to sorption data requires either the use of nonlinear regression or, alternatively, linear regression using one of the linearized versions of the model. Statistical limit...
Revisiting single photon avalanche diode current-voltage modeling and transient characteristics
Javitt, M.; Savuskan, V. Merhav, T.; Nemirovsky, Y.
2014-05-28
A model for the current-voltage and transient behavior of Single Photon Avalanche Diodes (SPADs) based on device physics is presented. The results of the model are compared to actual measurements and a reasonable fit is seen. Additionally, the model provides a useful tool for designing quenching circuitry and determining optimal operation conditions of the SPAD.
ERIC Educational Resources Information Center
Clarke, Paul; Crawford, Claire; Steele, Fiona; Vignoles, Anna
2015-01-01
The use of fixed (FE) and random effects (RE) in two-level hierarchical linear regression is discussed in the context of education research. We compare the robustness of FE models with the modelling flexibility and potential efficiency of those from RE models. We argue that the two should be seen as complementary approaches. We then compare bothâ€¦
ERIC Educational Resources Information Center
Clarke, Paul; Crawford, Claire; Steele, Fiona; Vignoles, Anna
2015-01-01
The use of fixed (FE) and random effects (RE) in two-level hierarchical linear regression is discussed in the context of education research. We compare the robustness of FE models with the modelling flexibility and potential efficiency of those from RE models. We argue that the two should be seen as complementary approaches. We then compare both…
Revisiting Gaussian Process Regression Modeling for Localization in Wireless Sensor Networks.
Richter, Philipp; Toledano-Ayala, Manuel
2015-01-01
Signal strength-based positioning in wireless sensor networks is a key technology for seamless, ubiquitous localization, especially in areas where Global Navigation Satellite System (GNSS) signals propagate poorly. To enable wireless local area network (WLAN) location fingerprinting in larger areas while maintaining accuracy, methods to reduce the effort of radio map creation must be consolidated and automatized. Gaussian process regression has been applied to overcome this issue, also with auspicious results, but the fit of the model was never thoroughly assessed. Instead, most studies trained a readily available model, relying on the zero mean and squared exponential covariance function, without further scrutinization. This paper studies the Gaussian process regression model selection for WLAN fingerprinting in indoor and outdoor environments. We train several models for indoor/outdoor- and combined areas; we evaluate them quantitatively and compare them by means of adequate model measures, hence assessing the fit of these models directly. To illuminate the quality of the model fit, the residuals of the proposed model are investigated, as well. Comparative experiments on the positioning performance verify and conclude the model selection. In this way, we show that the standard model is not the most appropriate, discuss alternatives and present our best candidate. PMID:26370996
Revisiting Gaussian Process Regression Modeling for Localization in Wireless Sensor Networks
Richter, Philipp; Toledano-Ayala, Manuel
2015-01-01
Signal strength-based positioning in wireless sensor networks is a key technology for seamless, ubiquitous localization, especially in areas where Global Navigation Satellite System (GNSS) signals propagate poorly. To enable wireless local area network (WLAN) location fingerprinting in larger areas while maintaining accuracy, methods to reduce the effort of radio map creation must be consolidated and automatized. Gaussian process regression has been applied to overcome this issue, also with auspicious results, but the fit of the model was never thoroughly assessed. Instead, most studies trained a readily available model, relying on the zero mean and squared exponential covariance function, without further scrutinization. This paper studies the Gaussian process regression model selection for WLAN fingerprinting in indoor and outdoor environments. We train several models for indoor/outdoor- and combined areas; we evaluate them quantitatively and compare them by means of adequate model measures, hence assessing the fit of these models directly. To illuminate the quality of the model fit, the residuals of the proposed model are investigated, as well. Comparative experiments on the positioning performance verify and conclude the model selection. In this way, we show that the standard model is not the most appropriate, discuss alternatives and present our best candidate. PMID:26370996
The Two-Capacitor Problem Revisited: A Mechanical Harmonic Oscillator Model Approach
ERIC Educational Resources Information Center
Lee, Keeyung
2009-01-01
The well-known two-capacitor problem, in which exactly half the stored energy disappears when a charged capacitor is connected to an identical capacitor, is discussed based on the mechanical harmonic oscillator model approach. In the mechanical harmonic oscillator model, it is shown first that "exactly half" the work done by a constant appliedâ€¦
Magal, Pierre; Ruan, Shigui
2014-04-01
The classical susceptible-infectious-recovered (SIR) model, originated from the seminal papers of Ross [51] and Ross and Hudson [52,53] in 1916-1917 and the fundamental contributions of Kermack and McKendrick [36-38] in 1927-1932, describes the transmission of infectious diseases between susceptible and infective individuals and provides the basic framework for almost all later epidemic models, including stochastic epidemic models using Monte Carlo simulations or individual-based models (IBM). In this paper, by defining the rules of contacts between susceptible and infective individuals, the rules of transmission of diseases through these contacts, and the time of transmission during contacts, we provide detailed comparisons between the classical deterministic SIR model and the IBM stochastic simulations of the model. More specifically, for the purpose of numerical and stochastic simulations we distinguish two types of transmission processes: that initiated by susceptible individuals and that driven by infective individuals. Our analysis and simulations demonstrate that in both cases the IBM converges to the classical SIR model only in some particular situations. In general, the classical and individual-based SIR models are significantly different. Our study reveals that the timing of transmission in a contact at the individual level plays a crucial role in determining the transmission dynamics of an infectious disease at the population level. PMID:24530806
Revisiting the Model of Creative Destruction: St. Jacobs, Ontario, a Decade Later
ERIC Educational Resources Information Center
Mitchell, Clare J. A.; de Waal, Sarah B.
2009-01-01
Ten years ago, the model of creative destruction was developed to predict the fate of communities that base their development on the commodification of rural heritage (Mitchell, C.J.A., 1998. Entrepreneurialism, commodification and creative destruction: a model of post-modern community development. Journal of Rural Studies 14, 273-286). Itsâ€¦
The Two-Capacitor Problem Revisited: A Mechanical Harmonic Oscillator Model Approach
ERIC Educational Resources Information Center
Lee, Keeyung
2009-01-01
The well-known two-capacitor problem, in which exactly half the stored energy disappears when a charged capacitor is connected to an identical capacitor, is discussed based on the mechanical harmonic oscillator model approach. In the mechanical harmonic oscillator model, it is shown first that "exactly half" the work done by a constant applied…
Revisiting the Model of Creative Destruction: St. Jacobs, Ontario, a Decade Later
ERIC Educational Resources Information Center
Mitchell, Clare J. A.; de Waal, Sarah B.
2009-01-01
Ten years ago, the model of creative destruction was developed to predict the fate of communities that base their development on the commodification of rural heritage (Mitchell, C.J.A., 1998. Entrepreneurialism, commodification and creative destruction: a model of post-modern community development. Journal of Rural Studies 14, 273-286). Its…
NASA Astrophysics Data System (ADS)
Marshall, David P.; Munday, David R.; Allison, Lesley C.; Hay, Russell J.; Johnson, Helen L.
2016-01-01
Gill's (1968) model of the Antarctic Circumpolar Current (ACC) is reinterpreted for a stratified, reduced-gravity ocean, where the barotropic streamfunction is replaced by the pycnocline depth, and the bottom drag coefficient by the Gent and McWilliams eddy diffusivity. The resultant model gives a simple description of the lateral structure of the ACC that is consistent with contemporary descriptions of ACC dynamics. The model is used to investigate and interpret the sensitivity of the ACC to the latitudinal profile of the surface wind stress. A substantial ACC remains when the wind jet is shifted north of the model Drake Passage, even by several thousand kilometers. The integral of the wind stress over the circumpolar streamlines is found to be a useful predictor of the magnitude of the volume transport through the model Drake Passage, although it is necessary to correct for basin-wide zonal pressure gradients in order to obtain good quantitative agreement.
Revisiting node-based SIR models in complex networks with degree correlations
NASA Astrophysics Data System (ADS)
Wang, Yi; Cao, Jinde; Alofi, Abdulaziz; AL-Mazrooei, Abdullah; Elaiw, Ahmed
2015-11-01
In this paper, we consider two growing networks which will lead to the degree-degree correlations between two nearest neighbors in the network. When the network grows to some certain size, we introduce an SIR-like disease such as pandemic influenza H1N1/09 to the population. Due to its rapid spread, the population size changes slowly, and thus the disease spreads on correlated networks with approximately fixed size. To predict the disease evolution on correlated networks, we first review two node-based SIR models incorporating degree correlations and an edge-based SIR model without considering degree correlation, and then compare the predictions of these models with stochastic SIR simulations, respectively. We find that the edge-based model, even without considering degree correlations, agrees much better than the node-based models incorporating degree correlations with stochastic SIR simulations in many respects. Moreover, simulation results show that for networks with positive correlation, the edge-based model provides a better upper bound of the cumulative incidence than the node-based SIR models, whereas for networks with negative correlation, it provides a lower bound of the cumulative incidence.
Leue, Anja; Beauducel, André
2011-03-01
The Positive and Negative Affect Schedule (PANAS) is a widely used inventory for the assessment of affect in psychology and other applied sciences. Despite its popularity, the structure of the PANAS is still under debate. On the one hand, there is evidence of the traditional 2-factor model with Positive Affect (PA) and Negative Affect (NA) as uncorrelated factors. On the other hand, a more complex structure of the PANAS has been discussed. To shed further light on the core dimensions of the PANAS, 2 studies investigated the structure of the PANAS in 2 German samples (N = 354 and N = 364, respectively) by means of confirmatory factor analysis. The factor analysis results of Study 1 for a traitlike time frame instruction suggested a suboptimal model fit for the uncorrelated 2-factor model and the 3-factor model with PA, Afraid, and Upset as factors, whereas a superior model fit occurred for a bifactor model with traitlike PA, NA, and a general 3rd factor named Affective Polarity. In Study 2, the bifactor model was replicated for a statelike PANAS time frame instruction and evidence of criterion validity was provided for PA, NA, and Affective Polarity factors in 2 sex offender subgroups and in a community sample. With Affective Polarity, we introduce an affect dimension that captures additional variance beyond PA and NA. Because of the adjectives with relevant loadings on Affective Polarity, this general factor represents an individual's orientation toward approach and withdrawal, respectively. PMID:21280952
BrainSignals Revisited: Simplifying a Computational Model of Cerebral Physiology
Caldwell, Matthew; Hapuarachchi, Tharindi; Highton, David; Elwell, Clare; Smith, Martin; Tachtsidis, Ilias
2015-01-01
Multimodal monitoring of brain state is important both for the investigation of healthy cerebral physiology and to inform clinical decision making in conditions of injury and disease. Near-infrared spectroscopy is an instrument modality that allows non-invasive measurement of several physiological variables of clinical interest, notably haemoglobin oxygenation and the redox state of the metabolic enzyme cytochrome c oxidase. Interpreting such measurements requires the integration of multiple signals from different sources to try to understand the physiological states giving rise to them. We have previously published several computational models to assist with such interpretation. Like many models in the realm of Systems Biology, these are complex and dependent on many parameters that can be difficult or impossible to measure precisely. Taking one such model, BrainSignals, as a starting point, we have developed several variant models in which specific regions of complexity are substituted with much simpler linear approximations. We demonstrate that model behaviour can be maintained whilst achieving a significant reduction in complexity, provided that the linearity assumptions hold. The simplified models have been tested for applicability with simulated data and experimental data from healthy adults undergoing a hypercapnia challenge, but relevance to different physiological and pathophysiological conditions will require specific testing. In conditions where the simplified models are applicable, their greater efficiency has potential to allow their use at the bedside to help interpret clinical data in near real-time. PMID:25961297
The Goodwin model revisited: Hopf bifurcation, limit-cycle, and periodic entrainment
NASA Astrophysics Data System (ADS)
Woller, Aurore; Gonze, Didier; Erneux, Thomas
2014-08-01
The three-variable Goodwin oscillator is a minimal model demonstrating the emergence of oscillations in simple biochemical feedback systems. As a prototypical oscillator, this model was extensively studied from a theoretical point of view and applied to various biological systems, including circadian clocks. Here, we reexamine this model, derive analytically the amplitude equation near the Hopf bifurcation and investigate the effect of a periodic modulation of the oscillator. In particular, we compare the entrainment performance when the free oscillator displays either self-sustained or damped oscillations. We discuss the results in the context of circadian oscillators.
Revisiting the Cape Cod Bacteria Injection Experiment Using a Stochastic Modeling Approach
Maxwell, R M; Welty, C; Harvey, R W
2006-11-22
Bromide and resting-cell bacteria tracer tests carried out in a sand and gravel aquifer at the USGS Cape Cod site in 1987 were reinterpreted using a three-dimensional stochastic approach and Lagrangian particle tracking numerical methods. Bacteria transport was strongly coupled to colloid filtration through functional dependence of local-scale colloid transport parameters on hydraulic conductivity and seepage velocity in a stochastic advection-dispersion/attachment-detachment model. Information on geostatistical characterization of the hydraulic conductivity (K) field from a nearby plot was utilized as input that was unavailable when the original analysis was carried out. A finite difference model for groundwater flow and a particle-tracking model of conservative solute transport was calibrated to the bromide-tracer breakthrough data using the aforementioned geostatistical parameters. An optimization routine was utilized to adjust the mean and variance of the lnK field over 100 realizations such that a best fit of a simulated, average bromide breakthrough curve is achieved. Once the optimal bromide fit was accomplished (based on adjusting the lnK statistical parameters in unconditional simulations), a stochastic particle-tracking model for the bacteria was run without adjustments to the local-scale colloid transport parameters. Good predictions of the mean bacteria breakthrough data were achieved using several approaches for modeling components of the system. Simulations incorporating the recent Tufenkji and Elimelech [1] equation for estimating single collector efficiency were compared to those using the Rajagopalan and Tien [2] model. Both appeared to work equally well at predicting mean bacteria breakthrough using a constant mean bacteria diameter for this set of field conditions, with the Rajagopalan and Tien model yielding approximately a 30% lower peak concentration and less tailing than the Tufenkji and Elimelech formulation. Simulations using a distribution of bacterial cell diameters available from original field notes yielded a slight improvement in the model and data agreement compared to simulations using an average bacteria diameter; variable bacterial cell diameters lowered the modeled peak concentrations and more significantly diminished the tailing behavior, particularly for the Rajagopalan and Tien model of collision frequency. Spatial variability in detachment had little effect on the results. The Lagrangian particle transport model representing the non-idealities of the colloid transport process appears to be a robust, grid-free method for modeling field-scale distribution problems where incorporation of fine-scale heterogeneity would necessitate large numbers of computational cells. The stochastic approach based on estimates of local-scale parameters for the bacteria-transport process both captures the mean field behavior of bacteria transport and calculates an envelope of uncertainty that brackets the observations in most simulation cases.
A NEW APPROACH TO HYDROLOGIC MODELING: DERIVED DISTRIBUTIONS REVISITED. (R824780)
A fractal geometric procedure to model hydrologic (geophysical) phenomena is introduced. The method consists of using derived distributions, obtained by transforming arbitrary multinomial multifractal measures via fractal interpolating functions, to represent observed hydrologic ...
The consensus in the two-feature two-state one-dimensional Axelrod model revisited
NASA Astrophysics Data System (ADS)
Biral, Elias J. P.; Tilles, Paulo F. C.; Fontanari, José F.
2015-04-01
The Axelrod model for the dissemination of culture exhibits a rich spatial distribution of cultural domains, which depends on the values of the two model parameters: F, the number of cultural features and q, the common number of states each feature can assume. In the one-dimensional model with F = q = 2, which is closely related to the constrained voter model, Monte Carlo simulations indicate the existence of multicultural absorbing configurations in which at least one macroscopic domain coexist with a multitude of microscopic ones in the thermodynamic limit. However, rigorous analytical results for the infinite system starting from the configuration where all cultures are equally likely show convergence to only monocultural or consensus configurations. Here we show that this disagreement is due simply to the order that the time-asymptotic limit and the thermodynamic limit are taken in the simulations. In addition, we show how the consensus-only result can be derived using Monte Carlo simulations of finite chains.
Revisiting Turbulence Model Validation for High-Mach Number Axisymmetric Compression Corner Flows
NASA Technical Reports Server (NTRS)
Georgiadis, Nicholas J.; Rumsey, Christopher L.; Huang, George P.
2015-01-01
Two axisymmetric shock-wave/boundary-layer interaction (SWBLI) cases are used to benchmark one- and two-equation Reynolds-averaged Navier-Stokes (RANS) turbulence models. This validation exercise was executed in the philosophy of the NASA Turbulence Modeling Resource and the AIAA Turbulence Model Benchmarking Working Group. Both SWBLI cases are from the experiments of Kussoy and Horstman for axisymmetric compression corner geometries with SWBLI inducing flares of 20 and 30 degrees, respectively. The freestream Mach number was approximately 7. The RANS closures examined are the Spalart-Allmaras one-equation model and the Menter family of kappa - omega two equation models including the Baseline and Shear Stress Transport formulations. The Wind-US and CFL3D RANS solvers are employed to simulate the SWBLI cases. Comparisons of RANS solutions to experimental data are made for a boundary layer survey plane just upstream of the SWBLI region. In the SWBLI region, comparisons of surface pressure and heat transfer are made. The effects of inflow modeling strategy, grid resolution, grid orthogonality, turbulent Prandtl number, and code-to-code variations are also addressed.
Vellela, Melissa; Qian, Hong
2009-01-01
Schlögl's model is the canonical example of a chemical reaction system that exhibits bistability. Because the biological examples of bistability and switching behaviour are increasingly numerous, this paper presents an integrated deterministic, stochastic and thermodynamic analysis of the model. After a brief review of the deterministic and stochastic modelling frameworks, the concepts of chemical and mathematical detailed balances are discussed and non-equilibrium conditions are shown to be necessary for bistability. Thermodynamic quantities such as the flux, chemical potential and entropy production rate are defined and compared across the two models. In the bistable region, the stochastic model exhibits an exchange of the global stability between the two stable states under changes in the pump parameters and volume size. The stochastic entropy production rate shows a sharp transition that mirrors this exchange. A new hybrid model that includes continuous diffusion and discrete jumps is suggested to deal with the multiscale dynamics of the bistable system. Accurate approximations of the exponentially small eigenvalue associated with the time scale of this switching and the full time-dependent solution are calculated using Matlab. A breakdown of previously known asymptotic approximations on small volume scales is observed through comparison with these and Monte Carlo results. Finally, in the appendix section is an illustration of how the diffusion approximation of the chemical master equation can fail to represent correctly the mesoscopically interesting steady-state behaviour of the system. PMID:19095615
NASA Astrophysics Data System (ADS)
Liu, C.; Wen, H.; Liu, J.; Ko, M.; Yan, H.; Chang, L.
2012-12-01
Repeat topography surveys provides a geometrically-corrected frame with relief information, which is crucial for studying geomorphic changes after a major slope hazard, such as the debris flow or landslides. The successful operation of Formosat-2 has proved the concept that the temporal resolution of a remote sensing system can be much improved by deploying a high-spatial-resolution sensor in a daily revisit orbit, as each accessible scene can be systematically observed from the same angle under similar illumination conditions. These characteristics make Formosat-2 an ideal satellite for site surveillance, and its images have been successfully applied in environmental monitoring, hazard assessment, orthomap generation, rapidly responding to a global disaster event, and land use management. The attempt of using a Formosat-2 stereo pair to generate a DSM, however, has not been very successful up-to-date. Ironically, it is mainly due to the characteristics of daily-revisit orbit as well. According to the parallax equation, to obtain an accurate height estimation requires a high disparity precision from the stereo pair. The most convenient approach is to maximize the baseline B or the baseline/height (B/H) ratio to a preferred range 0.6 to 1. It is not feasible, however, to acquire an across-track stereo pair with that range of baseline from the daily-revisit orbit using Formosat-2. Even taking the orbit drifting into consideration, it would take a few months to achieve a B/H ratio of approximately 0.15 across track. Another approach is to acquire an along-track stereo pair. But for the mountainous areas, such as the central mountain areas, in Taiwan, the shaded effect and geometrically distortion are apparent. This prohibits any attempt to employ the automatic image matching technique to generate a DSM based on the disparities retrieved from Frmosat-2 along-track stereo pair directly. Phase correlation is operated in the frequency-domain, which enables the relative translative offset between two similar images to be rapidly estimated. To meet the requirements in remote sensing and biomedical imaging, the technology of phase correlation has been extended to the sub-pixel level. Liu and Yan (2008) developed a robust phase correlation model using the based feature matching for image co-registration and DEM generation. Considering the fact that the Formosat-2 consecutive images are intrinsically stereo pairs with very narrow baselines, this innovative stereo-matching algorithm based on SPPC technique is employed to process Formosat-2 daily revisit stereo pairs with very narrow baselines. The detailed accuracy and efficiency analysis is investigated for the study area, Namasha, Kaohsiung, using the 50cm resolution aerial photo and the 2m resolution DEM derived from airborne LiDAR data. The archive of Formosat-2 images in Taiwan area collected from 2005 to 2012 was screened out, with the intention to select the consecutive pairs of those areas where major slope disasters occurred in the past eight years. This research encourages the repeated topography surveys of geomorphic changes using digital surface models deriving from Formosat-2 daily revisit stereo pair with very narrow baseline.
Fischer, Lukas P; Peter, Toni; Holm, Christian; de Graaf, Joost
2015-08-28
The so-called "raspberry" model refers to the hybrid lattice-Boltzmann and Langevin molecular dynamics scheme for simulating the dynamics of suspensions of colloidal particles, originally developed by Lobaskin and Dünweg [New J. Phys. 6, 54 (2004)], wherein discrete surface points are used to achieve fluid-particle coupling. This technique has been used in many simulation studies on the behavior of colloids. However, there are fundamental questions with regards to the use of this model. In this paper, we examine the accuracy with which the raspberry method is able to reproduce Stokes-level hydrodynamic interactions when compared to analytic expressions for solid spheres in simple-cubic crystals. To this end, we consider the quality of numerical experiments that are traditionally used to establish these properties and we discuss their shortcomings. We show that there is a discrepancy between the translational and rotational mobility reproduced by the simple raspberry model and present a way to numerically remedy this problem by adding internal coupling points. Finally, we examine a non-convex shape, namely, a colloidal dumbbell, and show that the filled raspberry model replicates the desired hydrodynamic behavior in bulk for this more complicated shape. Our investigation is continued in de Graaf et al. [J. Chem. Phys. 143, 084108 (2015)], wherein we consider the raspberry model in the confining geometry of two parallel plates. PMID:26328818
Revisiting the Rigidly Rotating Magnetosphere model for Ïƒ Ori E - I. Observations and data analysis
NASA Astrophysics Data System (ADS)
Oksala, M. E.; Wade, G. A.; Townsend, R. H. D.; Owocki, S. P.; Kochukhov, O.; Neiner, C.; Alecian, E.; Grunhut, J.
2012-01-01
We have obtained 18 new high-resolution spectropolarimetric observations of the B2Vp star Ïƒ Ori E with both the Narval and ESPaDOnS spectropolarimeters. The aim of these observations is to test, with modern data, the assumptions of the Rigidly Rotating Magnetosphere (RRM) model of Townsend & Owocki, applied to the specific case of Ïƒ Ori E by Townsend, Owocki & Groote. This model includes a substantially offset dipole magnetic field configuration, and approximately reproduces previous observational variations in longitudinal field strength, photometric brightness and HÎ± emission. We analyse new spectroscopy, including H I, He I, C II, Si III and Fe III lines, confirming the diversity of variability in photospheric lines, as well as the double S-wave variation of circumstellar hydrogen. Using the multiline analysis method of least-squares deconvolution (LSD), new, more precise longitudinal magnetic field measurements reveal a substantial variance between the shapes of the observed and RRM model time-varying field. The phase-resolved Stokes V profiles of He I 5876 and 6678 Ã… lines are fitted poorly by synthetic profiles computed from the magnetic topology assumed by Townsend et al.. These results challenge the offset dipole field configuration assumed in the application of the RRM model to Ïƒ Ori E, and indicate that future models of its magnetic field should also include complex, higher order components.