Infrared length scale and extrapolations for the no-core shell model
NASA Astrophysics Data System (ADS)
Wendt, K. A.; Forssén, C.; Papenbrock, T.; Sääf, D.
2015-06-01
We precisely determine the infrared (IR) length scale of the no-core shell model (NCSM). In the NCSM, the A -body Hilbert space is truncated by the total energy, and the IR length can be determined by equating the intrinsic kinetic energy of A nucleons in the NCSM space to that of A nucleons in a 3 (A -1 ) -dimensional hyper-radial well with a Dirichlet boundary condition for the hyper radius. We demonstrate that this procedure indeed yields a very precise IR length by performing large-scale NCSM calculations for 6Li. We apply our result and perform accurate IR extrapolations for bound states of 4He,6He,6Li , and 7Li . We also attempt to extrapolate NCSM results for 10B and 16O with bare interactions from chiral effective field theory over tens of MeV.
Infrared length scale and extrapolations for the no-core shell model
Wendt, K. A.; Forssén, C.; Papenbrock, T.; Sääf, D.
2015-06-03
In this paper, we precisely determine the infrared (IR) length scale of the no-core shell model (NCSM). In the NCSM, the A-body Hilbert space is truncated by the total energy, and the IR length can be determined by equating the intrinsic kinetic energy of A nucleons in the NCSM space to that of A nucleons in a 3(A-1)-dimensional hyper-radial well with a Dirichlet boundary condition for the hyper radius. We demonstrate that this procedure indeed yields a very precise IR length by performing large-scale NCSM calculations for 6Li. We apply our result and perform accurate IR extrapolations for bound statesmore » of 4He, 6He, 6Li, and 7Li. Finally, we also attempt to extrapolate NCSM results for 10B and 16O with bare interactions from chiral effective field theory over tens of MeV.« less
NASA Technical Reports Server (NTRS)
Mack, Robert J.; Kuhn, Neil S.
2006-01-01
A study was performed to determine a limiting separation distance for the extrapolation of pressure signatures from cruise altitude to the ground. The study was performed at two wind-tunnel facilities with two research low-boom wind-tunnel models designed to generate ground pressure signatures with "flattop" shapes. Data acquired at the first wind-tunnel facility showed that pressure signatures had not achieved the desired low-boom features for extrapolation purposes at separation distances of 2 to 5 span lengths. However, data acquired at the second wind-tunnel facility at separation distances of 5 to 20 span lengths indicated the "limiting extrapolation distance" had been achieved so pressure signatures could be extrapolated with existing codes to obtain credible predictions of ground overpressures.
Choi, Yoonseuk; Rosenblatt, Charles
2010-05-01
A herringbone "easy axis" pattern is scribed into a polyimide alignment layer for liquid-crystal orientation using the stylus of an atomic force microscope. Owing to the liquid crystal's bend elasticity K33 , the nematic director is unable to follow the sharp turn in the scribed easy axis, but instead relaxes over an extrapolation length L=K33/W2φ, where W2φ is the quadratic azimuthal anchoring strength coefficient. By immersing a tapered optical fiber into the liquid crystal, illuminating the fiber with polarized light, and scanning the fiber close to the substrate, a visualization and direct measurement of L are obtained on approaching the nematic-smectic- A phase transition temperature T NA from above. L is found to exhibit a sharp pretransitional increase near T NA, consistent with a diverging bend elastic constant. PMID:20866248
Richardson Extrapolation using DNAD
NASA Astrophysics Data System (ADS)
Celik, Ismail B.; Sezer, Hayri; Pakalapati, Suryanarayana R.; WVU-CFD Team
2013-11-01
Dual Number Automatic Derivation (DNAD) is a technique whereby a computer code can be executed with additional variable declarations to extend real number to a two dimensional space which is then used to evaluate derivatives to machine accuracy. In the literature this technique is usually applied to study sensitivities of calculations to model parameters, but not the mesh size. The current study explores possibilities of using the same technique to evaluate the derivative of the numerical solution with respect to mesh size which in turn can be used in the Taylor series expansion of the discretization error to calculate the error itself by way of extrapolation. Thus the new method enables explicit Richardson extrapolation by using only one set of calculations on a single grid. The extrapolation can be improved if an additional set of calculations are performed on a finer or a coarser mesh. The concept is demonstrated using one-dimensional example problems. Possible extension to multi-dimensions is discussed.
Biosimilars: Extrapolation for oncology.
Curigliano, Giuseppe; O'Connor, Darran P; Rosenberg, Julie A; Jacobs, Ira
2016-08-01
A biosimilar is a biologic that is highly similar to a licensed biologic (the reference product) in terms of purity, safety and efficacy. If the reference product is licensed to treat multiple therapeutic indications, extrapolation of indications, i.e., approval of a biosimilar for use in an indication held by the reference product but not directly studied in a comparative clinical trial with the biosimilar, may be possible but has to be scientifically justified. Here, we describe the data required to establish biosimilarity and emphasize that indication extrapolation is based on scientific principles and known mechanism of action. PMID:27354233
Ecotoxicological effects extrapolation models
Suter, G.W. II
1996-09-01
One of the central problems of ecological risk assessment is modeling the relationship between test endpoints (numerical summaries of the results of toxicity tests) and assessment endpoints (formal expressions of the properties of the environment that are to be protected). For example, one may wish to estimate the reduction in species richness of fishes in a stream reach exposed to an effluent and have only a fathead minnow 96 hr LC50 as an effects metric. The problem is to extrapolate from what is known (the fathead minnow LC50) to what matters to the decision maker, the loss of fish species. Models used for this purpose may be termed Effects Extrapolation Models (EEMs) or Activity-Activity Relationships (AARs), by analogy to Structure-Activity Relationships (SARs). These models have been previously reviewed in Ch. 7 and 9 of and by an OECD workshop. This paper updates those reviews and attempts to further clarify the issues involved in the development and use of EEMs. Although there is some overlap, this paper does not repeat those reviews and the reader is referred to the previous reviews for a more complete historical perspective, and for treatment of additional extrapolation issues.
Evidence, eminence and extrapolation.
Hlavin, Gerald; Koenig, Franz; Male, Christoph; Posch, Martin; Bauer, Peter
2016-06-15
A full independent drug development programme to demonstrate efficacy may not be ethical and/or feasible in small populations such as paediatric populations or orphan indications. Different levels of extrapolation from a larger population to smaller target populations are widely used for supporting decisions in this situation. There are guidance documents in drug regulation, where a weakening of the statistical rigour for trials in the target population is mentioned to be an option for dealing with this problem. To this end, we propose clinical trials designs, which make use of prior knowledge on efficacy for inference. We formulate a framework based on prior beliefs in order to investigate when the significance level for the test of the primary endpoint in confirmatory trials can be relaxed (and thus the sample size can be reduced) in the target population while controlling a certain posterior belief in effectiveness after rejection of the null hypothesis in the corresponding confirmatory statistical test. We show that point-priors may be used in the argumentation because under certain constraints, they have favourable limiting properties among other types of priors. The crucial quantity to be elicited is the prior belief in the possibility of extrapolation from a larger population to the target population. We try to illustrate an existing decision tree for extrapolation to paediatric populations within our framework. © 2016 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd. PMID:26753552
NASA Astrophysics Data System (ADS)
Van Zandt, James R.
2012-05-01
Steady-state performance of a tracking filter is traditionally evaluated immediately after a track update. However, there is commonly a further delay (e.g., processing and communications latency) before the tracks can actually be used. We analyze the accuracy of extrapolated target tracks for four tracking filters: Kalman filter with the Singer maneuver model and worst-case correlation time, with piecewise constant white acceleration, and with continuous white acceleration, and the reduced state filter proposed by Mookerjee and Reifler.1, 2 Performance evaluation of a tracking filter is significantly simplified by appropriate normalization. For the Kalman filter with the Singer maneuver model, the steady-state RMS error immediately after an update depends on only two dimensionless parameters.3 By assuming a worst case value of target acceleration correlation time, we reduce this to a single parameter without significantly changing the filter performance (within a few percent for air tracking).4 With this simplification, we find for all four filters that the RMS errors for the extrapolated state are functions of only two dimensionless parameters. We provide simple analytic approximations in each case.
The Extrapolation of Elementary Sequences
NASA Technical Reports Server (NTRS)
Laird, Philip; Saul, Ronald
1992-01-01
We study sequence extrapolation as a stream-learning problem. Input examples are a stream of data elements of the same type (integers, strings, etc.), and the problem is to construct a hypothesis that both explains the observed sequence of examples and extrapolates the rest of the stream. A primary objective -- and one that distinguishes this work from previous extrapolation algorithms -- is that the same algorithm be able to extrapolate sequences over a variety of different types, including integers, strings, and trees. We define a generous family of constructive data types, and define as our learning bias a stream language called elementary stream descriptions. We then give an algorithm that extrapolates elementary descriptions over constructive datatypes and prove that it learns correctly. For freely-generated types, we prove a polynomial time bound on descriptions of bounded complexity. An especially interesting feature of this work is the ability to provide quantitative measures of confidence in competing hypotheses, using a Bayesian model of prediction.
Infrared extrapolations for atomic nuclei
Furnstahl, R. J.; Hagen, Gaute; Papenbrock, Thomas F.; Wendt, Kyle A.
2015-01-01
Harmonic oscillator model-space truncations introduce systematic errors to the calculation of binding energies and other observables. We identify the relevant infrared (IR) scaling variable and give values for this nucleus-dependent quantity. We consider isotopes of oxygen computed with the coupled-cluster method from chiral nucleon–nucleon interactions at next-to-next-to-leading order and show that the IR component of the error is sufficiently understood to permit controlled extrapolations. By employing oscillator spaces with relatively large frequencies, that are well above the energy minimum, the ultraviolet corrections can be suppressed while IR extrapolations over tens of MeVs are accurate for ground-state energies. However, robust uncertainty quantification for extrapolated quantities that fully accounts for systematic errors is not yet developed.
Extrapolation methods for vector sequences
NASA Technical Reports Server (NTRS)
Smith, David A.; Ford, William F.; Sidi, Avram
1987-01-01
This paper derives, describes, and compares five extrapolation methods for accelerating convergence of vector sequences or transforming divergent vector sequences to convergent ones. These methods are the scalar epsilon algorithm (SEA), vector epsilon algorithm (VEA), topological epsilon algorithm (TEA), minimal polynomial extrapolation (MPE), and reduced rank extrapolation (RRE). MPE and RRE are first derived and proven to give the exact solution for the right 'essential degree' k. Then, Brezinski's (1975) generalization of the Shanks-Schmidt transform is presented; the generalized form leads from systems of equations to TEA. The necessary connections are then made with SEA and VEA. The algorithms are extended to the nonlinear case by cycling, the error analysis for MPE and VEA is sketched, and the theoretical support for quadratic convergence is discussed. Strategies for practical implementation of the methods are considered.
Infrared extrapolations for atomic nuclei
Furnstahl, R. J.; Hagen, Gaute; Papenbrock, Thomas F.; Wendt, Kyle A.
2015-01-01
Harmonic oscillator model-space truncations introduce systematic errors to the calculation of binding energies and other observables. We identify the relevant infrared (IR) scaling variable and give values for this nucleus-dependent quantity. We consider isotopes of oxygen computed with the coupled-cluster method from chiral nucleon–nucleon interactions at next-to-next-to-leading order and show that the IR component of the error is sufficiently understood to permit controlled extrapolations. By employing oscillator spaces with relatively large frequencies, that are well above the energy minimum, the ultraviolet corrections can be suppressed while IR extrapolations over tens of MeVs are accurate for ground-state energies. However, robust uncertaintymore » quantification for extrapolated quantities that fully accounts for systematic errors is not yet developed.« less
Extrapolation limitations of multilayer feedforward neural networks
NASA Technical Reports Server (NTRS)
Haley, Pamela J.; Soloway, Donald
1992-01-01
The limitations of backpropagation used as a function extrapolator were investigated. Four common functions were used to investigate the network's extrapolation capability. The purpose of the experiment was to determine whether neural networks are capable of extrapolation and, if so, to determine the range for which networks can extrapolate. The authors show that neural networks cannot extrapolate and offer an explanation to support this result.
Ethanol kinetics: extent of error in back extrapolation procedures.
al-Lanqawi, Y; Moreland, T A; McEwen, J; Halliday, F; Durnin, C J; Stevenson, I H
1992-01-01
1. Plasma ethanol concentrations were measured in 24 male volunteers for 9 h after a single oral dose of 710 mg kg-1. 2. The rate of decline of the plasma ethanol concentration (k0; mean +/- s.d.), was 186 +/- 26 mg l-1 h-1. 3. In each individual, three elimination rates were used to back-extrapolate plasma ethanol concentrations over 3 and 5 h periods from observed values at 4 h and 6 h post-dosing assuming zero-order kinetics. The extrapolated values were then compared with the observed concentrations. 4. Using the mean k0 values for the subjects the mean error in back extrapolation was small but highly variable. The variability in the error increased with the length of the extrapolation period. 5. When a k0 value of 150 mg l-1 h-1 (a value often cited as a population mean) was used for back extrapolation this resulted in significant under-estimation of actual values whereas the use of a k0 value of 238 mg l-1 h-1 (the highest value observed in the present study) resulted in significant over-estimation of actual values. 6. These results indicate that because the kinetics of ethanol are associated with substantial inter-subject variability the use of a single slope value to back calculate blood concentrations can give rise to considerable error. PMID:1457265
AXES OF EXTRAPOLATION IN RISK ASSESSMENTS
Extrapolation in risk assessment involves the use of data and information to estimate or predict something that has not been measured or observed. Reasons for extrapolation include that the number of combinations of environmental stressors and possible receptors is too large to c...
Builtin vs. auxiliary detection of extrapolation risk.
Munson, Miles Arthur; Kegelmeyer, W. Philip,
2013-02-01
A key assumption in supervised machine learning is that future data will be similar to historical data. This assumption is often false in real world applications, and as a result, prediction models often return predictions that are extrapolations. We compare four approaches to estimating extrapolation risk for machine learning predictions. Two builtin methods use information available from the classification model to decide if the model would be extrapolating for an input data point. The other two build auxiliary models to supplement the classification model and explicitly model extrapolation risk. Experiments with synthetic and real data sets show that the auxiliary models are more reliable risk detectors. To best safeguard against extrapolating predictions, however, we recommend combining builtin and auxiliary diagnostics.
Implicit Extrapolation Methods for Variable Coefficient Problems
NASA Technical Reports Server (NTRS)
Jung, M.; Ruede, U.
1996-01-01
Implicit extrapolation methods for the solution of partial differential equations are based on applying the extrapolation principle indirectly. Multigrid tau-extrapolation is a special case of this idea. In the context of multilevel finite element methods, an algorithm of this type can be used to raise the approximation order, even when the meshes are nonuniform or locally refined. Here previous results are generalized to the variable coefficient case and thus become applicable for nonlinear problems. The implicit extrapolation multigrid algorithm converges to the solution of a higher order finite element system. This is obtained without explicitly constructing higher order stiffness matrices but by applying extrapolation in a natural form within the algorithm. The algorithm requires only a small change of a basic low order multigrid method.
Extrapolation procedures in Mott electron polarimetry
NASA Technical Reports Server (NTRS)
Gay, T. J.; Khakoo, M. A.; Brand, J. A.; Furst, J. E.; Wijayaratna, W. M. K. P.; Meyer, W. V.; Dunning, F. B.
1992-01-01
In standard Mott electron polarimetry using thin gold film targets, extrapolation procedures must be used to reduce the experimentally measured asymmetries A to the values they would have for scattering from single atoms. These extrapolations involve the dependent of A on either the gold film thickness or the maximum detected electron energy loss in the target. A concentric cylindrical-electrode Mott polarimeter, has been used to study and compare these two types of extrapolations over the electron energy range 20-100 keV. The potential systematic errors which can result from such procedures are analyzed in detail, particularly with regard to the use of various fitting functions in thickness extrapolations, and the failure of perfect energy-loss discrimination to yield accurate polarizations when thick foils are used.
Cosmogony as an extrapolation of magnetospheric research
NASA Technical Reports Server (NTRS)
Alfven, H.
1984-01-01
A theory of the origin and evolution of the Solar System which considered electromagnetic forces and plasma effects is revised in light of information supplied by space research. In situ measurements in the magnetospheres and solar wind can be extrapolated outwards in space, to interstellar clouds, and backwards in time, to the formation of the solar system. The first extrapolation leads to a revision of cloud properties essential for the early phases in the formation of stars and solar nebulae. The latter extrapolation facilitates analysis of the cosmogonic processes by extrapolation of magnetospheric phenomena. Pioneer-Voyager observations of the Saturnian rings indicate that essential parts of their structure are fossils from cosmogonic times. By using detailed information from these space missions, it is possible to reconstruct events 4 to 5 billion years ago with an accuracy of a few percent.
Endangered species toxicity extrapolation using ICE models
The National Research Council’s (NRC) report on assessing pesticide risks to threatened and endangered species (T&E) included the recommendation of using interspecies correlation models (ICE) as an alternative to general safety factors for extrapolating across species. ...
Essentially nonoscillatory (ENO) reconstructions via extrapolation
NASA Technical Reports Server (NTRS)
Suresh, Ambady; Jorgenson, Philip C. E.
1995-01-01
In this paper, the algorithm for determining the stencil of a one-dimensional Essentially Nonoscillatory (ENO) reconstruction scheme on a uniform grid is reinterpreted as being based on extrapolation. This view leads to another extension of ENO reconstruction schemes to two-dimensional unstructured triangular meshes. The key idea here is to select several cells of the stencil in one step based on extrapolation rather than one cell at a time. Numerical experiments confirm that the new scheme yields sharp nonoscillatory reconstructions and that it is about five times faster than previous schemes.
Extrapolated implicit-explicit time stepping.
Constantinescu, E. M.; Sandu, A.; Mathematics and Computer Science; Virginia Polytechnic Inst. and State Univ.
2010-01-01
This paper constructs extrapolated implicit-explicit time stepping methods that allow one to efficiently solve problems with both stiff and nonstiff components. The proposed methods are based on Euler steps and can provide very high order discretizations of ODEs, index-1 DAEs, and PDEs in the method-of-lines framework. Implicit-explicit schemes based on extrapolation are simple to construct, easy to implement, and straightforward to parallelize. This work establishes the existence of perturbed asymptotic expansions of global errors, explains the convergence orders of these methods, and studies their linear stability properties. Numerical results with stiff ODE, DAE, and PDE test problems confirm the theoretical findings and illustrate the potential of these methods to solve multiphysics multiscale problems.
Chiral extrapolation of SU(3) amplitudes
Ecker, Gerhard
2011-05-23
Approximations of chiral SU(3) amplitudes at NNLO are proposed to facilitate the extrapolation of lattice data to the physical meson masses. Inclusion of NNLO terms is essential for investigating convergence properties of chiral SU(3) and for determining low-energy constants in a controllable fashion. The approximations are tested with recent lattice data for the ratio of decay constants F{sub K}/F{sub {pi}}.
Extrapolation discontinuous Galerkin method for ultraparabolic equations
NASA Astrophysics Data System (ADS)
Marcozzi, Michael D.
2009-02-01
Ultraparabolic equations arise from the characterization of the performance index of stochastic optimal control relative to ultradiffusion processes; they evidence multiple temporal variables and may be regarded as parabolic along characteristic directions. We consider theoretical and approximation aspects of a temporally order and step size adaptive extrapolation discontinuous Galerkin method coupled with a spatial Lagrange second-order finite element approximation for a prototype ultraparabolic problem. As an application, we value a so-called Asian option from mathematical finance.
Extrapolating phosphorus production to estimate resource reserves.
Vaccari, David A; Strigul, Nikolay
2011-08-01
Various indicators of resource scarcity and methods for extrapolating resource availability are examined for phosphorus. These include resource lifetime, and trends in resource price, ore grade and discovery rates, and Hubbert curve extrapolation. Several of these indicate increasing scarcity of phosphate resources. Calculated resource lifetime is subject to a number of caveats such as unanticipated future changes in resource discovery, mining and beneficiation technology, population growth or per-capita demand. Thus it should be used only as a rough planning index or as a relative indicator of potential scarcity. This paper examines the uncertainty in one method for estimating available resources from historical production data. The confidence intervals for the parameters and predictions of the Hubbert curves are computed as they relate to the amount of information available. These show that Hubbert-type extrapolations are not robust for predicting the ultimately recoverable reserves or year of peak production of phosphate rock. Previous successes of the Hubbert curve are for cases in which there exist alternative resources, which is not the situation for phosphate. It is suggested that data other than historical production, such as population growth, identified resources and economic factors, should be included in making such forecasts. PMID:21440285
Surface dose measurement using TLD powder extrapolation
Rapley, P. . E-mail: rapleyp@tbh.net
2006-10-01
Surface/near-surface dose measurements in therapeutic x-ray beams are important in determining the dose to the dermal and epidermal skin layers during radiation treatment. Accurate determination of the surface dose is a difficult but important task for proper treatment of patients. A new method of measuring surface dose in phantom through extrapolation of readings from various thicknesses of thermoluminescent dosimeter (TLD) powder has been developed and investigated. A device was designed, built, and tested that provides TLD powder thickness variation to a minimum thickness of 0.125 mm. Variations of the technique have been evaluated to optimize precision with consideration of procedural ease. Results of this study indicate that dose measurements (relative to D{sub max}) in regions of steep dose gradient in the beam axis direction are possible with a precision (2 standard deviations [SDs]) as good as {+-} 1.2% using the technique. The dosimeter was developed and evaluated using variation to the experimental method. A clinically practical procedure was determined, resulting in measured surface dose of 20.4 {+-} 2% of the D{sub max} dose for a 10 x 10 cm{sup 2}, 80-cm source-to-surface distance (SSD), Theratron 780 Cobalt-60 ({sup 60}C) beam. Results obtained with TLD powder extrapolation compare favorably to other methods presented in the literature. The TLD powder extrapolation tool has been used clinically at the Northwestern Ontario Regional Cancer Centre (NWORCC) to measure surface dose effects under a number of conditions. Results from these measurements are reported. The method appears to be a simple and economical tool for surface dose measurement, particularly for facilities with TLD powder measurement capabilities.
NASA Astrophysics Data System (ADS)
Kosek, Wieslaw
2016-04-01
Future Earth Orientation Parameters data are needed to compute real time transformation between the celestial and terrestrial reference frames. This transformation is realized by predictions of x, y pole coordinates data, UT1-UTC data and precesion-nutation extrapolation model. This paper is focused on the pole coordinates data prediction by combination of the least-squares (LS) extrapolation and autoregressive (AR) prediction models (LS+AR). The AR prediction which is applied to the LS extrapolation residuals of pole coordinates data does not able to predict all frequency bands of them and it is mostly tuned to predict subseasonal oscillations. The absolute values of differences between pole coordinates data and their LS+AR predictions increase with prediction length and depend mostly on starting prediction epochs, thus time series of these differences for 2, 4 and 8 weeks in the future were analyzed. Time frequency spectra of these differences for different prediction lengths are very similar showing some power in the frequency band corresponding to the prograde Chandler and annual oscillations, which means that the increase of prediction errors is caused by mismodelling of these oscillations by the LS extrapolation model. Thus, the LS+AR prediction method can be modified by taking into additional AR prediction correction computed from time series of these prediction differences for different prediction lengths. This additional AR prediction is mostly tuned to the seasonal frequency band of pole coordinates data.
Dioxin equivalency: Challenge to dose extrapolation
Brown, J.F. Jr.; Silkworth, J.B.
1995-12-31
Extensive research has shown that all biological effects of dioxin-like agents are mediated via a single biochemical target, the Ah receptor (AhR), and that the relative biologic potencies of such agents in any given system, coupled with their exposure levels, may be described in terms of toxic equivalents (TEQ). It has also shown that the TEQ sources include not only chlorinated species such as the dioxins (PCDDs), PCDFs, and coplanar PCBs, but also non-chlorinated substances such as the PAHs of wood smoke, the AhR agonists of cooked meat, and the indolocarbazol (ICZ) derived from cruciferous vegetables. Humans have probably had elevated exposures to these non-chlorinated TEQ sources ever since the discoveries of fire, cooking, and the culinary use of Brassica spp. Recent assays of CYP1A2 induction show that these ``natural`` or ``traditional`` AhR agonists are contributing 50--100 times as much to average human TEQ exposures as do the chlorinated xenobiotics. Currently, the safe doses of the xenobiotic TEQ sources are estimated from their NOAELs and large extrapolation factors, derived from arbitrary mathematical models, whereas the NOAELs themselves are regarded as the safe doses for the TEQs of traditional dietary components. Available scientific data can neither support nor refute either approach to assessing the health risk of an individual chemical substance. However, if two substances be toxicologically equivalent, then their TEQ-adjusted health risks must also be equivalent, and the same dose extrapolation procedure should be used for both.
Extrapolating Solar Dynamo Models Throughout the Heliosphere
NASA Astrophysics Data System (ADS)
Cox, B. T.; Miesch, M. S.; Augustson, K.; Featherstone, N. A.
2014-12-01
There are multiple theories that aim to explain the behavior of the solar dynamo, and their associated models have been fiercely contested. The two prevailing theories investigated in this project are the Convective Dynamo model that arises from the pure solving of the magnetohydrodynamic equations, as well as the Babcock-Leighton model that relies on sunspot dissipation and reconnection. Recently, the supercomputer simulations CASH and BASH have formed models of the behavior of the Convective and Babcock-Leighton models, respectively, in the convective zone of the sun. These models show the behavior of the models within the sun, while much less is known about the effects these models may have further away from the solar surface. The goal of this work is to investigate any fundamental differences between the Convective and Babcock-Leighton models of the solar dynamo outside of the sun and extending into the solar system via the use of potential field source surface extrapolations implemented via python code that operates on data from CASH and BASH. The use of real solar data to visualize supergranular flow data in the BASH model is also used to learn more about the behavior of the Babcock-Leighton Dynamo. From the process of these extrapolations it has been determined that the Babcock-Leighton model, as represented by BASH, maintains complex magnetic fields much further into the heliosphere before reverting into a basic dipole field, providing 3D visualisations of the models distant from the sun.
Extrapolation methods for dynamic partial differential equations
NASA Technical Reports Server (NTRS)
Turkel, E.
1978-01-01
Several extrapolation procedures are presented for increasing the order of accuracy in time for evolutionary partial differential equations. These formulas are based on finite difference schemes in both the spatial and temporal directions. On practical grounds the methods are restricted to schemes that are fourth order in time and either second, fourth or sixth order in space. For hyperbolic problems the second order in space methods are not useful while the fourth order methods offer no advantage over the Kreiss-Oliger method unless very fine meshes are used. Advantages are first achieved using sixth order methods in space coupled with fourth order accuracy in time. Computational results are presented confirming the analytic discussions.
Extrapolation of toxic indices among test objects
Tichý, Miloň; Rucki, Marián; Roth, Zdeněk; Hanzlíková, Iveta; Vlková, Alena; Tumová, Jana; Uzlová, Rút
2010-01-01
Oligochaeta Tubifex tubifex, fish fathead minnow (Pimephales promelas), hepatocytes isolated from rat liver and ciliated protozoan are absolutely different organisms and yet their acute toxicity indices correlate. Correlation equations for special effects were developed for a large heterogeneous series of compounds (QSAR, quantitative structure-activity relationships). Knowing those correlation equations and their statistic evaluation, one can extrapolate the toxic indices. The reason is that a common physicochemical property governs the biological effect, namely the partition coefficient between two unmissible phases, simulated generally by n-octanol and water. This may mean that the transport of chemicals towards a target is responsible for the magnitude of the effect, rather than reactivity, as one would assume suppose. PMID:21331180
Extrapolation of toxic indices among test objects.
Tichý, Miloň; Rucki, Marián; Roth, Zdeněk; Hanzlíková, Iveta; Vlková, Alena; Tumová, Jana; Uzlová, Rút
2010-12-01
Oligochaeta Tubifex tubifex, fish fathead minnow (Pimephales promelas), hepatocytes isolated from rat liver and ciliated protozoan are absolutely different organisms and yet their acute toxicity indices correlate. Correlation equations for special effects were developed for a large heterogeneous series of compounds (QSAR, quantitative structure-activity relationships). Knowing those correlation equations and their statistic evaluation, one can extrapolate the toxic indices. The reason is that a common physicochemical property governs the biological effect, namely the partition coefficient between two unmissible phases, simulated generally by n-octanol and water. This may mean that the transport of chemicals towards a target is responsible for the magnitude of the effect, rather than reactivity, as one would assume suppose. PMID:21331180
The Role of Motion Extrapolation in Amphibian Prey Capture
2015-01-01
Sensorimotor delays decouple behaviors from the events that drive them. The brain compensates for these delays with predictive mechanisms, but the efficacy and timescale over which these mechanisms operate remain poorly understood. Here, we assess how prediction is used to compensate for prey movement that occurs during visuomotor processing. We obtained high-speed video records of freely moving, tongue-projecting salamanders catching walking prey, emulating natural foraging conditions. We found that tongue projections were preceded by a rapid head turn lasting ∼130 ms. This motor lag, combined with the ∼100 ms phototransduction delay at photopic light levels, gave a ∼230 ms visuomotor response delay during which prey typically moved approximately one body length. Tongue projections, however, did not significantly lag prey position but were highly accurate instead. Angular errors in tongue projection accuracy were consistent with a linear extrapolation model that predicted prey position at the time of tongue contact using the average prey motion during a ∼175 ms period one visual latency before the head movement. The model explained successful strikes where the tongue hit the fly, and unsuccessful strikes where the fly turned and the tongue hit a phantom location consistent with the fly's earlier trajectory. The model parameters, obtained from the data, agree with the temporal integration and latency of retinal responses proposed to contribute to motion extrapolation. These results show that the salamander predicts future prey position and that prediction significantly improves prey capture success over a broad range of prey speeds and light levels. SIGNIFICANCE STATEMENT Neural processing delays cause actions to lag behind the events that elicit them. To cope with these delays, the brain predicts what will happen in the future. While neural circuits in the retina and beyond have been suggested to participate in such predictions, few behaviors have been
Frequency extrapolation by nonconvex compressive sensing
Chartrand, Rick; Sidky, Emil Y; Pan, Xiaochaun
2010-12-03
Tomographic imaging modalities sample subjects with a discrete, finite set of measurements, while the underlying object function is continuous. Because of this, inversion of the imaging model, even under ideal conditions, necessarily entails approximation. The error incurred by this approximation can be important when there is rapid variation in the object function or when the objects of interest are small. In this work, we investigate this issue with the Fourier transform (FT), which can be taken as the imaging model for magnetic resonance imaging (MRl) or some forms of wave imaging. Compressive sensing has been successful for inverting this data model when only a sparse set of samples are available. We apply the compressive sensing principle to a somewhat related problem of frequency extrapolation, where the object function is represented by a super-resolution grid with many more pixels than FT measurements. The image on the super-resolution grid is obtained through nonconvex minimization. The method fully utilizes the available FT samples, while controlling aliasing and ringing. The algorithm is demonstrated with continuous FT samples of the Shepp-Logan phantom with additional small, high-contrast objects.
Hard hadronic collisions: extrapolation of standard effects
Ali, A.; Aurenche, P.; Baier, R.; Berger, E.; Douiri, A.; Fontannaz, M.; Humpert, B.; Ingelman, G.; Kinnunen, R.; Pietarinen, E.
1984-01-01
We study hard hadronic collisions for the proton-proton (pp) and the proton-antiproton (p anti p) option in the CERN LEP tunnel. Based on our current knowledge of hard collisions at the present CERN p anti p Collider, and with the help of quantum chromodynamics (QCD), we extrapolate to the next generation of hadron colliders with a centre-of-mass energy E/sub cm/ = 10 to 20 TeV. We estimate various signatures, trigger rates, event topologies, and associated distributions for a variety of old and new physical processes, involving prompt photons, leptons, jets, W/sup + -/ and Z bosons in the final state. We also calculate the maximum fermion and boson masses accessible at the LEP Hadron Collider. The standard QCD and electroweak processes studied here, being the main body of standard hard collisions, quantify the challenge of extracting new physics with hadron colliders. We hope that our estimates will provide a useful profile of the final states, and that our experimental physics colleagues will find this of use in the design of their detectors. 84 references.
Crawford, D.J.; Richmond, C.R.
1980-01-01
The rationale underlying interspecies extrapolation of metabolic data has been based primarily on pragmatic concerns. Little attention has been given to the extent to which such extrapolations have a firm epistemological basis. The strength of this approach for model-free (purely empirical) extrapolation and for extrapolation involving a variety of theoretical constructs is examined in this paper. An attempt is made to provide some understanding of the degree of confidence that can be placed in the extrapolation of metabolic data from one species to another. Published results for a wide variety of radionuclides are analyzed and the importance of these results to the field of nuclear medicine is explored. Problems inherent in the logic of extrapolation are then delineated in view of these historical data.
CROSS-SPECIES DOSE EXTRAPOLATION FOR DIESEL EMISSIONS
Models for cross-species (rat to human) dose extrapolation of diesel emission were evaluated for purposes of establishing guidelines for human exposure to diesel emissions (DE) based on DE toxicological data obtained in rats. Ideally, a model for this extrapolation would provide...
Direct Extrapolation of Biota-sediment Accumulation Factors (BSAFs)
Biota-sediment accumulation factors (BSAFs) for fish and shellfish were extrapolated directly from one location and species to other species, to other locations within a site, to other sites, and their combinations. The median errors in the extrapolations across species at a loc...
Fully vectorial laser resonator modeling by vector extrapolation methods
NASA Astrophysics Data System (ADS)
Asoubar, Daniel; Kuhn, Michael; Wyrowski, Frank
2015-02-01
The optimization of multi-parameter resonators requires flexible simulation techniques beyond the scalar approximation. Therefore we generalize the scalar Fox and Li algorithm for the transversal eigenmode calculation to a fully vectorial model. This modified eigenvalue problem is solved by two polynomial-type vector extrapolation methods, namely the minimal polynomial extrapolation and the reduced rank extrapolation. Compared to other eigenvalue solvers these techniques can also be applied to resonators including nonlinear components. As an example we show the calculation of an azimuthally polarized eigenmode emitted by a resonator containing a discontinuous phase element and a nonlinear active medium. The simulation is verified by experiments.
An algorithm for a generalization of the Richardson extrapolation process
NASA Technical Reports Server (NTRS)
Ford, William F.; Sidi, Avram
1987-01-01
The paper presents a recursive method, designated the W exp (m)-algorithm, for implementing a generalization of the Richardson extrapolation process. Compared to the direct solution of the linear sytems of equations defining the extrapolation procedure, this method requires a small number of arithmetic operations and very little storage. The technique is also applied to solve recursively the coefficient problem associated with the rational approximations obtained by applying a d-transformation to power series. In the course of development a new recursive algorithm for implementing a very general extrapolation procedure is introduced, for solving the same problem. A FORTRAN program for the W exp (m)-algorithm is also appended.
3D Hail Size Distribution Interpolation/Extrapolation Algorithm
NASA Technical Reports Server (NTRS)
Lane, John
2013-01-01
Radar data can usually detect hail; however, it is difficult for present day radar to accurately discriminate between hail and rain. Local ground-based hail sensors are much better at detecting hail against a rain background, and when incorporated with radar data, provide a much better local picture of a severe rain or hail event. The previous disdrometer interpolation/ extrapolation algorithm described a method to interpolate horizontally between multiple ground sensors (a minimum of three) and extrapolate vertically. This work is a modification to that approach that generates a purely extrapolated 3D spatial distribution when using a single sensor.
Ecotoxicological effects assessment: A comparison of several extrapolation procedures
Okkerman, P.C.; v.d. Plassche, E.J.; Slooff, W.; Van Leeuwen, C.J.; Canton, J.H. , Bilthoven )
1991-04-01
In the future, extrapolation procedures will become more and more important for the effect assessment of compounds in aquatic systems. For achieving a reliable method these extrapolation procedures have to be evaluated thoroughly. As a first step three extrapolation procedures are compared by means of two sets of data, consisting of (semi)chronic and acute toxicity test results for 11 aquatic species and 8 compounds. Because of its statistical basis the extrapolation procedure of Van Straalen and Denneman is preferred over the procedures of the EPA and Stephan et al. The results of the calculations showed that lower numbers of toxicity data increase the chance of underestimating the risk of a compound. Therefore it is proposed to extend the OECD guidelines for algae, Daphnia, and fish with chronic (aquatic) toxicity tests for more species of different taxonomic groups.
Extrapolation technique pitfalls in asymmetry measurements at colliders
NASA Astrophysics Data System (ADS)
Colletti, Katrina; Hong, Ziqing; Toback, David; Wilson, Jonathan S.
2016-09-01
Asymmetry measurements are common in collider experiments and can sensitively probe particle properties. Typically, data can only be measured in a finite region covered by the detector, so an extrapolation from the visible asymmetry to the inclusive asymmetry is necessary. Often a constant multiplicative factor is advantageous for the extrapolation and this factor can be readily determined using simulation methods. However, there is a potential, avoidable pitfall involved in the determination of this factor when the asymmetry in the simulated data sample is small. We find that to obtain a reliable estimate of the extrapolation factor, the number of simulated events required rises as the inverse square of the simulated asymmetry; this can mean that an unexpectedly large sample size is required when determining the extrapolation factor.
NASA Astrophysics Data System (ADS)
Mirus, B. B.; Halford, K. J.; Sweetkind, D. S.; Fenelon, J.
2014-12-01
The utility of geologic frameworks for extrapolating hydraulic conductivities to length scales that are commensurate with hydraulic data has been assessed at the Nevada National Security Site in highly-faulted volcanic rocks. Observed drawdowns from eight, large-scale, aquifer tests on Pahute Mesa provided the necessary constraints to test assumed relations between hydraulic conductivity and interpretations of the geology. The investigated volume of rock encompassed about 40 cubic miles where drawdowns were detected more than 2 mi from pumping wells and traversed major fault structures. Five sets of hydraulic conductivities at about 500 pilot points were estimated by simultaneously interpreting all aquifer tests with a different geologic framework for each set. Each geologic framework was incorporated as prior information that assumed homogeneous hydraulic conductivities within each geologic unit. Complexity of the geologic frameworks ranged from an undifferentiated mass of rock with a single unit to 14 unique geologic units. Analysis of the model calibrations showed that a maximum of four geologic units could be differentiated where each was hydraulically unique as defined by the mean and standard deviation of log-hydraulic conductivity. Consistency of hydraulic property estimates within extents of investigation and effects of geologic frameworks on extrapolation were evaluated qualitatively with maps of transmissivity. Distributions of transmissivity were similar within the investigated extents regardless of geologic framework except for a transmissive streak along a fault in the Fault-Structure framework. Extrapolation was affected by underlying geologic frameworks where the variability of transmissivity increased as the number of units increased.
Implicit extrapolation methods for multilevel finite element computations
Jung, M.; Ruede, U.
1994-12-31
The finite element package FEMGP has been developed to solve elliptic and parabolic problems arising in the computation of magnetic and thermomechanical fields. FEMGP implements various methods for the construction of hierarchical finite element meshes, a variety of efficient multilevel solvers, including multigrid and preconditioned conjugate gradient iterations, as well as pre- and post-processing software. Within FEMGP, multigrid {tau}-extrapolation can be employed to improve the finite element solution iteratively to higher order. This algorithm is based on an implicit extrapolation, so that the algorithm differs from a regular multigrid algorithm only by a slightly modified computation of the residuals on the finest mesh. Another advantage of this technique is, that in contrast to explicit extrapolation methods, it does not rely on the existence of global error expansions, and therefore neither requires uniform meshes nor global regularity assumptions. In the paper the authors will analyse the {tau}-extrapolation algorithm and present experimental results in the context of the FEMGP package. Furthermore, the {tau}-extrapolation results will be compared to higher order finite element solutions.
Extrapolation of carcinogenicity between species: Qualitative and quantitative factors
Gold, L.S. Univ. of California, Berkeley ); Manley, N.B.; Ames, B.N. )
1992-12-01
Prediction of human cancer risk from the results of rodent bioassays requires two types of extrapolation: a qualitative extrapolation from short-lived rodent species to long-lived humans, and a quantitative extrapolation from near-toxic doses in the bioassay to low-level human exposures. Experimental evidence on the accuracy of prediction between closely related species tested under similar experimental conditions (rats, mice, and hamsters) indicates that: (1) if a chemical is positive in one species, it will be positive in the second species about 75% of the time; however, since about 50% of test chemicals are positive in each species, by chance alone one would expect a predictive value between species of about 50%. (2) If a chemical induces tumors in a particular target organ in one species, it will induce tumors in the same organ in the second species about 50% of the time. Similar predictive values are obtained in an analysis of prediction from humans to rats or from humans to mice for known human carcinogens. Limitations of bioassay data for use in quantitative extrapolation are discussed, including constraints on both estimates of carcinogenic potency and of the dose-response in experiments with only two doses and a control. Quantitative extrapolation should be based on an understanding of mechanisms of carcinogenesis, particularly mitogenic effects that are present at high and not low doses.
Rule-based extrapolation: a continuing challenge for exemplar models.
Denton, Stephen E; Kruschke, John K; Erickson, Michael A
2008-08-01
Erickson and Kruschke (1998, 2002) demonstrated that in rule-plus-exception categorization, people generalize category knowledge by extrapolating in a rule-like fashion, even when they are presented with a novel stimulus that is most similar to a known exception. Although exemplar models have been found to be deficient in explaining rule-based extrapolation, Rodrigues and Murre (2007) offered a variation of an exemplar model that was better able to account for such performance. Here, we present the results of a new rule-plus-exception experiment that yields rule-like extrapolation similar to that of previous experiments, and yet the data are not accounted for by Rodrigues and Murre's augmented exemplar model. Further, a hybrid rule-and-exemplar model is shown to better describe the data. Thus, we maintain that rule-plus-exception categorization continues to be a challenge for exemplar-only models. PMID:18792504
Chiral Extrapolation of Lattice Data for Heavy Meson Hyperfine Splittings
X.-H. Guo; P.C. Tandy; A.W. Thomas
2006-03-01
We investigate the chiral extrapolation of the lattice data for the light-heavy meson hyperfine splittings D*-D and B*-B to the physical region for the light quark mass. The chiral loop corrections providing non-analytic behavior in m{sub {pi}} are consistent with chiral perturbation theory for heavy mesons. Since chiral loop corrections tend to decrease the already too low splittings obtained from linear extrapolation, we investigate two models to guide the form of the analytic background behavior: the constituent quark potential model, and the covariant model of QCD based on the ladder-rainbow truncation of the Dyson-Schwinger equations. The extrapolated hyperfine splittings remain clearly below the experimental values even allowing for the model dependence in the description of the analytic background.
Efficient implementation of minimal polynomial and reduced rank extrapolation methods
NASA Technical Reports Server (NTRS)
Sidi, Avram
1990-01-01
The minimal polynomial extrapolation (MPE) and reduced rank extrapolation (RRE) are two effective techniques that have been used in accelerating the convergence of vector sequences, such as those that are obtained from iterative solution of linear and nonlinear systems of equation. Their definitions involve some linear least squares problems, and this causes difficulties in their numerical implementation. Timewise efficient and numerically stable implementations for MPE and RRE are developed. A computer program written in FORTRAN 77 is also appended and applied to some model problems.
Extrapolation method for the no-core shell model
NASA Astrophysics Data System (ADS)
Zhan, H.; Nogga, A.; Barrett, B. R.; Vary, J. P.; Navrátil, P.
2004-03-01
Nuclear many-body calculations are computationally demanding. An estimate of their accuracy is often hampered by the limited amount of computational resources even on present-day supercomputers. We provide an extrapolation method based on perturbation theory, so that the binding energy of a large basis-space calculation can be estimated without diagonalizing the Hamiltonian in this space. The extrapolation method is tested for 3H and 6Li nuclei. It will extend our computational abilities significantly and allow for reliable error estimates.
40 CFR 86.435-78 - Extrapolated emission values.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 40 Protection of Environment 18 2010-07-01 2010-07-01 false Extrapolated emission values. 86.435-78 Section 86.435-78 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NEW AND IN-USE HIGHWAY VEHICLES AND ENGINES Emission Regulations for 1978 and Later New Motorcycles,...
40 CFR 86.435-78 - Extrapolated emission values.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 40 Protection of Environment 18 2011-07-01 2011-07-01 false Extrapolated emission values. 86.435-78 Section 86.435-78 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NEW AND IN-USE HIGHWAY VEHICLES AND ENGINES Emission Regulations for 1978 and Later New Motorcycles,...
40 CFR 86.435-78 - Extrapolated emission values.
Code of Federal Regulations, 2013 CFR
2013-07-01
... 40 Protection of Environment 19 2013-07-01 2013-07-01 false Extrapolated emission values. 86.435-78 Section 86.435-78 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NEW AND IN-USE HIGHWAY VEHICLES AND ENGINES Emission Regulations for 1978 and Later New Motorcycles,...
40 CFR 86.435-78 - Extrapolated emission values.
Code of Federal Regulations, 2012 CFR
2012-07-01
... 40 Protection of Environment 19 2012-07-01 2012-07-01 false Extrapolated emission values. 86.435-78 Section 86.435-78 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NEW AND IN-USE HIGHWAY VEHICLES AND ENGINES Emission Regulations for 1978 and Later New Motorcycles,...
40 CFR 86.435-78 - Extrapolated emission values.
Code of Federal Regulations, 2014 CFR
2014-07-01
... 40 Protection of Environment 19 2014-07-01 2014-07-01 false Extrapolated emission values. 86.435-78 Section 86.435-78 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONTROL OF EMISSIONS FROM NEW AND IN-USE HIGHWAY VEHICLES AND ENGINES Emission Regulations for 1978 and Later New Motorcycles,...
Analytic Approximations for the Extrapolation of Lattice Data
Masjuan, Pere
2010-12-22
We present analytic approximations of chiral SU(3) amplitudes for the extrapolation of lattice data to the physical masses and the determination of Next-to-Next-to-Leading-Order low-energy constants. Lattice data for the ratio F{sub K}/F{sub {pi}} is used to test the method.
Properties of infrared extrapolations in a harmonic oscillator basis
NASA Astrophysics Data System (ADS)
Coon, Sidney A.; Kruse, Michael K. G.
2016-02-01
The success and utility of effective field theory (EFT) in explaining the structure and reactions of few-nucleon systems has prompted the initiation of EFT-inspired extrapolations to larger model spaces in ab initio methods such as the no-core shell model (NCSM). In this contribution, we review and continue our studies of infrared (ir) and ultraviolet (uv) regulators of NCSM calculations in which the input is phenomenological NN and NNN interactions fitted to data. We extend our previous findings that an extrapolation in the ir cutoff with the uv cutoff above the intrinsic uv scale of the interaction is quite successful, not only for the eigenstates of the Hamiltonian but also for expectation values of operators, such as r2, considered long range. The latter results are obtained with Hamiltonians transformed by the similarity renormalization group (SRG) evolution. On the other hand, a possible extrapolation of ground state energies in the uv cutoff when the ir cutoff is below the intrinsic ir scale is not robust and does not agree with the ir extrapolation of the same data or with independent calculations using other methods.
MODELING TOXICOKINETICS FOR CROSS-SPECIES EXTRAPOLATION OF DEVELOPMENTAL EFFECTS
Animal toxicology studies used to evaluate the potential for effects due to exposures during developmental periods are extrapolated to humans based upon the maternal exposure dose. The approach does not address whether the toxicokinetics are similar across species during the rel...
MULTIPLE SOLVENT EXPOSURE IN HUMANS: CROSS-SPECIES EXTRAPOLATIONS
Multiple Solvent Exposures in Humans:
Cross-Species Extrapolations
(Future Research Plan)
Vernon A. Benignus1, Philip J. Bushnell2 and William K. Boyes2
A few solvents can be safely studied in acute experiments in human subjects. Data exist in rats f...
Parallel solution of partial differential equations by extrapolation methods
Leland, Robert W.; Rollett, J. S.
2015-02-01
We have found, in the ROGE algorithm, an extrapolation process which is robust, effective and practically simple to implement. It removes the difficulty of needing to make a precise estimate of the over-relaxation parameter for Successive Over-Relaxation (SOR) type methods.
An objective analysis technique for extrapolating tidal fields
NASA Technical Reports Server (NTRS)
Sanchez, B. V.
1984-01-01
An interpolation technique which allows accurate extrapolation of tidal height fields in the ocean basins by making use of selected satellite altimetry measurements and/or conventional gauge measurements was developed and tested. A normal mode solution for the Atlantic and Indian Oceans was obtained by means of a finite difference grid. Normal mode amplitude maps are presented.
NASA Astrophysics Data System (ADS)
Mirus, Benjamin B.; Halford, Keith; Sweetkind, Don; Fenelon, Joe
2016-02-01
The suitability of geologic frameworks for extrapolating hydraulic conductivity (K) to length scales commensurate with hydraulic data is difficult to assess. A novel method is presented for evaluating assumed relations between K and geologic interpretations for regional-scale groundwater modeling. The approach relies on simultaneous interpretation of multiple aquifer tests using alternative geologic frameworks of variable complexity, where each framework is incorporated as prior information that assumes homogeneous K within each model unit. This approach is tested at Pahute Mesa within the Nevada National Security Site (USA), where observed drawdowns from eight aquifer tests in complex, highly faulted volcanic rocks provide the necessary hydraulic constraints. The investigated volume encompasses 40 mi3 (167 km3) where drawdowns traversed major fault structures and were detected more than 2 mi (3.2 km) from pumping wells. Complexity of the five frameworks assessed ranges from an undifferentiated mass of rock with a single unit to 14 distinct geologic units. Results show that only four geologic units can be justified as hydraulically unique for this location. The approach qualitatively evaluates the consistency of hydraulic property estimates within extents of investigation and effects of geologic frameworks on extrapolation. Distributions of transmissivity are similar within the investigated extents irrespective of the geologic framework. In contrast, the extrapolation of hydraulic properties beyond the volume investigated with interfering aquifer tests is strongly affected by the complexity of a given framework. Testing at Pahute Mesa illustrates how this method can be employed to determine the appropriate level of geologic complexity for large-scale groundwater modeling.
Mirus, Benjamin B.; Halford, Keith J.; Sweetkind, Donald; Fenelon, Joseph M.
2016-01-01
The suitability of geologic frameworks for extrapolating hydraulic conductivity (K) to length scales commensurate with hydraulic data is difficult to assess. A novel method is presented for evaluating assumed relations between K and geologic interpretations for regional-scale groundwater modeling. The approach relies on simultaneous interpretation of multiple aquifer tests using alternative geologic frameworks of variable complexity, where each framework is incorporated as prior information that assumes homogeneous K within each model unit. This approach is tested at Pahute Mesa within the Nevada National Security Site (USA), where observed drawdowns from eight aquifer tests in complex, highly faulted volcanic rocks provide the necessary hydraulic constraints. The investigated volume encompasses 40 mi3 (167 km3) where drawdowns traversed major fault structures and were detected more than 2 mi (3.2 km) from pumping wells. Complexity of the five frameworks assessed ranges from an undifferentiated mass of rock with a single unit to 14 distinct geologic units. Results show that only four geologic units can be justified as hydraulically unique for this location. The approach qualitatively evaluates the consistency of hydraulic property estimates within extents of investigation and effects of geologic frameworks on extrapolation. Distributions of transmissivity are similar within the investigated extents irrespective of the geologic framework. In contrast, the extrapolation of hydraulic properties beyond the volume investigated with interfering aquifer tests is strongly affected by the complexity of a given framework. Testing at Pahute Mesa illustrates how this method can be employed to determine the appropriate level of geologic complexity for large-scale groundwater modeling.
NASA Astrophysics Data System (ADS)
Mirus, Benjamin B.; Halford, Keith; Sweetkind, Don; Fenelon, Joe
2016-08-01
The suitability of geologic frameworks for extrapolating hydraulic conductivity ( K) to length scales commensurate with hydraulic data is difficult to assess. A novel method is presented for evaluating assumed relations between K and geologic interpretations for regional-scale groundwater modeling. The approach relies on simultaneous interpretation of multiple aquifer tests using alternative geologic frameworks of variable complexity, where each framework is incorporated as prior information that assumes homogeneous K within each model unit. This approach is tested at Pahute Mesa within the Nevada National Security Site (USA), where observed drawdowns from eight aquifer tests in complex, highly faulted volcanic rocks provide the necessary hydraulic constraints. The investigated volume encompasses 40 mi3 (167 km3) where drawdowns traversed major fault structures and were detected more than 2 mi (3.2 km) from pumping wells. Complexity of the five frameworks assessed ranges from an undifferentiated mass of rock with a single unit to 14 distinct geologic units. Results show that only four geologic units can be justified as hydraulically unique for this location. The approach qualitatively evaluates the consistency of hydraulic property estimates within extents of investigation and effects of geologic frameworks on extrapolation. Distributions of transmissivity are similar within the investigated extents irrespective of the geologic framework. In contrast, the extrapolation of hydraulic properties beyond the volume investigated with interfering aquifer tests is strongly affected by the complexity of a given framework. Testing at Pahute Mesa illustrates how this method can be employed to determine the appropriate level of geologic complexity for large-scale groundwater modeling.
Mirus, Benjamin B.; Halford, Keith J.; Sweetkind, Donald; Fenelon, Joseph M.
2016-02-18
The suitability of geologic frameworks for extrapolating hydraulic conductivity (K) to length scales commensurate with hydraulic data is difficult to assess. A novel method is presented for evaluating assumed relations between K and geologic interpretations for regional-scale groundwater modeling. The approach relies on simultaneous interpretation of multiple aquifer tests using alternative geologic frameworks of variable complexity, where each framework is incorporated as prior information that assumes homogeneous K within each model unit. This approach is tested at Pahute Mesa within the Nevada National Security Site (USA), where observed drawdowns from eight aquifer tests in complex, highly faulted volcanic rocks providemore » the necessary hydraulic constraints. The investigated volume encompasses 40 mi3 (167 km3) where drawdowns traversed major fault structures and were detected more than 2 mi (3.2 km) from pumping wells. Complexity of the five frameworks assessed ranges from an undifferentiated mass of rock with a single unit to 14 distinct geologic units. Results show that only four geologic units can be justified as hydraulically unique for this location. The approach qualitatively evaluates the consistency of hydraulic property estimates within extents of investigation and effects of geologic frameworks on extrapolation. Distributions of transmissivity are similar within the investigated extents irrespective of the geologic framework. In contrast, the extrapolation of hydraulic properties beyond the volume investigated with interfering aquifer tests is strongly affected by the complexity of a given framework. As a result, testing at Pahute Mesa illustrates how this method can be employed to determine the appropriate level of geologic complexity for large-scale groundwater modeling.« less
Extrapolation techniques applied to matrix methods in neutron diffusion problems
NASA Technical Reports Server (NTRS)
Mccready, Robert R
1956-01-01
A general matrix method is developed for the solution of characteristic-value problems of the type arising in many physical applications. The scheme employed is essentially that of Gauss and Seidel with appropriate modifications needed to make it applicable to characteristic-value problems. An iterative procedure produces a sequence of estimates to the answer; and extrapolation techniques, based upon previous behavior of iterants, are utilized in speeding convergence. Theoretically sound limits are placed on the magnitude of the extrapolation that may be tolerated. This matrix method is applied to the problem of finding criticality and neutron fluxes in a nuclear reactor with control rods. The two-dimensional finite-difference approximation to the two-group neutron fluxes in a nuclear reactor with control rods. The two-dimensional finite-difference approximation to the two-group neutron-diffusion equations is treated. Results for this example are indicated.
Survival extrapolation using the poly-Weibull model
Lunn, David; Sharples, Linda D
2015-01-01
Recent studies of (cost-) effectiveness in cardiothoracic transplantation have required estimation of mean survival over the lifetime of the recipients. In order to calculate mean survival, the complete survivor curve is required but is often not fully observed, so that survival extrapolation is necessary. After transplantation, the hazard function is bathtub-shaped, reflecting latent competing risks which operate additively in overlapping time periods. The poly-Weibull distribution is a flexible parametric model that may be used to extrapolate survival and has a natural competing risks interpretation. In addition, treatment effects and subgroups can be modelled separately for each component of risk. We describe the model and develop inference procedures using freely available software. The methods are applied to two problems from cardiothoracic transplantation. PMID:21937472
Survival extrapolation using the poly-Weibull model.
Demiris, Nikolaos; Lunn, David; Sharples, Linda D
2015-04-01
Recent studies of (cost-) effectiveness in cardiothoracic transplantation have required estimation of mean survival over the lifetime of the recipients. In order to calculate mean survival, the complete survivor curve is required but is often not fully observed, so that survival extrapolation is necessary. After transplantation, the hazard function is bathtub-shaped, reflecting latent competing risks which operate additively in overlapping time periods. The poly-Weibull distribution is a flexible parametric model that may be used to extrapolate survival and has a natural competing risks interpretation. In addition, treatment effects and subgroups can be modelled separately for each component of risk. We describe the model and develop inference procedures using freely available software. The methods are applied to two problems from cardiothoracic transplantation. PMID:21937472
Extrapolated gradientlike algorithms for molecular dynamics and celestial mechanics simulations.
Omelyan, I P
2006-09-01
A class of symplectic algorithms is introduced to integrate the equations of motion in many-body systems. The algorithms are derived on the basis of an advanced gradientlike decomposition approach. Its main advantage over the standard gradient scheme is the avoidance of time-consuming evaluations of force gradients by force extrapolation without any loss of precision. As a result, the efficiency of the integration improves significantly. The algorithms obtained are analyzed and optimized using an error-function theory. The best among them are tested in actual molecular dynamics and celestial mechanics simulations for comparison with well-known nongradient and gradient algorithms such as the Störmer-Verlet, Runge-Kutta, Cowell-Numerov, Forest-Ruth, Suzuki-Chin, and others. It is demonstrated that for moderate and high accuracy, the extrapolated algorithms should be considered as the most efficient for the integration of motion in molecular dynamics simulations. PMID:17025782
Population movement studied at microscale: experience and extrapolation.
Chapman, M
1987-12-01
The context for this paper is the nature of generalization deriving from the study of particular cases of human behavior. With specific reference to field research on population movement conducted among individuals, households, small groups, and village communities in 3rd world societies, it challenges the convention that both generalization and extrapolation are based inevitably and exclusively on the number of events subject to examination. An evaluation is made of the methodological aspects of 4 different studies of population mobility at microscale, undertaken between 1965 and 1977 in the Solomon Islands and northwest Thailand. On this basis, integrated field designs that incorporate a range of intersecting instruments are favored for their technical flexibility and logical strength. With case studies of 3rd world villages, market centers, and urban neighborhoods, generalization and extrapolation is based on depth of understanding and power of theoretical connections. PMID:12315702
Extrapolation from occupational studies: a substitute for environmental epidemiology.
Enterline, P E
1981-01-01
Extrapolation from occupational data to general environmental exposures gives some interesting results, and these results might be useful in our decision-making process. These results could never be observed by environmental epidemiology and this method probably represents the only way of quantifying the health effects of low-exposure levels. Three linear models for extrapolating to low levels are presented--one from Canadian data, one from American data and one from British data. One or more of these is applied to two recently publicized asbestos exposures; exposures resulting from asbestos heat shields in hair dryers and exposures in public school buildings. Predictions are derived as to the effects of asbestos exposures on cancer mortality. A comparison is made between predictions made on the basis of a linear and nonlinear model. PMID:7333259
An efficient method to evaluate energy variances for extrapolation methods
NASA Astrophysics Data System (ADS)
Puddu, G.
2012-08-01
The energy variance extrapolation method consists of relating the approximate energies in many-body calculations to the corresponding energy variances and inferring eigenvalues by extrapolating to zero variance. The method needs a fast evaluation of the energy variances. For many-body methods that expand the nuclear wavefunctions in terms of deformed Slater determinants, the best available method for the evaluation of energy variances scales with the sixth power of the number of single-particle states. We propose a new method which depends on the number of single-particle orbits and the number of particles rather than the number of single-particle states. We discuss as an example the case of 4He using the chiral N3LO interaction in a basis consisting up to 184 single-particle states.
Limitations on wind-tunnel pressure signature extrapolation
NASA Technical Reports Server (NTRS)
Mack, Robert J.; Darden, Christine M.
1992-01-01
Analysis of some recent experimental sonic boom data has revived the hypothesis that there is a closeness limit to the near-field separation distance from which measured wind tunnel pressure signatures can be extrapolated to the ground as though generated by a supersonic-cruise aircraft. Geometric acoustic theory is used to derive an estimate of this distance and the sample data is used to provide a preliminary indication of practical separation distance values.
Chiral extrapolation of the X(3872) binding energy
NASA Astrophysics Data System (ADS)
Baru, V.; Epelbaum, E.; Filin, A. A.; Gegelia, J.; Nefediev, A. V.
2016-02-01
The role of pion dynamics in the X(3872) charmonium-like state is studied in the framework of a renormalisable effective quantum field theory approach and they are found to play a substantial role in the formation of the X. Chiral extrapolation from the physical point to unphysically large pion masses is performed and the results are confronted with the lattice predictions. The proposed approach overrides the gap between the lattice calculations and the physical limit in mπ.
An efficient extrapolation to the (T)/CBS limit
NASA Astrophysics Data System (ADS)
Ranasinghe, Duminda S.; Barnes, Ericka C.
2014-05-01
We extrapolate to the perturbative triples (T)/complete basis set (CBS) limit using double ζ basis sets without polarization functions (Wesleyan-1-Triples-2ζ or "Wes1T-2Z") and triple ζ basis sets with a single level of polarization functions (Wesleyan-1-Triples-3ζ or "Wes1T-3Z"). These basis sets were optimized for 102 species representing the first two rows of the Periodic Table. The species include the entire set of neutral atoms, positive and negative atomic ions, as well as several homonuclear diatomic molecules, hydrides, rare gas dimers, polar molecules, such as oxides and fluorides, and a few transition states. The extrapolated Wes1T-(2,3)Z triples energies agree with (T)/CBS benchmarks to within ±0.65 mEh, while the rms deviations of comparable model chemistries W1, CBS-APNO, and CBS-QB3 for the same test set are ±0.23 mEh, ±2.37 mEh, and ±5.80 mEh, respectively. The Wes1T-(2,3)Z triples calculation time for the largest hydrocarbon in the G2/97 test set, C6H5Me+, is reduced by a factor of 25 when compared to W1. The cost-effectiveness of the Wes1T-(2,3)Z extrapolation validates the usefulness of the Wes1T-2Z and Wes1T-3Z basis sets which are now available for a more efficient extrapolation of the (T) component of any composite model chemistry.
An efficient extrapolation to the (T)/CBS limit
Ranasinghe, Duminda S.; Barnes, Ericka C.
2014-05-14
We extrapolate to the perturbative triples (T)/complete basis set (CBS) limit using double ζ basis sets without polarization functions (Wesleyan-1-Triples-2ζ or “Wes1T-2Z”) and triple ζ basis sets with a single level of polarization functions (Wesleyan-1-Triples-3ζ or “Wes1T-3Z”). These basis sets were optimized for 102 species representing the first two rows of the Periodic Table. The species include the entire set of neutral atoms, positive and negative atomic ions, as well as several homonuclear diatomic molecules, hydrides, rare gas dimers, polar molecules, such as oxides and fluorides, and a few transition states. The extrapolated Wes1T-(2,3)Z triples energies agree with (T)/CBS benchmarks to within ±0.65 mE{sub h}, while the rms deviations of comparable model chemistries W1, CBS-APNO, and CBS-QB3 for the same test set are ±0.23 mE{sub h}, ±2.37 mE{sub h}, and ±5.80 mE{sub h}, respectively. The Wes1T-(2,3)Z triples calculation time for the largest hydrocarbon in the G2/97 test set, C{sub 6}H{sub 5}Me{sup +}, is reduced by a factor of 25 when compared to W1. The cost-effectiveness of the Wes1T-(2,3)Z extrapolation validates the usefulness of the Wes1T-2Z and Wes1T-3Z basis sets which are now available for a more efficient extrapolation of the (T) component of any composite model chemistry.
A simple extrapolation of thermodynamic perturbation theory to infinite order.
Ghobadi, Ahmadreza F; Elliott, J Richard
2015-09-21
Recent analyses of the third and fourth order perturbation contributions to the equations of state for square well spheres and Lennard-Jones chains show trends that persist across orders and molecular models. In particular, the ratio between orders (e.g., A3/A2, where A(i) is the ith order perturbation contribution) exhibits a peak when plotted with respect to density. The trend resembles a Gaussian curve with the peak near the critical density. This observation can form the basis for a simple recursion and extrapolation from the highest available order to infinite order. The resulting extrapolation is analytic and therefore cannot fully characterize the critical region, but it remarkably improves accuracy, especially for the binodal curve. Whereas a second order theory is typically accurate for the binodal at temperatures within 90% of the critical temperature, the extrapolated result is accurate to within 99% of the critical temperature. In addition to square well spheres and Lennard-Jones chains, we demonstrate how the method can be applied semi-empirically to the Perturbed Chain - Statistical Associating Fluid Theory (PC-SAFT). PMID:26395687
A simple extrapolation of thermodynamic perturbation theory to infinite order
Ghobadi, Ahmadreza F.; Elliott, J. Richard
2015-09-21
Recent analyses of the third and fourth order perturbation contributions to the equations of state for square well spheres and Lennard-Jones chains show trends that persist across orders and molecular models. In particular, the ratio between orders (e.g., A{sub 3}/A{sub 2}, where A{sub i} is the ith order perturbation contribution) exhibits a peak when plotted with respect to density. The trend resembles a Gaussian curve with the peak near the critical density. This observation can form the basis for a simple recursion and extrapolation from the highest available order to infinite order. The resulting extrapolation is analytic and therefore cannot fully characterize the critical region, but it remarkably improves accuracy, especially for the binodal curve. Whereas a second order theory is typically accurate for the binodal at temperatures within 90% of the critical temperature, the extrapolated result is accurate to within 99% of the critical temperature. In addition to square well spheres and Lennard-Jones chains, we demonstrate how the method can be applied semi-empirically to the Perturbed Chain - Statistical Associating Fluid Theory (PC-SAFT)
Determination of Extrapolation Distance with Measured Pressure Signatures from Two Low-Boom Models
NASA Technical Reports Server (NTRS)
Mack, Robert J.; Kuhn, Neil
2004-01-01
A study to determine a limiting distance to span ratio for the extrapolation of near-field pressure signatures is described and discussed. This study was to be done in two wind-tunnel facilities with two wind-tunnel models. At this time, only the first half had been completed, so the scope of this report is limited to the design of the models, and to an analysis of the first set of measured pressure signatures. The results from this analysis showed that the pressure signatures measured at separation distances of 2 to 5 span lengths did not show the desired low-boom shapes. However, there were indications that the pressure signature shapes were becoming 'flat-topped'. This trend toward a 'flat-top' pressure signatures shape was seen to be a gradual one at the distance ratios employed in this first series of wind-tunnel tests.
Villemot, François; Capelli, Riccardo; Colombo, Giorgio; van der Vaart, Arjan
2016-06-14
Improvements to the confinement method for the calculation of conformational free energy differences are presented. By taking advantage of phase space overlap between simulations at different frequencies, significant gains in accuracy and speed are reached. The optimal frequency spacing for the simulations is obtained from extrapolations of the confinement energy, and relaxation time analysis is used to determine time steps, simulation lengths, and friction coefficients. At postprocessing, interpolation of confinement energies is used to significantly reduce discretization errors in the calculation of conformational free energies. The efficiency of this protocol is illustrated by applications to alanine n-peptides and lactoferricin. For the alanine-n-peptide, errors were reduced between 2- and 10-fold and sampling times between 8- and 67-fold, while for lactoferricin the long sampling times at low frequencies were reduced 10-100-fold. PMID:27120438
Verwichte, E.; Foullon, C.; White, R. S.; Van Doorsselaere, T.
2013-04-10
Two transversely oscillating coronal loops are investigated in detail during a flare on the 2011 September 6 using data from the Atmospheric Imaging Assembly (AIA) on board the Solar Dynamics Observatory. We compare two independent methods to determine the Alfven speed inside these loops. Through the period of oscillation and loop length, information about the Alfven speed inside each loop is deduced seismologically. This is compared with the Alfven speed profiles deduced from magnetic extrapolation and spectral methods using AIA bandpass. We find that for both loops the two methods are consistent. Also, we find that the average Alfven speed based on loop travel time is not necessarily a good measure to compare with the seismological result, which explains earlier reported discrepancies. Instead, the effect of density and magnetic stratification on the wave mode has to be taken into account. We discuss the implications of combining seismological, extrapolation, and spectral methods in deducing the physical properties of coronal loops.
Extrapolated HPGe efficiency estimates based on a single calibration measurement
Winn, W.G.
1994-07-01
Gamma spectroscopists often must analyze samples with geometries for which their detectors are not calibrated. The effort to experimentally recalibrate a detector for a new geometry can be quite time consuming, causing delay in reporting useful results. Such concerns have motivated development of a method for extrapolating HPGe efficiency estimates from an existing single measured efficiency. Overall, the method provides useful preliminary results for analyses that do not require exceptional accuracy, while reliably bracketing the credible range. The estimated efficiency {element_of} for a uniform sample in a geometry with volume V is extrapolated from the measured {element_of}{sub 0} of the base sample of volume V{sub 0}. Assuming all samples are centered atop the detector for maximum efficiency, {element_of} decreases monotonically as V increases about V{sub 0}, and vice versa. Extrapolation of high and low efficiency estimates {element_of}{sub h} and {element_of}{sub L} provides an average estimate of {element_of} = 1/2 [{element_of}{sub h} + {element_of}{sub L}] {plus_minus} 1/2 [{element_of}{sub h} {minus} {element_of}{sub L}] (general) where an uncertainty D{element_of} = 1/2 ({element_of}{sub h} {minus} {element_of}{sub L}] brackets limits for a maximum possible error. The {element_of}{sub h} and {element_of}{sub L} both diverge from {element_of}{sub 0} as V deviates from V{sub 0}, causing D{element_of} to increase accordingly. The above concepts guided development of both conservative and refined estimates for {element_of}.
Smooth extrapolation of unknown anatomy via statistical shape models
NASA Astrophysics Data System (ADS)
Grupp, R. B.; Chiang, H.; Otake, Y.; Murphy, R. J.; Gordon, C. R.; Armand, M.; Taylor, R. H.
2015-03-01
Several methods to perform extrapolation of unknown anatomy were evaluated. The primary application is to enhance surgical procedures that may use partial medical images or medical images of incomplete anatomy. Le Fort-based, face-jaw-teeth transplant is one such procedure. From CT data of 36 skulls and 21 mandibles separate Statistical Shape Models of the anatomical surfaces were created. Using the Statistical Shape Models, incomplete surfaces were projected to obtain complete surface estimates. The surface estimates exhibit non-zero error in regions where the true surface is known; it is desirable to keep the true surface and seamlessly merge the estimated unknown surface. Existing extrapolation techniques produce non-smooth transitions from the true surface to the estimated surface, resulting in additional error and a less aesthetically pleasing result. The three extrapolation techniques evaluated were: copying and pasting of the surface estimate (non-smooth baseline), a feathering between the patient surface and surface estimate, and an estimate generated via a Thin Plate Spline trained from displacements between the surface estimate and corresponding vertices of the known patient surface. Feathering and Thin Plate Spline approaches both yielded smooth transitions. However, feathering corrupted known vertex values. Leave-one-out analyses were conducted, with 5% to 50% of known anatomy removed from the left-out patient and estimated via the proposed approaches. The Thin Plate Spline approach yielded smaller errors than the other two approaches, with an average vertex error improvement of 1.46 mm and 1.38 mm for the skull and mandible respectively, over the baseline approach.
Acute toxicity value extrapolation with fish and aquatic invertebrates
Buckler, D.R.; Mayer, F.L.; Ellersieck, Mark R.; Asfaw, A.
2005-01-01
Assessment of risk posed by an environmental contaminant to an aquatic community requires estimation of both its magnitude of occurrence (exposure) and its ability to cause harm (effects). Our ability to estimate effects is often hindered by limited toxicological information. As a result, resource managers and environmental regulators are often faced with the need to extrapolate across taxonomic groups in order to protect the more sensitive members of the aquatic community. The goals of this effort were to 1) compile and organize an extensive body of acute toxicity data, 2) characterize the distribution of toxicant sensitivity across taxa and species, and 3) evaluate the utility of toxicity extrapolation methods based upon sensitivity relations among species and chemicals. Although the analysis encompassed a wide range of toxicants and species, pesticides and freshwater fish and invertebrates were emphasized as a reflection of available data. Although it is obviously desirable to have high-quality acute toxicity values for as many species as possible, the results of this effort allow for better use of available information for predicting the sensitivity of untested species to environmental contaminants. A software program entitled "Ecological Risk Analysis" (ERA) was developed that predicts toxicity values for sensitive members of the aquatic community using species sensitivity distributions. Of several methods evaluated, the ERA program used with minimum data sets comprising acute toxicity values for rainbow trout, bluegill, daphnia, and mysids provided the most satisfactory predictions with the least amount of data. However, if predictions must be made using data for a single species, the most satisfactory results were obtained with extrapolation factors developed for rainbow trout (0.412), bluegill (0.331), or scud (0.041). Although many specific exceptions occur, our results also support the conventional wisdom that invertebrates are generally more sensitive to
Extrapolation of vertical target motion through a brief visual occlusion.
Zago, Myrka; Iosa, Marco; Maffei, Vincenzo; Lacquaniti, Francesco
2010-03-01
It is known that arbitrary target accelerations along the horizontal generally are extrapolated much less accurately than target speed through a visual occlusion. The extent to which vertical accelerations can be extrapolated through an occlusion is much less understood. Here, we presented a virtual target rapidly descending on a blank screen with different motion laws. The target accelerated under gravity (1g), decelerated under reversed gravity (-1g), or moved at constant speed (0g). Probability of each type of acceleration differed across experiments: one acceleration at a time, or two to three different accelerations randomly intermingled could be presented. After a given viewing period, the target disappeared for a brief, variable period until arrival (occluded trials) or it remained visible throughout (visible trials). Subjects were asked to press a button when the target arrived at destination. We found that, in visible trials, the average performance with 1g targets could be better or worse than that with 0g targets depending on the acceleration probability, and both were always superior to the performance with -1g targets. By contrast, the average performance with 1g targets was always superior to that with 0g and -1g targets in occluded trials. Moreover, the response times of 1g trials tended to approach the ideal value with practice in occluded protocols. To gain insight into the mechanisms of extrapolation, we modeled the response timing based on different types of threshold models. We found that occlusion was accompanied by an adaptation of model parameters (threshold time and central processing time) in a direction that suggests a strategy oriented to the interception of 1g targets at the expense of the interception of the other types of tested targets. We argue that the prediction of occluded vertical motion may incorporate an expectation of gravity effects. PMID:19882150
Optical tomography by the temporally extrapolated absorbance method
NASA Astrophysics Data System (ADS)
Oda, Ichiro; Eda, Hideo; Tsunazawa, Yoshio; Takada, Michinosuke; Yamada, Yukio; Nishimura, Goro; Tamura, Mamoru
1996-01-01
The concept of the temporally extrapolated absorbance method (TEAM) for optical tomography of turbid media has been verified by fundamental experiments and image reconstruction. The TEAM uses the time-resolved spectroscopic data of the reference and object to provide projection data that are processed by conventional backprojection. Optical tomography images of a phantom consisting of axisymmetric double cylinders were experimentally obtained with the TEAM and time-gating and continuous-wave (CW) methods. The reconstructed TEAM images are compared with those obtained with the time-gating and CW methods and are found to have better spatial resolution.
A critical note on extrapolated helium pair potentials
NASA Astrophysics Data System (ADS)
Klopper, Wim
2001-07-01
It is difficult, if not impossible, to extrapolate the helium pair potential to the limit of a complete basis to within the accuracy needed to improve significantly on existing, directly computed potentials. Even though the basis-set convergence of calculations in a correlation-consistent basis with cardinal number X is dominated by the X-3 term, it is important to account for energy terms that converge more rapidly than ∝X-3. The electron-correlation contribution to the potential will be overestimated noticeably when these terms are not properly taken into account.
Nonlinear Force-Free Field Extrapolation of NOAA AR 0696
NASA Astrophysics Data System (ADS)
Thalmann, J. K.; Wiegelmann, T.
2007-12-01
We investigate the 3D coronal magnetic field structure of NOAA AR 0696 in the period of November 09-11, 2004, before and after an X2.5 flare (occurring around 02:13 UT on November 10, 2004). The coronal magnetic field dominates the structure of the solar corona and consequently plays a key role for the understanding of the initiation of flares. The most accurate presently available method to derive the coronal magnetic field is currently the nonlinear force-free field extrapolation from measurements of the photospheric magnetic field vector. These vector-magnetograms were processed from stokes I, Q, U, and V measurements of the Big Bear Solar Observatory and extrapolated into the corona with the nonlinear force-free optimization code developed by Wiegelmann (2004). We analyze the corresponding time series of coronal equilibria regarding topology changes of the 3D coronal magnetic field during the flare. Furthermore, quantities such as the temporal evolution of the magnetic energy and helicity are computed.
Extrapolating Single Organic Ion Solvation Thermochemistry from Simulated Water Nanodroplets.
Coles, Jonathan P; Houriez, Céline; Meot-Ner Mautner, Michael; Masella, Michel
2016-09-01
We compute the ion/water interaction energies of methylated ammonium cations and alkylated carboxylate anions solvated in large nanodroplets of 10 000 water molecules using 10 ns molecular dynamics simulations and an all-atom polarizable force-field approach. Together with our earlier results concerning the solvation of these organic ions in nanodroplets whose molecular sizes range from 50 to 1000, these new data allow us to discuss the reliability of extrapolating absolute single-ion bulk solvation energies from small ion/water droplets using common power-law functions of cluster size. We show that reliable estimates of these energies can be extrapolated from a small data set comprising the results of three droplets whose sizes are between 100 and 1000 using a basic power-law function of droplet size. This agrees with an earlier conclusion drawn from a model built within the mean spherical framework and paves the road toward a theoretical protocol to systematically compute the solvation energies of complex organic ions. PMID:27420562
Extrapolating W -associated jet-production ratios at the LHC
NASA Astrophysics Data System (ADS)
Bern, Z.; Dixon, L. J.; Febres Cordero, F.; Höche, S.; Ita, H.; Kosower, D. A.; Maître, D.
2015-07-01
Electroweak vector-boson production, accompanied by multiple jets, is an important background to searches for physics beyond the standard model. A precise and quantitative understanding of this process is helpful in constraining deviations from known physics. We study four key ratios in W +n -jet production at the LHC. We compute the ratio of cross sections for W +n - to W +(n -1 )-jet production as a function of the minimum jet transverse momentum. We also study the ratio differentially, as a function of the W -boson transverse momentum; as a function of the scalar sum of the jet transverse energy, HTjets; and as a function of certain jet transverse momenta. We show how to use such ratios to extrapolate differential cross sections to W +6 -jet production at next-to-leading order, and we cross-check the method against a direct calculation at leading order. We predict the differential distribution in HTjets for W +6 jets at next-to-leading order using such an extrapolation. We use the BlackHat software library together with SHERPA to perform the computations.
Estimation of macro velocity models by wave field extrapolation
NASA Astrophysics Data System (ADS)
Cox, Hendricus Lambertus Hubertus
A method to estimate accurate macro velocity models for prediction of traveltimes of seismic waves in the earth's subsurface is developed. The sensitivity of prestack migration is used to estimate the model and since model errors are expressed in the quality of the migration result, the migration process itself can be used to determine these errors. Using an initial model, shot records are downward extrapolated to grid points (depth points) in the subsurface. The extrapolated data can be reordered into so called common depth point (CDP) gathers, image gathers and focus panels. The deviation from horizontal alignment is used to quantify the errors in the model and to apply update corrections accordingly. The analysis can be done before or after stacking over all shot records (CDP stacking). the previously mentioned focus panels are generated by CDP stacking. The alignment analysis reduces then to a simple focusing analysis. The examples discussed show that horizontal alignment gives accurate macro velocity models for prestack depth migration. Focus panels can be difficult to interpret in complicated situations, where it is impossible to converge to the correct solution with focus panels only. The process should be guided by macrogeologic models of the area. In complicated situations, a layer stripping strategy is preferred.
California's Proposition 65: extrapolating animal toxicity to humans.
Kilgore, W W
1990-01-01
In 1986, the voters of California passed a law regarding the concept of extrapolating animal toxicity data to humans. The California Safe Drinking Water and Toxic Enforcement Act of 1986, known as Proposition 65, does five things: 1. It creates a list of chemicals (including a number of agricultural chemicals) known to cause cancer or reproductive toxicity; 2. It limits discharges of listed chemicals to drinking water sources; 3. It requires prior warning before exposure to listed chemicals by anyone in the course of doing business; 4. It creates a list of chemicals requiring testing for carcinogenicity or reproductive toxicity; and 5. It requires the Governor to consult with qualified experts (a 12-member "Scientific Advisory Panel" was appointed) as necessary to carry out his duties. This paper discusses the details and implications of this proposition. Areas of responsibility have been assigned. The definition of significant risk is being addressed. PMID:2248253
Survival Extrapolation in the Presence of Cause Specific Hazards
Benaglia, Tatiana; Jackson, Christopher H.; Sharples, Linda D.
2016-01-01
Health economic evaluations require estimates of expected survival from patients receiving different interventions, often over a lifetime. However, data on the patients of interest are typically only available for a much shorter follow-up time, from randomised trials or cohorts. Previous work showed how to use general population mortality to improve extrapolations of the short-term data, assuming a constant additive or multiplicative effect on the hazards for all-cause mortality for study patients relative to the general population. A more plausible assumption may be a constant effect on the hazard for the specific cause of death targeted by the treatments. To address this problem, we use independent parametric survival models for cause-specific mortality among the general population. Since causes of death are unobserved for the patients of interest, a polyhazard model is used to express their all-cause mortality as a sum of latent cause-specific hazards. Assuming proportional cause-specific hazards between the general and study populations then allows us to extrapolate mortality of the patients of interest to the long term. A Bayesian framework is used to jointly model all sources of data. By simulation we show that ignoring cause-specific hazards leads to biased estimates of mean survival when the proportion of deaths due to the cause of interest changes through time. The methods are applied to an evaluation of implantable cardioverter defibrillators (ICD) for the prevention of sudden cardiac death among patients with cardiac arrhythmia. After accounting for cause-specific mortality, substantial differences are seen in estimates of life years gained from ICD. PMID:25413028
ERIC Educational Resources Information Center
Boudreaux, Gregory M.; Wells, M. Scott
2007-01-01
Everyone with a thorough knowledge of single variable calculus knows that integration can be used to find the length of a curve on a given interval, called its arc length. Fortunately, if one endeavors to pose and solve more interesting problems than simply computing lengths of various curves, there are techniques available that do not require an…
Theoretical Basis for Finite Difference Extrapolation of Sonic Boom Signatures
NASA Technical Reports Server (NTRS)
Plotkin, Kenneth J.
1996-01-01
Calculation of sonic boom signatures for aircraft has traditionally followed the methods of Whitham' and Walkden. The wave disturbance generated by the vehicle is obtained by area rule linearized supersonic flow methods, which yield a locally axisymmetric asymptotic solution. This solution is acoustic in nature, i.e., first order in disturbance quantities, and corresponds to ray acoustics. Cumulative nonlinear distortion of the signature is incorporated by using this solution to adjust propagation speed to first order, thus yielding a solution second order in disturbance quantities. The effects of atmospheric gradients are treated by Blokhintzov's method of geometrical acoustics. Both nonlinear signature evolution and ray tracing are applied as if the pressure field very close to the vehicle were actually that given by the source term (the 'F-function') of the asymptotic linearized flow solution. The viewpoint is thus that the flow solution exists at a small radius near the vehicle, and may be treated as an input to an extrapolation procedure consisting of ray tracing and nonlinear aging. The F-function is often regarded as a representation of a near-field pressure signature, and it is common for computational implementations to treat it interchangeably with the pressure signature. There is a 'matching radius' between the source function and the subsequent propagation extrapolation. This viewpoint has been supported by wind tunnel tests of simple models, and very typically yields correct results for actual flight vehicles. The assumption that the F-function and near-field signature are interchangeable is generally not correct. The flowfield of a vehicle which is not axisymmetric contains crossflow components which are very significant at small radii and less so at larger distances. From an acoustical viewpoint, the crossflow is equivalent to source diffraction portions of the wave field. Use of the F-function as a near field signature effectively assumes that the
Measuring Thermodynamic Length
Crooks, Gavin E
2007-09-07
Thermodynamic length is a metric distance between equilibrium thermodynamic states. Among other interesting properties, this metric asymptotically bounds the dissipation induced by a finite time transformation of a thermodynamic system. It is also connected to the Jensen-Shannon divergence, Fisher information, and Rao's entropy differential metric. Therefore, thermodynamic length is of central interestin understanding matter out of equilibrium. In this Letter, we will consider how to denethermodynamic length for a small system described by equilibrium statistical mechanics and how to measure thermodynamic length within a computer simulation. Surprisingly, Bennett's classic acceptance ratio method for measuring free energy differences also measures thermodynamic length.
Biological bases for cancer dose-response extrapolation procedures
Wilson, J.D. )
1991-01-01
The Moolgavkar-Knudson theory of carcinogenesis of 1981 incorporates the viable portions of earlier multistage theories and provides the basis for both the linearized multistage and biologically based dose-response extrapolation methodologies. This theory begins with the premise that cancer occurs because irreversible genetic changes (mutations) are required for transformation of normal cells to cancer cells; incidence data are consistent with only two critical changes begin required, but a small contribution from three or higher mutation pathways cannot be rules out. Events or agents that increase the rate of cell division also increase the probability that one of these critical mutations will occur by reducing the time available for repair of DNA lesions before mitosis. The DNA lesions can occur from background causes or from treatment with mutagenic agents. Thus, the equations describing incidence as a function of exposure to carcinogenic agents include two separate terms, one accounting for mutagenic and one for mitogenic stimuli. At high exposures these interact, producing synergism and high incidence rates, but at low exposures they are effectively independent. The multistage models that are now used include only terms corresponding to the mutagenic stimuli and this fail to adequately describe incidence at high dose rates. Biologically based models attempt to include mitogenic effects, as well; they are usually limited by data availability.
An empirical relationship for extrapolating sparse experimental lap joint data.
Segalman, Daniel Joseph; Starr, Michael James
2010-10-01
Correctly incorporating the influence of mechanical joints in built-up mechanical systems is a critical element for model development for structural dynamics predictions. Quality experimental data are often difficult to obtain and is rarely sufficient to determine fully parameters for relevant mathematical models. On the other hand, fine-mesh finite element (FMFE) modeling facilitates innumerable numerical experiments at modest cost. Detailed FMFE analysis of built-up structures with frictional interfaces reproduces trends among problem parameters found experimentally, but there are qualitative differences. Those differences are currently ascribed to the very approximate nature of the friction model available in most finite element codes. Though numerical simulations are insufficient to produce qualitatively correct behavior of joints, some relations, developed here through observations of a multitude of numerical experiments, suggest interesting relationships among joint properties measured under different loading conditions. These relationships can be generalized into forms consistent with data from physical experiments. One such relationship, developed here, expresses the rate of energy dissipation per cycle within the joint under various combinations of extensional and clamping load in terms of dissipation under other load conditions. The use of this relationship-though not exact-is demonstrated for the purpose of extrapolating a representative set of experimental data to span the range of variability observed from real data.
Ramp response estimation and spectrum extrapolation for ultrasonic scattering
NASA Astrophysics Data System (ADS)
Clark, G. A.
1984-08-01
A combined application of digital signal processing, estimation theory, and scattering theory is used to attack the important problem of target identification. The basic problem is that of examining an object with an ultrasonic pulse and using the reflected signal to determine various properties of the object. Typically, an impulse reponse h(t) can be calculated from knowledge of the (input x(t)) signal and the reflected (output y(t)) signal. If the impulse response can be found, it can sometimes contain important information about the object. We have studied some new algorithms for impulse response estimation. It is also well-known that the ramp response contains information about the cross-sectional area of the scatterer. The ramp response can be calculated directly from the impulse response, but the estimate of cross-sectional area is degraded by the fact that the ultrasonic transducer severely bandlimits the data. Algorithms have been produced for extrapolating the ramp response spectrum to improve the cross-sectional area estimates from the ramp response technique. Experimental results demonstrating the estimation of properties of a known flaw (created by a saw cut) in a block of aluminum are presented.
Detail enhancement of blurred infrared images based on frequency extrapolation
NASA Astrophysics Data System (ADS)
Xu, Fuyuan; Zeng, Deguo; Zhang, Jun; Zheng, Ziyang; Wei, Fei; Wang, Tiedan
2016-05-01
A novel algorithm for enhancing the details of the blurred infrared images based on frequency extrapolation has been raised in this paper. Unlike other researchers' work, this algorithm mainly focuses on how to predict the higher frequency information based on the Laplacian pyramid separation of the blurred image. This algorithm uses the first level of the high frequency component of the pyramid of the blurred image to reverse-generate a higher, non-existing frequency component, and adds back to the histogram equalized input blurred image. A simple nonlinear operator is used to analyze the extracted first level high frequency component of the pyramid. Two critical parameters are participated in the calculation known as the clipping parameter C and the scaling parameter S. The detailed analysis of how these two parameters work during the procedure is figure demonstrated in this paper. The blurred image will become clear, and the detail will be enhanced due to the added higher frequency information. This algorithm has the advantages of computational simplicity and great performance, and it can definitely be deployed in the real-time industrial applications. We have done lots of experiments and gave illustrations of the algorithm's performance in this paper to convince its effectiveness.
Interspecies Gene Name Extrapolation--A New Approach.
Petric, Roxana Cojocneanu; Braicu, Cornelia; Bassi, Cristian; Pop, Laura; Taranu, Ionelia; Dragos, Nicolae; Dumitrascu, Dan; Negrini, Massimo; Berindan-Neagoe, Ioana
2015-01-01
The use of animal models has facilitated numerous scientific developments, especially when employing "omics" technologies to study the effects of various environmental factors on humans. Our study presents a new bioinformatics pipeline suitable when the generated microarray data from animal models does not contain the necessary human gene name annotation. We conducted single color gene expression microarray on duodenum and spleen tissue obtained from pigs which have been exposed to zearalenone and Escherichia coli contamination, either alone or combined. By performing a combination of file format modifications and data alignments using various online tools as well as a command line environment, we performed the pig to human gene name extrapolation with an average yield of 58.34%, compared to 3.64% when applying more simple methods. In conclusion, while online data analysis portals on their own are of great importance in data management and assessment, our new pipeline provided a more effective approach for a situation which can be frequently encountered by researchers in the "omics" era. PMID:26407293
Impact ejecta dynamics in an atmosphere - Experimental results and extrapolations
NASA Technical Reports Server (NTRS)
Schultz, P. H.; Gault, D. E.
1982-01-01
It is noted that the impacts of 0.635-cm aluminum projectiles at 6 km/sec into fine pumice dust, at 1 atm, generate a ball of ionized gas behind an expanding curtain of upward moving ejecta. The gas ball forms a toroid which dissolves as it is driven along the interior of the ejecta curtain, by contrast to near-surface explosions in which a fireball envelops early-time crater growth. High frame rate Schlieren photographs show that the atmosphere at the base of the ejecta curtain is initially turbulent, but later forms a vortex. These experiments suggest that although small size ejecta may be decelerated by air drag, they are not simply lofted and suspended but become incorporated in an ejecta cloud that is controlled by air flow which is produced by the response of the atmosphere to the impact. The extrapolation of these results to large body impacts on the earth suggests such contrasts with laboratory experiments as a large quantity of impact-generated vapor, the supersonic advance of the ejecta curtain, the lessened effect of air drag due to the tenuous upper atmosphere, and the role of secondary cratering.
Time-domain incident-field extrapolation technique based on the singularity-expansion method
Klaasen, J.J.
1991-05-01
In this report, a method presented to extrapolate measurements from Nuclear Electromagnetic Pulse (NEMP) assessments directly in the time domain. This method is based on a time-domain extrapolation function which is obtained from the Singularity Expansion Method representation of the measured incident field of the NEMP simulator. Once the time-domain extrapolation function is determined, the responses recorded during an assessment can be extrapolated simply by convolving them with the time domain extrapolation function. It is found that to obtain useful extrapolated responses, the incident field measurements needs to be made minimum phase; otherwise unbounded results can be obtained. Results obtained with this technique are presented, using data from actual assessments.
Complexities of glucuronidation affecting in vitro in vivo extrapolation.
Lin, Jiunn H; Wong, Bradley K
2002-12-01
Glucuronidation is responsible for the clearance of a diverse range of drug and chemicals whose topology confers properties that complicate in vitro-in vivo clearance correlations as compared to those possible for oxidative metabolism. The active site of the UGTs faces the inside of the luminal space of the endoplasmic reticulum, thus presenting diffusional barriers for substrates, the cosubstrate, UDPGA, and resultant glucuronide products. Transport processes for the cosubstrate UDPGA and glucuronidated products likely contribute to the well-known latency phenomena in which exogenous detergents or alamethicin are required for maximal UGT activity in microsomes. This complicates the extrapolation of results of in vitro clearance studies to the in vivo situation. Even with activation, the microsomal-based clearance values still underestimate the actual in vivo UGT-mediated clearance; therefore latency is not the only explanation for the poor correlation. Recent data indicate that hepatocytes are a promising in vitro system that can be used for the early evaluation of human clearance behavior of drug candidates. Both induction and inhibition of UGT-mediated clearance are a source of clinical drug-drug interactions. Emerging evidence indicates that the same mechanisms identified in the regulation of CYP enzymes also are involved in regulation of the UGTs, i.e., CAR, AH and probably PXR mediate regulation of UGT1A1, 1A6 and UGT2B7, respectively. In contrast to CYP-mediated interactions, with a few exceptions, the magnitude of UGT-mediated interactions are less than 2-fold because of the relatively high UGT Km values and substrate overlap among the multiple isozymes. PMID:12369890
Trinkaus, Erik; Holliday, Trenton W.; Auerbach, Benjamin M.
2014-01-01
The Late Pleistocene archaic humans from western Eurasia (the Neandertals) have been described for a century as exhibiting absolutely and relatively long clavicles. This aspect of their body proportions has been used to distinguish them from modern humans, invoked to account for other aspects of their anatomy and genetics, used in assessments of their phylogenetic polarities, and used as evidence for Late Pleistocene population relationships. However, it has been unclear whether the usual scaling of Neandertal clavicular lengths to their associated humeral lengths reflects long clavicles, short humeri, or both. Neandertal clavicle lengths, along with those of early modern humans and latitudinally diverse recent humans, were compared with both humeral lengths and estimated body masses (based on femoral head diameters). The Neandertal do have long clavicles relative their humeri, even though they fall within the ranges of variation of early and recent humans. However, when scaled to body masses, their humeral lengths are relatively short, and their clavicular lengths are indistinguishable from those of Late Pleistocene and recent modern humans. The few sufficiently complete Early Pleistocene Homo clavicles seem to have relative lengths also well within recent human variation. Therefore, appropriately scaled clavicular length seems to have varied little through the genus Homo, and it should not be used to account for other aspects of Neandertal biology or their phylogenetic status. PMID:24616525
NASA Technical Reports Server (NTRS)
Darden, C. M.
1984-01-01
A method for analyzing shock coalescence which includes three dimensional effects was developed. The method is based on an extension of the axisymmetric solution, with asymmetric effects introduced through an additional set of governing equations, derived by taking the second circumferential derivative of the standard shock equations in the plane of symmetry. The coalescence method is consistent with and has been combined with a nonlinear sonic boom extrapolation program which is based on the method of characteristics. The extrapolation program, is able to extrapolate pressure signatures which include embedded shocks from an initial data line in the plane of symmetry at approximately one body length from the axis of the aircraft to the ground. The axisymmetric shock coalescence solution, the asymmetric shock coalescence solution, the method of incorporating these solutions into the extrapolation program, and the methods used to determine spatial derivatives needed in the coalescence solution are described. Results of the method are shown for a body of revolution at a small, positive angle of attack.
The evolution of σ γP with coherence length
NASA Astrophysics Data System (ADS)
Caldwell, Allen
2016-07-01
Assuming the form {σ }γ {{P}}\\propto {l}{λ {{eff}}} at fixed Q 2 for the behavior of the virtual-photon proton scattering cross section, where l is the coherence length of the photon fluctuations, it is seen that the extrapolated values of {σ }γ {{P}} for different Q 2 cross for l≈ {10}8 fm. It is argued that this behavior is not physical, and that the behavior of the cross sections must change before this coherence length l is reached. This could set the scale for the onset of saturation of the parton densities in the photon, and thereby saturation of parton densities in general.
NASA Astrophysics Data System (ADS)
Fang, Jun; Gao, Xingyu; Song, Haifeng; Wang, Han
2016-06-01
Wavefunction extrapolation greatly reduces the number of self-consistent field (SCF) iterations and thus the overall computational cost of Born-Oppenheimer molecular dynamics (BOMD) that is based on the Kohn-Sham density functional theory. Going against the intuition that the higher order of extrapolation possesses a better accuracy, we demonstrate, from both theoretical and numerical perspectives, that the extrapolation accuracy firstly increases and then decreases with respect to the order, and an optimal extrapolation order in terms of minimal number of SCF iterations always exists. We also prove that the optimal order tends to be larger when using larger MD time steps or more strict SCF convergence criteria. By example BOMD simulations of a solid copper system, we show that the optimal extrapolation order covers a broad range when varying the MD time step or the SCF convergence criterion. Therefore, we suggest the necessity for BOMD simulation packages to open the user interface and to provide more choices on the extrapolation order. Another factor that may influence the extrapolation accuracy is the alignment scheme that eliminates the discontinuity in the wavefunctions with respect to the atomic or cell variables. We prove the equivalence between the two existing schemes, thus the implementation of either of them does not lead to essential difference in the extrapolation accuracy.
Zhong, Sheng-hua; Ma, Zheng; Wilson, Colin; Liu, Yan; Flombaum, Jonathan I
2014-01-01
Intuitively, extrapolating object trajectories should make visual tracking more accurate. This has proven to be true in many contexts that involve tracking a single item. But surprisingly, when tracking multiple identical items in what is known as “multiple object tracking,” observers often appear to ignore direction of motion, relying instead on basic spatial memory. We investigated potential reasons for this behavior through probabilistic models that were endowed with perceptual limitations in the range of typical human observers, including noisy spatial perception. When we compared a model that weights its extrapolations relative to other sources of information about object position, and one that does not extrapolate at all, we found no reliable difference in performance, belying the intuition that extrapolation always benefits tracking. In follow-up experiments we found this to be true for a variety of models that weight observations and predictions in different ways; in some cases we even observed worse performance for models that use extrapolations compared to a model that does not at all. Ultimately, the best performing models either did not extrapolate, or extrapolated very conservatively, relying heavily on observations. These results illustrate the difficulty and attendant hazards of using noisy inputs to extrapolate the trajectories of multiple objects simultaneously in situations with targets and featurally confusable nontargets. PMID:25311300
Cross-species extrapolation of chemical effects: Challenges and new insights
One of the greatest uncertainties in chemical risk assessment is extrapolation of effects from tested to untested species. While this undoubtedly is a challenge in the human health arena, species extrapolation is a particularly daunting task in ecological assessments, where it is...
Coefficients of Effective Length.
ERIC Educational Resources Information Center
Edwards, Roger H.
1981-01-01
Under certain conditions, a validity Coefficient of Effective Length (CEL) can produce highly misleading results. A modified coefficent is suggested for use when empirical studies indicate that underlying assumptions have been violated. (Author/BW)
ERIC Educational Resources Information Center
Martins, Roberto de A.
1978-01-01
Describes a thought experiment using a general analysis approach with Lorentz transformations to show that the apparent self-contradictions of special relativity concerning the length-paradox are really non-existant. (GA)
[Sonographic leg length measurement].
Holst, A; Thomas, W
1989-03-01
After brief presentation of the clinical and radiological methods to measure the leg length and the leg length difference the authors outline the new diagnostic method for measuring the leg length and the leg length difference by means of real time sonography. Postmortem tests and clinical examples show that ultrasound is ideal to determine exactly the length of femur and tibia. The joint gaps on the hip, knee and upper ankle joint can be demonstrated by means of a 5 MHz linear scanner. A 1 mm strong metal bar on the skin and under the scanner is placed at right angles to the longitudinal axis of the body so that the bar can be seen in the centre. A measuring device gives the distances of the joint gaps in cm so that the difference correspond to the real length of femur and tibia. This standardised measuring is done by a particularly developed bearing and measuring device. The results of the sonographical measurements on 20 corpses and checking after consecutive dissections showed in 75% of cases a 100% sonographic measuring accuracy of the total leg length. The separately considered results for femur (85%) and tibia (90) were even better. The maximum sonographic measuring fault was 1.0 cm for the femur (in one case) and 0.5 cm for the tibia, respectively. All sonographic measurements were performed with the Sonoline SL-1 of the Siemens Company (Erlangen, W-Germany). Thus, sonographical measuring of the leg length offers a reliable, non-invasive method that can be repeated as often as necessary and is simply executed.(ABSTRACT TRUNCATED AT 250 WORDS) PMID:2652268
Sprouse, Gene D.
2011-07-15
Technological changes have moved publishing to electronic-first publication where the print version has been relegated to simply another display mode. Distribution in HTML and EPUB formats, for example, changes the reading environment and reduces the need for strict pagination. Therefore, in an effort to streamline the calculation of length, the APS journals will no longer use the printed page as the determining factor for length. Instead the journals will now use word counts (or word equivalents for tables, figures, and equations) to establish length; for details please see http://publish.aps.org/authors/length-guide. The title, byline, abstract, acknowledgment, and references will not be included in these counts allowing authors the freedom to appropriately credit coworkers, funding sources, and the previous literature, bringing all relevant references to the attention of readers. This new method for determining length will be easier for authors to calculate in advance, and lead to fewer length-associated revisions in proof, yet still retain the quality of concise communication that is a virtue of short papers.
Area, length and thickness conservation: Dogma or reality?
NASA Astrophysics Data System (ADS)
Moretti, Isabelle; Callot, Jean Paul
2012-08-01
The basic assumption of quantitative structural geology is the preservation of material during deformation. However the hypothesis of volume conservation alone does not help to predict past or future geometries and so this assumption is usually translated into bed length in 2D (or area in 3D) and thickness conservation. When subsurface data are missing, geologists may extrapolate surface data to depth using the kink-band approach. These extrapolations, preserving both thicknesses and dips, lead to geometries which are restorable but often erroneous, due to both disharmonic deformation and internal deformation of layers. First, the Bolivian Sub-Andean Zone case is presented to highlight the evolution of the concepts on which balancing is based, and the important role played by a decoupling level in enhancing disharmony. Second, analogue models are analyzed to test the validity of the balancing techniques. Chamberlin's excess area approach is shown to be on average valid. However, neither the length nor the thicknesses are preserved. We propose that in real cases, the length preservation hypothesis during shortening could also be a wrong assumption. If the data are good enough to image the decollement level, the Chamberlin excess area method could be used to compute the bed length changes.
CONSTRAINING THREE-DIMENSIONAL MAGNETIC FIELD EXTRAPOLATIONS USING THE TWIN PERSPECTIVES OF STEREO
Conlon, Paul A.; Gallagher, Peter T.
2010-05-20
The three-dimensional magnetic topology of a solar active region (NOAA 10956) was reconstructed using a linear force-free field extrapolation constrained using the twin perspectives of STEREO. A set of coronal field configurations was initially generated from extrapolations of the photospheric magnetic field observed by the Michelson Doppler Imager on SOHO. Using an EUV intensity-based cost function, the extrapolated field lines that were most consistent with 171 A passband images from the Extreme UltraViolet Imager on STEREO were identified. This facilitated quantitative constraints to be placed on the twist ({alpha}) of the extrapolated field lines, where {nabla} x B = {alpha}B. Using the constrained values of {alpha}, the evolution in time of twist, connectivity, and magnetic energy were then studied. A flux emergence event was found to result in significant changes in the magnetic topology and total magnetic energy of the region.
Melting of “non-magic” argon clusters and extrapolation to the bulk limit
Senn, Florian Wiebke, Jonas; Schumann, Ole; Gohr, Sebastian; Schwerdtfeger, Peter; Pahl, Elke
2014-01-28
The melting of argon clusters Ar{sub N} is investigated by applying a parallel-tempering Monte Carlo algorithm for all cluster sizes in the range from 55 to 309 atoms. Extrapolation to the bulk gives a melting temperature of 85.9 K in good agreement with the previous value of 88.9 K using only Mackay icosahedral clusters for the extrapolation [E. Pahl, F. Calvo, L. Koči, and P. Schwerdtfeger, “Accurate melting temperatures for neon and argon from ab initio Monte Carlo simulations,” Angew. Chem., Int. Ed. 47, 8207 (2008)]. Our results for argon demonstrate that for the extrapolation to the bulk one does not have to restrict to magic number cluster sizes in order to obtain good estimates for the bulk melting temperature. However, the extrapolation to the bulk remains a problem, especially for the systematic selection of suitable cluster sizes.
Can Tauc plot extrapolation be used for direct-band-gap semiconductor nanocrystals?
Feng, Y. Lin, S.; Huang, S.; Shrestha, S.; Conibeer, G.
2015-03-28
Despite that Tauc plot extrapolation has been widely adopted for extracting bandgap energies of semiconductors, there is a lack of theoretical support for applying it to nanocrystals. In this paper, direct-allowed optical transitions in semiconductor nanocrystals have been formulated based on a purely theoretical approach. This result reveals a size-dependant transition of the power factor used in Tauc plot, increasing from one half used in the 3D bulk case to one in the 0D case. This size-dependant intermediate value of power factor allows a better extrapolation of measured absorption data. Being a material characterization technique, the generalized Tauc extrapolation gives a more reasonable and accurate acquisition of the intrinsic bandgap, while the unjustified purpose of extrapolating any elevated bandgap caused by quantum confinement is shown to be incorrect.
Controversy on toxicological dose-response relationships and low-dose extrapolation of respective risks is often the consequence of misleading data presentation, lack of differentiation between types of response variables, and diverging mechanistic interpretation. In this chapter...
Cross-species extrapolation of toxicity data from limited surrogate test organisms to all wildlife with potential of chemical exposure remains a key challenge in ecological risk assessment. A number of factors affect extrapolation, including the chemical exposure, pharmacokinetic...
Document Length Normalization.
ERIC Educational Resources Information Center
Singhal, Amit; And Others
1996-01-01
Describes a study that investigated document retrieval relevance based on document length in an experimental text collection. Topics include term weighting and document ranking, retrieval strategies such as the vector-space cosine match, and a modified technique called the pivoted cosine normalization. (LRW)
NASA Astrophysics Data System (ADS)
Tay, Kim Gaik; Kek, Sie Long; Abdul-Kahar, Rosmila
2015-05-01
In this paper, we have further improved the limitations of our previous two Richardson's extrapolation spreadsheet calculators for computing differentiations numerically. The new feature in this new Richardson's extrapolation spreadsheet calculator is fully automated up to any level based on the stopping criteria using VBA programming. The new version is more flexible because it is controlled by programming. Furthermore, it reduces computational time and CPU memory.
Video Extrapolation Method Based on Time-Varying Energy Optimization and CIP.
Sakaino, Hidetomo
2016-09-01
Video extrapolation/prediction methods are often used to synthesize new videos from images. For fluid-like images and dynamic textures as well as moving rigid objects, most state-of-the-art video extrapolation methods use non-physics-based models that learn orthogonal bases from a number of images but at high computation cost. Unfortunately, data truncation can cause image degradation, i.e., blur, artifact, and insufficient motion changes. To extrapolate videos that more strictly follow physical rules, this paper proposes a physics-based method that needs only a few images and is truncation-free. We utilize physics-based equations with image intensity and velocity: optical flow, Navier-Stokes, continuity, and advection equations. These allow us to use partial difference equations to deal with the local image feature changes. Image degradation during extrapolation is minimized by updating model parameters, where a novel time-varying energy balancer model that uses energy based image features, i.e., texture, velocity, and edge. Moreover, the advection equation is discretized by high-order constrained interpolation profile for lower quantization error than can be achieved by the previous finite difference method in long-term videos. Experiments show that the proposed energy based video extrapolation method outperforms the state-of-the-art video extrapolation methods in terms of image quality and computation cost. PMID:27305677
How useful are corpus-based methods for extrapolating psycholinguistic variables?
Mandera, Paweł; Keuleers, Emmanuel; Brysbaert, Marc
2015-01-01
Subjective ratings for age of acquisition, concreteness, affective valence, and many other variables are an important element of psycholinguistic research. However, even for well-studied languages, ratings usually cover just a small part of the vocabulary. A possible solution involves using corpora to build a semantic similarity space and to apply machine learning techniques to extrapolate existing ratings to previously unrated words. We conduct a systematic comparison of two extrapolation techniques: k-nearest neighbours, and random forest, in combination with semantic spaces built using latent semantic analysis, topic model, a hyperspace analogue to language (HAL)-like model, and a skip-gram model. A variant of the k-nearest neighbours method used with skip-gram word vectors gives the most accurate predictions but the random forest method has an advantage of being able to easily incorporate additional predictors. We evaluate the usefulness of the methods by exploring how much of the human performance in a lexical decision task can be explained by extrapolated ratings for age of acquisition and how precisely we can assign words to discrete categories based on extrapolated ratings. We find that at least some of the extrapolation methods may introduce artefacts to the data and produce results that could lead to different conclusions that would be reached based on the human ratings. From a practical point of view, the usefulness of ratings extrapolated with the described methods may be limited. PMID:25695623
The accuracy of far-field noise obtained by the mathematical extrapolation of near-field noise data
NASA Technical Reports Server (NTRS)
Ahtye, W. F.; Karel, S.
1975-01-01
Results are presented for an analytical study of the accuracy and limitations of a technique that permits the mathematical extrapolation of near-field noise data to far-field conditions. The effects of the following variables on predictive accuracy of the far-field pressure were examined: (1) number of near-field microphones; (2) length of source distribution; (3) complexity of near-field and far-field distributions; (4) source-to-microphone distance; and (5) uncertainties in microphone data and imprecision in the location of the near-field microphones. It is shown that the most important parameters describing predictive accuracy are the number of microphones, the ratio of source length to acoustic wavelength (L/lambda), and the error in location of near-field microphones. For maximum microphone location errors of plus or minus 1 cm, only an accuracy of plus or minus 2.5 dB can be attained with approximately 40 microphones for the highest L/lambda of 10.
The accuracy of far-field noise obtained by the mathematical extrapolation of near-field noise data
NASA Technical Reports Server (NTRS)
Ahtye, W. F.; Karel, S.
1975-01-01
Results are described of an analytical study of the accuracy and limitations of a technique that permits the mathematical extrapolation of near-field noise data to far-field conditions. The effects of the following variables on predictive accuracy of the far-field pressure were examined: (1) number of near-field microphones; (2) length of source distribution; (3) complexity of near-field and far-field distributions; (4) source-to-microphone distance; and (5) uncertainties in microphone data and imprecision in the location of the near-field microphones. It is shown that the most important parameters describing predictive accuracy are the number of microphones, the ratio of source length to acoustic wavelength, (L/wavelength), and the error in location of near-field microphones. If microphone measurement and location errors are not included, then far-field pressures can be accurately predicted up to L/wavelength values of 15 using approximately 50 microphones. For maximum microphone location errors of + or - 1 cm, only an accuracy of + or - 2-1/2 db can be attained with approximately 40 microphones for the highest L/wavelength of 10.
The K+ K+ scattering length from Lattice QCD
Silas Beane; Thomas Luu; Konstantinos Orginos; Assumpta Parreno; Martin Savage; Aaron Torok; Andre Walker-Loud
2007-09-11
The K+K+ scattering length is calculated in fully-dynamical lattice QCD with domain-wall valence quarks on the MILC asqtad-improved gauge configurations with fourth-rooted staggered sea quarks. Three-flavor mixed-action chiral perturbation theory at next-to-leading order, which includes the leading effects of the finite lattice spacing, is used to extrapolate the results of the lattice calculation to the physical value of mK + /fK + . We find mK^+ aK^+ K^+ = â~0.352 Â± 0.016, where the statistical and systematic errors have been combined in quadrature.
Inoue, S.; Magara, T.; Choe, G. S.; Kim, K. S.; Pandey, V. S.; Shiota, D.; Kusano, K.
2014-01-01
We develop a nonlinear force-free field (NLFFF) extrapolation code based on the magnetohydrodynamic (MHD) relaxation method. We extend the classical MHD relaxation method in two important ways. First, we introduce an algorithm initially proposed by Dedner et al. to effectively clean the numerical errors associated with ∇ · B . Second, the multigrid type method is implemented in our NLFFF to perform direct analysis of the high-resolution magnetogram data. As a result of these two implementations, we successfully extrapolated the high resolution force-free field introduced by Low and Lou with better accuracy in a drastically shorter time. We also applied our extrapolation method to the MHD solution obtained from the flux-emergence simulation by Magara. We found that NLFFF extrapolation may be less effective for reproducing areas higher than a half-domain, where some magnetic loops are found in a state of continuous upward expansion. However, an inverse S-shaped structure consisting of the sheared and twisted loops formed in the lower region can be captured well through our NLFFF extrapolation method. We further discuss how well these sheared and twisted fields are reconstructed by estimating the magnetic topology and twist quantitatively.
The Extrapolation of High Altitude Solar Cell I(V) Characteristics to AM0
NASA Technical Reports Server (NTRS)
Snyder, David B.; Scheiman, David A.; Jenkins, Phillip P.; Reinke, William; Blankenship, Kurt; Demers, James
2007-01-01
The high altitude aircraft method has been used at NASA GRC since the early 1960's to calibrate solar cell short circuit current, ISC, to Air Mass Zero (AMO). This method extrapolates ISC to AM0 via the Langley plot method, a logarithmic extrapolation to 0 air mass, and includes corrections for the varying Earth-Sun distance to 1.0 AU and compensating for the non-uniform ozone distribution in the atmosphere. However, other characteristics of the solar cell I(V) curve do not extrapolate in the same way. Another approach is needed to extrapolate VOC and the maximum power point (PMAX) to AM0 illumination. As part of the high altitude aircraft method, VOC and PMAX can be obtained as ISC changes during the flight. These values can then the extrapolated, sometimes interpolated, to the ISC(AM0) value. This approach should be valid as long as the shape of the solar spectra in the stratosphere does not change too much from AMO. As a feasibility check, the results are compared to AMO I(V) curves obtained using the NASA GRC X25 based multi-source simulator. This paper investigates the approach on both multi-junction solar cells and sub-cells.
NASA Technical Reports Server (NTRS)
Lueck, Dale E. (Inventor)
1994-01-01
Payload customers for the Space Shuttle have recently expressed concerns about the possibility of their payloads at an adjacent pad being contaminated by plume effluents from a shuttle at an active pad as they await launch on an inactive pad. As part of a study to satisfy such concerns a ring of inexpensive dosimeters was deployed around the active pad at the inter-pad distance. However, following a launch, dosimeters cannot be read for several hours after the exposure. As a consequence factors such as different substrates, solvent systems, and possible volatilization of HCl from the badges were studied. This observation led to the length of stain (LOS) dosimeters of this invention. Commercial passive LOS dosimeters are sensitive only to the extent of being capable of sensing 2 ppm to 20 ppm if the exposure is 8 hours. To map and quantitate the HCl generated by Shuttle launches, and in the atmosphere within a radius of 1.5 miles from the active pad, a sensitivity of 2 ppm HCl in the atmospheric gases on an exposure of 5 minutes is required. A passive length of stain dosimeter has been developed having a sensitivity rendering it capable of detecting a gas in a concentration as low as 2 ppm on an exposure of five minutes.
Forced Field Extrapolation of the Magnetic Structure of the Hα fibrils in the Solar Chromosphere
NASA Astrophysics Data System (ADS)
Xiaoshuai, Zhu; Huaning, Wang; Zhanle, Du; Han, He
2016-07-01
We present a careful assessment of forced field extrapolation using the Solar Dynamics Observatory/Helioseismic and Magnetic Imager magnetogram. We use several metrics to check the convergence property. The extrapolated field lines below 3600 km appear to be aligned with most of the Hα fibrils observed by the New Vacuum Solar Telescope. In the region where magnetic energy is far larger than potential energy, the field lines computed by forced field extrapolation are still consistent with the patterns of Hα fibrils while the nonlinear force-free field results show a large misalignment. The horizontal average of the lorentz force ratio shows that the forced region where the force-free assumption fails can reach heights of 1400–1800 km. The non-force-free state of the chromosphere is also confirmed based on recent radiation magnetohydrodynamics simulations.
Larsen, Ross E.
2016-04-12
In this study, we introduce two simple tight-binding models, which we call fragment frontier orbital extrapolations (FFOE), to extrapolate important electronic properties to the polymer limit using electronic structure calculations on only a few small oligomers. In particular, we demonstrate by comparison to explicit density functional theory calculations that for long oligomers the energies of the highest occupied molecular orbital (HOMO), the lowest unoccupied molecular orbital (LUMO), and of the first electronic excited state are accurately described as a function of number of repeat units by a simple effective Hamiltonian parameterized from electronic structure calculations on monomers, dimers and, optionally,more » tetramers. For the alternating copolymer materials that currently comprise some of the most efficient polymer organic photovoltaic devices one can use these simple but rigorous models to extrapolate computed properties to the polymer limit based on calculations on a small number of low-molecular-weight oligomers.« less
NASA Astrophysics Data System (ADS)
Sandu, Adrian; Schlick, Tamar
1999-05-01
Numerical resonance artifacts have become recognized recently as a limiting factor to increasing the timestep in multiple-timestep (MTS) biomolecular dynamics simulations. At certain timesteps correlated to internal motions (e.g., 5 fs, around half the period of the fastest bond stretch, Tmin), visible inaccuracies or instabilities can occur. Impulse-MTS schemes are vulnerable to these resonance errors since large energy pulses are introduced to the governing dynamics equations when the slow forces are evaluated. We recently showed that such resonance artifacts can be masked significantly by applying extrapolative splitting to stochastic dynamics. Theoretical and numerical analyses of force-splitting integrators based on the Verlet discretization are reported here for linear models to explain these observations and to suggest how to construct effective integrators for biomolecular dynamics that balance stability with accuracy. Analyses for Newtonian dynamics demonstrate the severe resonance patterns of the Impulse splitting, with this severity worsening with the outer timestep, Δ t; Constant Extrapolation is generally unstable, but the disturbances do not grow with Δ t. Thus, the stochastic extrapolative combination can counteract generic instabilities and largely alleviate resonances with a sufficiently strong Langevin heat-bath coupling (γ), estimates for which are derived here based on the fastest and slowest motion periods. These resonance results generally hold for nonlinear test systems: a water tetramer and solvated protein. Proposed related approaches such as Extrapolation/Correction and Midpoint Extrapolation work better than Constant Extrapolation only for timesteps less than Tmin/2. An effective extrapolative stochastic approach for biomolecules that balances long-timestep stability with good accuracy for the fast subsystem is then applied to a biomolecule using a three-class partitioning: the medium forces are treated by Midpoint Extrapolationvia
Small angle x-ray scattering of chromatin. Radius and mass per unit length depend on linker length.
Williams, S P; Langmore, J P
1991-01-01
Analyses of low angle x-ray scattering from chromatin, isolated by identical procedures but from different species, indicate that fiber diameter and number of nucleosomes per unit length increase with the amount of nucleosome linker DNA. Experiments were conducted at physiological ionic strength to obtain parameters reflecting the structure most likely present in living cells. Guinier analyses were performed on scattering from solutions of soluble chromatin from Necturus maculosus erythrocytes (linker length 48 bp), chicken erythrocytes (linker length 64 bp), and Thyone briareus sperm (linker length 87 bp). The results were extrapolated to infinite dilution to eliminate interparticle contributions to the scattering. Cross-sectional radii of gyration were found to be 10.9 +/- 0.5, 12.1 +/- 0.4, and 15.9 +/- 0.5 nm for Necturus, chicken, and Thyone chromatin, respectively, which are consistent with fiber diameters of 30.8, 34.2, and 45.0 nm. Mass per unit lengths were found to be 6.9 +/- 0.5, 8.3 +/- 0.6, and 11.8 +/- 1.4 nucleosomes per 10 nm for Necturus, chicken, and Thyone chromatin, respectively. The geometrical consequences of the experimental mass per unit lengths and radii of gyration are consistent with a conserved interaction among nucleosomes. Cross-linking agents were found to have little effect on fiber external geometry, but significant effect on internal structure. The absolute values of fiber diameter and mass per unit length, and their dependencies upon linker length agree with the predictions of the double-helical crossed-linker model. A compilation of all published x-ray scattering data from the last decade indicates that the relationship between chromatin structure and linker length is consistent with data obtained by other investigators. Images FIGURE 1 PMID:2049522
Improvement of the Quality of Reconstructed Holographic Images by Extrapolation of Digital Holograms
NASA Astrophysics Data System (ADS)
Dyomin, V. V.; Olshukov, A. S.
2016-02-01
The work is devoted to investigation of noise in reconstructed holographic images in the form of a system of fringes parallel to the hologram frame boundaries. Mathematical and physical interpretation is proposed together with an algorithm for reduction of this effect by extrapolation of digital holograms using bicubic splines. The efficiency of the algorithm is estimated and examples of its application are presented.
Route-to-route extrapolation of the toxic potency of MTBE.
Dourson, M L; Felter, S P
1997-12-01
MTBE is a volatile organic compound used as an oxygenating agent in gasoline. Inhalation from fumes while refueling automobiles is the principle route of exposure for humans, and toxicity by this route has been well studied. Oral exposures to MTBE exist as well, primarily due to groundwater contamination from leaking stationary sources, such as underground storage tanks. Assessing the potential public health impacts of oral exposures to MTBE is problematic because drinking water studies do not exist for MTBE, and the few oil-gavage studies from which a risk assessment could be derived are limited. This paper evaluates the suitability of the MTBE database for conducting an inhalation route-to-oral route extrapolation of toxicity. This includes evaluating the similarity of critical effect between these two routes, quantifiable differences in absorption, distribution, metabolism, and excretion, and sufficiency of toxicity data by the inhalation route. We conclude that such an extrapolation is appropriate and have validated the extrapolation by finding comparable toxicity between a subchronic gavage oral bioassay and oral doses we extrapolate from a subchronic inhalation bioassay. Our results are extended to the 2-year inhalation toxicity study by Chun et al. (1992) in which rats were exposed to 0, 400, 3000, or 8000 ppm MTBE for 6 hr/d, 5 d/wk. We have estimated the equivalent oral doses to be 0, 130, 940, or 2700 mg/kg/d. These equivalent doses may be useful in conducting noncancer and cancer risk assessments. PMID:9463928
THE MISUSE OF HYDROLOGIC UNIT MAPS FOR EXTRAPOLATION, REPORTING, AND ECOSYSTEM MANAGEMENT
The use of watersheds to conduct research on land-water relationships has expanded recently to include both extrapolation and reporting of water resource information and ecosystem management. More often than not, hydrologic units, and hydrologic unit codes (HUCs) in particular, a...
Extrapolation of multiplicity distribution in p+p(\\bar{p}) collisions to LHC energies
NASA Astrophysics Data System (ADS)
Dash, Ajay Kumar; Mohanty, Bedangadas
2010-02-01
The multiplicity (Nch) and pseudorapidity distribution (dNch/dη) of primary charged particles in p + p collisions at Large Hadron Collider (LHC) energies of \\sqrt{s} = 10 and 14 TeV are obtained from extrapolation of existing measurements at lower \\sqrt{s}. These distributions are then compared to calculations from PYTHIA and PHOJET models. The existing \\sqrt{s} measurements are unable to distinguish between a logarithmic and power law dependence of the average charged particle multiplicity (langNchrang) on \\sqrt{s}, and their extrapolation to energies accessible at LHC give very different values. Assuming a reasonably good description of inclusive charged particle multiplicity distributions by negative binomial distribution (NBD) at lower \\sqrt{s} to hold for LHC energies, we observe that the logarithmic \\sqrt{s} dependences of langNchrang are favored by the models at midrapidity. The dNch/dη versus η distributions for the existing measurements are found to be reasonably well described by a function with three parameters which accounts for the basic features of the distribution, height at midrapidity, central rapidity plateau and the higher rapidity fall-off. Extrapolation of these parameters as a function of \\sqrt{s} is used to predict the pseudorapidity distributions of charged particles at LHC energies. dNch/dη calculations from PYTHIA and PHOJET models are found to be lower compared to those obtained from the extrapolated dNch/dη versus η distributions for a broad η range.
Technology Transfer Automated Retrieval System (TEKTRAN)
In this study, six extrapolation methods have been compared for their ability to estimate daily crop evapotranspiration (ETd) from instantaneous latent heat flux estimates derived from digital airborne multispectral remote sensing imagery. Data used in this study were collected during an experiment...
NASA Technical Reports Server (NTRS)
Mendelson, A.; Manson, S. S.
1960-01-01
A method using finite-difference recurrence relations is presented for direct extrapolation of families of curves. The method is illustrated by applications to creep-rupture data for several materials and it is shown that good results can be obtained without the necessity for any of the usual parameter concepts.
Extrapolation of inhaled particulate toxicity data from experimental animals to humans. Final report
Morgan, D.L.; Steele, V.E.; Hatch, G.E.
1988-02-01
Significant progress has been made over the past three years to develop methodology and assess various tissues and tissue-sensitivity endpoints with the ultimate goal of validating a proposed extrapolation model. This model allows the quantitative extrapolation of inhaled particulate toxicology data from experimental animals to man. Methodology was developed to accurately measure nucleotide levels in small tissue samples, to determine cellular toxicant levels, to isolate tissue or cells at various levels of the respiratory tract, and to culture animal and human cells for identical treatment conditions. The tissues assessed were nasal turbinate epithelial, olfactory epithelial, and alveolar macrophages. Nasal and pulmonary lavage fluids were also studied. These comprehensive, comparative animal-to-human extrapolation studies were unique in that, for the first time, tissue response was measured as a function of actual cellular dose, and common endpoints were used for the same target-cell types in different species. The described cell-sensitivity model could then be used to provide quantitative animal-to-human extrapolation values for inhaled particulates, which will be extremely valuable for human risk assessment.
Jager, Tjalling; Klok, Chris
2010-11-12
The interest of environmental management is in the long-term health of populations and ecosystems. However, toxicity is usually assessed in short-term experiments with individuals. Modelling based on dynamic energy budget (DEB) theory aids the extraction of mechanistic information from the data, which in turn supports educated extrapolation to the population level. To illustrate the use of DEB models in this extrapolation, we analyse a dataset for life cycle toxicity of copper in the earthworm Dendrobaena octaedra. We compare four approaches for the analysis of the toxicity data: no model, a simple DEB model without reserves and maturation (the Kooijman-Metz formulation), a more complex one with static reserves and simplified maturation (as used in the DEBtox software) and a full-scale DEB model (DEB3) with explicit calculation of reserves and maturation. For the population prediction, we compare two simple demographic approaches (discrete time matrix model and continuous time Euler-Lotka equation). In our case, the difference between DEB approaches and population models turned out to be small. However, differences between DEB models increased when extrapolating to more field-relevant conditions. The DEB3 model allows for a completely consistent assessment of toxic effects and therefore greater confidence in extrapolating, but poses greater demands on the available data. PMID:20921051
A NEW CODE FOR NONLINEAR FORCE-FREE FIELD EXTRAPOLATION OF THE GLOBAL CORONA
Jiang Chaowei; Feng Xueshang; Xiang Changqing
2012-08-10
Reliable measurements of the solar magnetic field are still restricted to the photosphere, and our present knowledge of the three-dimensional coronal magnetic field is largely based on extrapolations from photospheric magnetograms using physical models, e.g., the nonlinear force-free field (NLFFF) model that is usually adopted. Most of the currently available NLFFF codes have been developed with computational volume such as a Cartesian box or a spherical wedge, while a global full-sphere extrapolation is still under development. A high-performance global extrapolation code is in particular urgently needed considering that the Solar Dynamics Observatory can provide a full-disk magnetogram with resolution up to 4096 Multiplication-Sign 4096. In this work, we present a new parallelized code for global NLFFF extrapolation with the photosphere magnetogram as input. The method is based on the magnetohydrodynamics relaxation approach, the CESE-MHD numerical scheme, and a Yin-Yang spherical grid that is used to overcome the polar problems of the standard spherical grid. The code is validated by two full-sphere force-free solutions from Low and Lou's semi-analytic force-free field model. The code shows high accuracy and fast convergence, and can be ready for future practical application if combined with an adaptive mesh refinement technique.
NASA Technical Reports Server (NTRS)
Kahn, M. M. S.; Cahill, J. F.
1983-01-01
Use of this analytical parameter, it is shown, highlights the distinction between cases which are dominated by trailing-edge separation, and those for which separation at the shock foot is dominant. Use of the analytical parameter and the distinction noted above greatly improves the correlation of separation data and the extrapolation of wind tunnel data to flight conditions.
Atomization Energies of SO and SO2; Basis Set Extrapolation Revisted
NASA Technical Reports Server (NTRS)
Bauschlicher, Charles W., Jr.; Ricca, Alessandra; Arnold, James (Technical Monitor)
1998-01-01
The addition of tight functions to sulphur and extrapolation to the complete basis set limit are required to obtain accurate atomization energies. Six different extrapolation procedures are tried. The best atomization energies come from the series of basis sets that yield the most consistent results for all extrapolation techniques. In the variable alpha approach, alpha values larger than 4.5 or smaller than 3, appear to suggest that the extrapolation may not be reliable. It does not appear possible to determine a reliable basis set series using only the triple and quadruple zeta based sets. The scalar relativistic effects reduce the atomization of SO and SO2 by 0.34 and 0.81 kcal/mol, respectively, and clearly must be accounted for if a highly accurate atomization energy is to be computed. The magnitude of the core-valence (CV) contribution to the atomization is affected by missing diffuse valence functions. The CV contribution is much more stable if basis set superposition errors are accounted for. A similar study of SF, SF(+), and SF6 shows that the best family of basis sets varies with the nature of the S bonding.
Buckler, Denny R., Foster L. Mayer, Mark R. Ellersieck and Amha Asfaw. 2003. Evaluation of Minimum Data Requirements for Acute Toxicity Value Extrapolation with Aquatic Organisms. EPA/600/R-03/104. U.S. Environmental Protection Agency, National Health and Environmental Effects Re...
Simulation technique for extrapolation curves in 4πβ-γ coincidence counting method using EGS5 code.
Unno, Y; Sanami, T; Sasaki, S; Hagiwara, M; Yunoki, A
2016-03-01
A simulation technique was developed for the extrapolation technique in 4πβ-γ coincidence counting method. Simultaneous emissions of β and γ rays were calculated using EGS5 code to obtain coincidence counting between both β and γ channels. The simulated extrapolation curves were compared with experimental data obtained with (134)Cs measurements using a plastic scintillator in the β channel. The variation of the extrapolation curves with γ-gate configuration was investigated by the simulation technique. PMID:26688354
Kissling, Wilm Daniel; Dalby, Lars; Fløjgaard, Camilla; Lenoir, Jonathan; Sandel, Brody; Sandom, Christopher; Trøjelsgaard, Kristian; Svenning, Jens-Christian
2014-01-01
Ecological trait data are essential for understanding the broad-scale distribution of biodiversity and its response to global change. For animals, diet represents a fundamental aspect of species’ evolutionary adaptations, ecological and functional roles, and trophic interactions. However, the importance of diet for macroevolutionary and macroecological dynamics remains little explored, partly because of the lack of comprehensive trait datasets. We compiled and evaluated a comprehensive global dataset of diet preferences of mammals (“MammalDIET”). Diet information was digitized from two global and cladewide data sources and errors of data entry by multiple data recorders were assessed. We then developed a hierarchical extrapolation procedure to fill-in diet information for species with missing information. Missing data were extrapolated with information from other taxonomic levels (genus, other species within the same genus, or family) and this extrapolation was subsequently validated both internally (with a jack-knife approach applied to the compiled species-level diet data) and externally (using independent species-level diet information from a comprehensive continentwide data source). Finally, we grouped mammal species into trophic levels and dietary guilds, and their species richness as well as their proportion of total richness were mapped at a global scale for those diet categories with good validation results. The success rate of correctly digitizing data was 94%, indicating that the consistency in data entry among multiple recorders was high. Data sources provided species-level diet information for a total of 2033 species (38% of all 5364 terrestrial mammal species, based on the IUCN taxonomy). For the remaining 3331 species, diet information was mostly extrapolated from genus-level diet information (48% of all terrestrial mammal species), and only rarely from other species within the same genus (6%) or from family level (8%). Internal and external
Kissling, Wilm Daniel; Dalby, Lars; Fløjgaard, Camilla; Lenoir, Jonathan; Sandel, Brody; Sandom, Christopher; Trøjelsgaard, Kristian; Svenning, Jens-Christian
2014-07-01
Ecological trait data are essential for understanding the broad-scale distribution of biodiversity and its response to global change. For animals, diet represents a fundamental aspect of species' evolutionary adaptations, ecological and functional roles, and trophic interactions. However, the importance of diet for macroevolutionary and macroecological dynamics remains little explored, partly because of the lack of comprehensive trait datasets. We compiled and evaluated a comprehensive global dataset of diet preferences of mammals ("MammalDIET"). Diet information was digitized from two global and cladewide data sources and errors of data entry by multiple data recorders were assessed. We then developed a hierarchical extrapolation procedure to fill-in diet information for species with missing information. Missing data were extrapolated with information from other taxonomic levels (genus, other species within the same genus, or family) and this extrapolation was subsequently validated both internally (with a jack-knife approach applied to the compiled species-level diet data) and externally (using independent species-level diet information from a comprehensive continentwide data source). Finally, we grouped mammal species into trophic levels and dietary guilds, and their species richness as well as their proportion of total richness were mapped at a global scale for those diet categories with good validation results. The success rate of correctly digitizing data was 94%, indicating that the consistency in data entry among multiple recorders was high. Data sources provided species-level diet information for a total of 2033 species (38% of all 5364 terrestrial mammal species, based on the IUCN taxonomy). For the remaining 3331 species, diet information was mostly extrapolated from genus-level diet information (48% of all terrestrial mammal species), and only rarely from other species within the same genus (6%) or from family level (8%). Internal and external
André, M; Malmström, M E; Neretnieks, I
2009-11-01
Permanent storage of spent nuclear fuel in crystalline bedrock is investigated in several countries. For this storage scenario, the host rock is the third and final barrier for radionuclide migration. Sorption reactions in the crystalline rock matrix have strong retardative effects on the transport of radionuclides. To assess the barrier properties of the host rock it is important to have sorption data representative of the undisturbed host rock conditions. Sorption data is in the majority of reported cases determined using crushed rock. Crushing has been shown to increase a rock samples sorption capacity by creating additional surfaces. There are several problems with such an extrapolation. In studies where this problem is addressed, simple models relating the specific surface area to the particle size are used to extrapolate experimental data to a value representative of the host rock conditions. In this article, we report and compare surface area data of five size fractions of crushed granite and of 100 mm long drillcores as determined by the Brunauer Emmet Teller (BET)-method using N(2)-gas. Special sample holders that could hold large specimen were developed for the BET measurements. Surface area data on rock samples as large as the drillcore has not previously been published. An analysis of this data show that the extrapolated value for intact rock obtained from measurements on crushed material was larger than the determined specific surface area of the drillcores, in some cases with more than 1000%. Our results show that the use of data from crushed material and current models to extrapolate specific surface areas for host rock conditions can lead to over estimation interpretations of sorption ability. The shortcomings of the extrapolation model are discussed and possible explanations for the deviation from experimental data are proposed. PMID:19781807
Accurate Complete Basis Set Extrapolation of Direct Random Phase Correlation Energies.
Mezei, Pál D; Csonka, Gábor I; Ruzsinszky, Adrienn
2015-08-11
The direct random phase approximation (dRPA) is a promising way to obtain improvements upon the standard semilocal density functional results in many aspects of computational chemistry. In this paper, we address the slow convergence of the calculated dRPA correlation energy with the increase of the quality and size of the popular Gaussian-type Dunning's correlation consistent aug-cc-pVXZ split valence atomic basis set family. The cardinal number X controls the size of the basis set, and we use X = 3-6 in this study. It is known that even the very expensive X = 6 basis sets lead to large errors for the dRPA correlation energy, and thus complete basis set extrapolation is necessary. We study the basis set convergence of the dRPA correlation energies on a set of 65 hydrocarbon isomers from CH4 to C6H6. We calculate the iterative density fitted dRPA correlation energies using an efficient algorithm based on the CC-like form of the equations using the self-consistent HF orbitals. We test the popular inverse cubic, the optimized exponential, and inverse power formulas for complete basis set extrapolation. We have found that the optimized inverse power based extrapolation delivers the best energies. Further analysis showed that the optimal exponent depends on the molecular structure, and the most efficient two-point energy extrapolations that use X = 3 and 4 can be improved considerably by considering the atomic composition and hybridization states of the atoms in the molecules. Our results also show that the optimized exponents that yield accurate X = 3 and 4 extrapolated dRPA energies for atoms or small molecules might be inaccurate for larger molecules. PMID:26574475
Yoshinaga, N.; Arima, A.
2010-04-15
We propose some new, efficient, and practical extrapolation methods to obtain a few low-lying eigenenergies of a large-dimensional Hamiltonian matrix in the nuclear shell model. We obtain those energies at the desired accuracy by extrapolation after diagonalizing small-dimensional submatrices of the sorted Hamiltonian matrix.
Approaches for extrapolating in vitro toxicity testing results for prediction of human in vivo outcomes are needed. The purpose of this case study was to employ in vitro toxicokinetics and PBPK modeling to perform in vitro to in vivo extrapolation (IVIVE) of lindane neurotoxicit...
Latychevskaia, Tatiana; Fink, Hans-Werner
2015-01-12
Previously reported crystalline structures obtained by an iterative phase retrieval reconstruction of their diffraction patterns seem to be free from displaying any irregularities or defects in the lattice, which appears to be unrealistic. We demonstrate here that the structure of a nanocrystal including its atomic defects can unambiguously be recovered from its diffraction pattern alone by applying a direct phase retrieval procedure not relying on prior information of the object shape. Individual point defects in the atomic lattice are clearly apparent. Conventional phase retrieval routines assume isotropic scattering. We show that when dealing with electrons, the quantitatively correct transmission function of the sample cannot be retrieved due to anisotropic, strong forward scattering specific to electrons. We summarize the conditions for this phase retrieval method and show that the diffraction pattern can be extrapolated beyond the original record to even reveal formerly not visible Bragg peaks. Such extrapolated wave field pattern leads to enhanced spatial resolution in the reconstruction.
Neutron spectroscopy results of JET high-performance plasmas and extrapolations to DT performance
Hellesen, C.; Andersson Sunden, E.; Conroy, S.; Ericsson, G.; Eriksson, J.; Gatu Johnson, M.; Weiszflog, M.; Collaboration: JET-EFDA Contributors
2010-10-15
In a fusion reactor with high energy gain, the fusion power will be mainly thermonuclear (THN). Measurements of the THN neutron rate are a good performance indicator of a fusion plasma, requiring neutron emission spectroscopy (NES) measurements to distinguish thermal and nonthermal contributions. We report here on recent NES results from JET high-performance plasmas with high fractions (about 65%) of THN emission. The analysis is made with a framework for analyzing NES data, taking into account THN reactions and beam-target reactions. The results are used to extrapolate to the equivalent DT rates. Finally, we discuss the applicability of using NES in the deuterium phase of ITER, both for the extrapolations to ITER's future DT performance as well as for the measurements of confined energetic ions.
Chiral extrapolations in 2+1 flavor domain wall fermion simulations
NASA Astrophysics Data System (ADS)
Lin, Meifeng
2006-12-01
Simulations with 2+1 flavors of domain wall fermions provide us with the opportunity to compare the lattice data directly to the predictions of continuum chiral perturbation theory, up to correc- tions from the residual chiral symmetry breaking, mres , and O(a) lattice artefacts, which are rela- tively small for domain wall fermions. We present preliminary results for the pseudoscalar meson masses and decay constants from partially quenched simulations and examine the next-to-leading order chiral extrapolations at small quark masses. The simulations were carried out on two lattice volumes : 163 × 32 and 243 × 64, with the lattice spacing fixed at about 0.1 fm. The subtleties of the chiral fits are discussed. We also explore the roles of mres and O(a) terms in the NLO chiral expansions and their effects on the chiral extrapolations for the pseudoscalar masses and decay constants.
Agarwal, Amit B; McBride, Ali
2016-08-01
The World Health Organization defines a biosimilar as "a biotherapeutic product which is similar in terms of quality, safety and efficacy to an already licensed reference biotherapeutic product." Biosimilars are biologic medical products that are very distinct from small-molecule generics, as their active substance is a biological agent derived from a living organism. Approval processes are highly regulated, with guidance issued by the European Medicines Agency and US Food and Drug Administration. Approval requires a comparability exercise consisting of extensive analytical and preclinical in vitro and in vivo studies, and confirmatory clinical studies. Extrapolation of biosimilars from their original indication to another is a feasible but highly stringent process reliant on rigorous scientific justification. This review focuses on the processes involved in gaining biosimilar approval and extrapolation and details the comparability exercise undertaken in the European Union between originator erythropoietin-stimulating agent, Eprex(®), and biosimilar, Retacrit™. PMID:27317353
Neutron spectroscopy results of JET high-performance plasmas and extrapolations to DT performance.
Hellesen, C; Andersson Sundén, E; Conroy, S; Ericsson, G; Eriksson, J; Gatu Johnson, M; Weiszflog, M
2010-10-01
In a fusion reactor with high energy gain, the fusion power will be mainly thermonuclear (THN). Measurements of the THN neutron rate are a good performance indicator of a fusion plasma, requiring neutron emission spectroscopy (NES) measurements to distinguish thermal and nonthermal contributions. We report here on recent NES results from JET high-performance plasmas with high fractions (about 65%) of THN emission. The analysis is made with a framework for analyzing NES data, taking into account THN reactions and beam-target reactions. The results are used to extrapolate to the equivalent DT rates. Finally, we discuss the applicability of using NES in the deuterium phase of ITER, both for the extrapolations to ITER’s future DT performance as well as for the measurements of confined energetic ions. PMID:21058461
New method of extrapolation of the resistance of a model planing boat to full size
NASA Technical Reports Server (NTRS)
Sottorf, W
1942-01-01
The previously employed method of extrapolating the total resistance to full size with lambda(exp 3) (model scale) and thereby foregoing a separate appraisal of the frictional resistance, was permissible for large models and floats of normal size. But faced with the ever increasing size of aircraft a reexamination of the problem of extrapolation to full size is called for. A method is described by means of which, on the basis of an analysis of tests on planing surfaces, the variation of the wetted surface over the take-off range is analytically obtained. The friction coefficients are read from Prandtl's curve for turbulent boundary layer with laminar approach. With these two values a correction for friction is obtainable.
Applications and limitations of interspecies scaling and in vitro extrapolation in pharmacokinetics.
Lin, J H
1998-12-01
The search for new drugs is an extremely time-consuming and costly endeavor. Much of the time and cost are expended on generating data that support the efficacy and safety profiles of the drug. Because of ethical constraints, relevant pharmacological and toxicological assessments must be made in laboratory animals and in in vitro systems before human testing can begin. In support of the efficacy and safety evaluation during drug development, two fundamental challenges facing industrial drug metabolism scientists are (1) how to "scale-up" the pharmacokinetic data from animals to humans and (2) how to extrapolate the in vitro data to the in vivo situation. This review examines the applications and limitations of interspecies scaling and in vitro extrapolation in pharmacokinetics. PMID:9860929
Latychevskaia, Tatiana Fink, Hans-Werner
2013-11-11
Conventional microscopic records represent intensity distributions whereby local sample information is mapped onto local information at the detector. In coherent microscopy, the superposition principle of waves holds; field amplitudes are added, not intensities. This non-local representation is spread out in space and interference information combined with wave continuity allows extrapolation beyond the actual detected data. Established resolution criteria are thus circumvented and hidden object details can retrospectively be recovered from just a fraction of an interference pattern.
NASA Astrophysics Data System (ADS)
Mazas, Franck; Hamm, Luc; Kergadallan, Xavier
2013-04-01
In France, the storm Xynthia of February 27-28th, 2010 reminded engineers and stakeholders of the necessity for an accurate estimation of extreme sea levels for the risk assessment in coastal areas. Traditionally, two main approaches exist for the statistical extrapolation of extreme sea levels: the direct approach performs a direct extrapolation on the sea level data, while the indirect approach carries out a separate analysis of the deterministic component (astronomical tide) and stochastic component (meteorological residual, or surge). When the tidal component is large compared with the surge one, the latter approach is known to perform better. In this approach, the statistical extrapolation is performed on the surge component then the distribution of extreme seal levels is obtained by convolution of the tide and surge distributions. This model is often referred to as the Joint Probability Method. Different models from the univariate extreme theory have been applied in the past for extrapolating extreme surges, in particular the Annual Maxima Method (AMM) and the r-largest method. In this presentation, we apply the Peaks-Over-Threshold (POT) approach for declustering extreme surge events, coupled with the Poisson-GPD model for fitting extreme surge peaks. This methodology allows a sound estimation of both lower and upper tails of the stochastic distribution, including the estimation of the uncertainties associated to the fit by computing the confidence intervals. After convolution with the tide signal, the model yields the distribution for the whole range of possible sea level values. Particular attention is paid to the necessary distinction between sea level values observed at a regular time step, such as hourly, and sea level events, such as those occurring during a storm. Extremal indexes for both surges and levels are thus introduced. This methodology will be illustrated with a case study at Brest, France.
A Spatial Extrapolation Approach to Assess the Impact of Climate Change on Water Resource Systems
NASA Astrophysics Data System (ADS)
Pina, J.; Tilmant, A.; Anctil, F.
2015-12-01
The typical approach to assess climate change impacts on water resources systems is based on a vertical integration/coupling of models: GCM models are run to project future precipitations and temperatures, which are then downscaled and used as inputs to hydrologic models whose outputs are processed by water systems models. From a decision-making point of view, this top-down vertical approach presents some challenges. For example, since the range of uncertainty that can be explored with GCM is limited, researchers are relying on ensembles to enlarge the spread, making the modeling approach even more demanding in terms of computation time and resource. When a particular water system must be analyzed, the question is to know whether this computationally intensive vertical approach is necessary in the first place or if we could extrapolate projections available in neighboring systems to feed the water system model? This would be equivalent to a horizontal approach. The proposed study addresses this question by comparing the performance of a water resource system under future climate conditions using the vertical and horizontal approaches. The methodology is illustrated with the hydropower system of the Gatineau River Basin in Quebec, Canada. Vertically obtained hydrologic projections available in those river basins are extrapolated and used as inputs to a stochastic multireservoir optimization model. Two different extrapolation techniques are tested. The first one simply relies on the ratios between the drainage areas. The second exploits the covariance structure found in historical flow data throughout the region. The analysis of the simulation results reveals that the annual and weekly energy productions of the system derived from the horizontal approach are statistically equivalent to those obtained with the vertical one, regardless of the extrapolation technique used.
NASA Astrophysics Data System (ADS)
Steinhausen, Heinz C.; Martín, Rodrigo; den Brok, Dennis; Hullin, Matthias B.; Klein, Reinhard
2015-03-01
Numerous applications in computer graphics and beyond benefit from accurate models for the visual appearance of real-world materials. Data-driven models like photographically acquired bidirectional texture functions (BTFs) suffer from limited sample sizes enforced by the common assumption of far-field illumination. Several materials like leather, structured wallpapers or wood contain structural elements on scales not captured by typical BTF measurements. We propose a method extending recent research by Steinhausen et al. to extrapolate BTFs for large-scale material samples from a measured and compressed BTF for a small fraction of the material sample, guided by a set of constraints. We propose combining color constraints with surface descriptors similar to normal maps as part of the constraints guiding the extrapolation process. This helps narrowing down the search space for suitable ABRDFs per texel to a large extent. To acquire surface descriptors for nearly at materials, we build upon the idea of photometrically estimating normals. Inspired by recent work by Pan and Skala, we obtain images of the sample in four different rotations with an off-the-shelf flatbed scanner and derive surface curvature information from these. Furthermore, we simplify the extrapolation process by using a pixel-based texture synthesis scheme, reaching computational efficiency similar to texture optimization.
Interspecies extrapolation for mammals and birds: What`s a risk assessor to do?
Norton, S.R.; Rhomberg, L.
1995-12-31
Ecological risk assessors are often faced with the need to extrapolate laboratory data on a particular species to another species of interest. For aquatic organisms, an extensive database is available that has been used to develop empirical intertaxa extrapolation factors. For birds and mammals, however, such a database does not exist. This presentation explores the strengths and limitations of using knowledge of allometric scaling to address issues of interspecies differences in toxic response. In particular, the authors focus on approaches for oral exposure. Most approaches to interspecies extrapolation begin with implicit or explicit assumptions about how to express equivalent doses for different body sizes. It is in these calculations that knowledge of allometric scaling has its greatest applicability. Options include adjusting dose for equivalent organ weights (i.e., approximately BW{sup 1}), equivalent surface area (i.e., approximately BW{sup 2/3}), or equivalent metabolic rate (i.e., approximately BW{sup 3/4}). The authors will discuss the basis for each of these options, precedents for their use, and implications for analyzing ecotoxicological data.
NASA Astrophysics Data System (ADS)
Moraitis, Kostas; Archontis, Vasilis; Tziotziou, Konstantinos; Georgoulis, Manolis K.
We calculate the instantaneous free magnetic energy and relative magnetic helicity of solar active regions using two independent approaches: a) a non-linear force-free (NLFF) method that requires only a single photospheric vector magnetogram, and b) well known semi-analytical formulas that require the full three-dimensional (3D) magnetic field structure. The 3D field is obtained either from MHD simulations, or from observed magnetograms via respective NLFF field extrapolations. We find qualitative agreement between the two methods and, quantitatively, a discrepancy not exceeding a factor of 4. The comparison of the two methods reveals, as a byproduct, two independent tests for the quality of a given force-free field extrapolation. We find that not all extrapolations manage to achieve the force-free condition in a valid, divergence-free, magnetic configuration. This research has been co-financed by the European Union (European Social Fund - ESF) and Greek national funds through the Operational Program "Education and Lifelong Learning" of the National Strategic Reference Framework (NSRF) - Research Funding Program: Thales. Investing in knowledge society through the European Social Fund.
NASA Astrophysics Data System (ADS)
Guo, J. Y.; Li, Y. B.; Dai, C. L.; Shum, C. K.
2013-10-01
We present a technique to improve the least squares (LS) extrapolation of Earth orientation parameters (EOPs), consisting of fixing the last observed data point on the LS extrapolation curve, which customarily includes a polynomial and a few sinusoids. For the polar motion (PM), a more sophisticated two steps approach has been developed, which consists of estimating the amplitude of the more stable one of the annual (AW) and Chandler (CW) wobbles using data of longer time span, and then estimating the other parameters using a shorter time span. The technique is studied using hindcast experiments, and justified using year-by-year statistics of 8 years. In order to compare with the official predictions of the International Earth Rotation and Reference Systems Service (IERS) performed at the U.S. Navy Observatory (USNO), we have enforced short-term predictions by applying the ARIMA method to the residuals computed by subtracting the LS extrapolation curve from the observation data. The same as at USNO, we have also used atmospheric excitation function (AEF) to further improve predictions of UT1-UTC. As results, our short-term predictions are comparable to the USNO predictions, and our long-term predictions are marginally better, although not for every year. In addition, we have tested the use of AEF and oceanic excitation function (OEF) in PM prediction. We find that use of forecasts of AEF alone does not lead to any apparent improvement or worsening, while use of forecasts of AEF + OEF does lead to apparent improvement.
A model for the data extrapolation of greenhouse gas emissions in the Brazilian hydroelectric system
NASA Astrophysics Data System (ADS)
Pinguelli Rosa, Luiz; Aurélio dos Santos, Marco; Gesteira, Claudio; Elias Xavier, Adilson
2016-06-01
Hydropower reservoirs are artificial water systems and comprise a small proportion of the Earth’s continental territory. However, they play an important role in the aquatic biogeochemistry and may affect the environment negatively. Since the 90s, as a result of research on organic matter decay in manmade flooded areas, some reports have associated greenhouse gas emissions with dam construction. Pioneering work carried out in the early period challenged the view that hydroelectric plants generate completely clean energy. Those estimates suggested that GHG emissions into the atmosphere from some hydroelectric dams may be significant when measured per unit of energy generated and should be compared to GHG emissions from fossil fuels used for power generation. The contribution to global warming of greenhouse gases emitted by hydropower reservoirs is currently the subject of various international discussions and debates. One of the most controversial issues is the extrapolation of data from different sites. In this study, the extrapolation from a site sample where measurements were made to the complete set of 251 reservoirs in Brazil, comprising a total flooded area of 32 485 square kilometers, was derived from the theory of self-organized criticality. We employed a power law for its statistical representation. The present article reviews the data generated at that time in order to demonstrate how, with the help of mathematical tools, we can extrapolate values from one reservoir to another without compromising the reliability of the results.
Landis, W.G.
1995-12-31
One of the central problems in environmental toxicology has been the extrapolation from laboratory tests to the field and from biomonitoring results to ecological impacts. The crossing of the boundary from molecular mechanisms to population impacts has always been difficult. Perhaps the problem in extrapolation is not so much the effects of physical scale as much as the transition boundary between two different types of systems, organismal and non-organismal. The basic properties of these systems are quite distinct. Organismal systems possess a central core of information, subject to natural selection, that can impose homeostasis (body temperature) or diversity (immune system) upon the constituents of that system. Unless there are changes in the genetic structure of the germ line, impacts to the somatic cells and structure of the organism are erased upon the establishment of a new generation. The integrity of the germplasm means that organismal systems are largely a historical. In contrast, non-organismal systems contain no central and inheritable repository of information, analogous to the genome, that serves as the blueprint for an ecological system. Non-organismal systems are historical in the terminology of complex systems. The irreversibility and historical nature of ecological systems has also been observed experimentally. Historical events and the derived heterogeneity in the field must be taken into account when the extrapolations are conducted. Genetic structure of the populations, the current spatial distribution of species, and the physical structure of the system must all be taken into account if accurate forecasts from experimental results can be made.
Study on Two Methods for Nonlinear Force-Free Extrapolation Based on Semi-Analytical Field
NASA Astrophysics Data System (ADS)
Liu, S.; Zhang, H. Q.; Su, J. T.; Song, M. T.
2011-03-01
In this paper, two semi-analytical solutions of force-free fields (Low and Lou, Astrophys. J. 352, 343, 1990) have been used to test two nonlinear force-free extrapolation methods. One is the boundary integral equation (BIE) method developed by Yan and Sakurai ( Solar Phys. 195, 89, 2000), and the other is the approximate vertical integration (AVI) method developed by Song et al. ( Astrophys. J. 649, 1084, 2006). Some improvements have been made to the AVI method to avoid the singular points in the process of calculation. It is found that the correlation coefficients between the first semi-analytical field and extrapolated field using the BIE method, and also that obtained by the improved AVI method, are greater than 90% below a height 10 of the 64×64 lower boundary. For the second semi-analytical field, these correlation coefficients are greater than 80% below the same relative height. Although differences between the semi-analytical solutions and the extrapolated fields exist for both the BIE and AVI methods, these two methods can give reliable results for heights of about 15% of the extent of the lower boundary.
NASA Astrophysics Data System (ADS)
Zhang, C.; Chuckpaiwong, I.; Liang, S. Y.; Seth, B. B.
2002-07-01
Life testing under nominal operating conditions of mechanical parts with high mean lifetime between failure (MTBF) often consumes a significant amount of time and resources, rendering such procedures expensive and impractical. As a result, the technology of accelerated life testing (ALT) has been developed for testing at high stress levels (e.g. temperature, voltage, pressure, corrosive media, load, vibration amplitude, etc.) so that it can be extrapolated—through a physically reasonable statistical model—to obtain estimations of life at lower, normal stress levels or even limit stress levels. However, the issue of prediction accuracy associated with extrapolating data outside the range of testing, or even to a singularity level (no stress), has not yet been fully addressed. In this research, an accelerator factor is introduced into an inverse power law model to estimate the life distribution in terms of time and stresses. Also, a generalized Eyring model is set up for singularity extrapolation in handling limit stress level conditions. The procedure to calibrate the associated shape factors based on the maximum likelihood principle is also formulated. The methodology implementation, based on a one-main-step, multiple-step-stress test scheme, is experimentally illustrated with tapered roller bearing under the stress of environmental corrosion as a case study. The experimental results show that the developed accelerated life test model can effectively evaluate the life probability of a bearing based on accelerated testing data when extrapolating to the stress levels within or outside the range of testing.
Multi-state extrapolation of UV/Vis absorption spectra with QM/QM hybrid methods
NASA Astrophysics Data System (ADS)
Ren, Sijin; Caricato, Marco
2016-05-01
In this work, we present a simple approach to simulate absorption spectra from hybrid QM/QM calculations. The goal is to obtain reliable spectra for compounds that are too large to be treated efficiently at a high level of theory. The present approach is based on the extrapolation of the entire absorption spectrum obtained by individual subcalculations. Our program locates the main spectral features in each subcalculation, e.g., band peaks and shoulders, and fits them to Gaussian functions. Each Gaussian is then extrapolated with a formula similar to that of ONIOM (Our own N-layered Integrated molecular Orbital molecular Mechanics). However, information about individual excitations is not necessary so that difficult state-matching across subcalculations is avoided. This multi-state extrapolation thus requires relatively low implementation effort while affording maximum flexibility in the choice of methods to be combined in the hybrid approach. The test calculations show the efficacy and robustness of this methodology in reproducing the spectrum computed for the entire molecule at a high level of theory.
Multi-state extrapolation of UV/Vis absorption spectra with QM/QM hybrid methods.
Ren, Sijin; Caricato, Marco
2016-05-14
In this work, we present a simple approach to simulate absorption spectra from hybrid QM/QM calculations. The goal is to obtain reliable spectra for compounds that are too large to be treated efficiently at a high level of theory. The present approach is based on the extrapolation of the entire absorption spectrum obtained by individual subcalculations. Our program locates the main spectral features in each subcalculation, e.g., band peaks and shoulders, and fits them to Gaussian functions. Each Gaussian is then extrapolated with a formula similar to that of ONIOM (Our own N-layered Integrated molecular Orbital molecular Mechanics). However, information about individual excitations is not necessary so that difficult state-matching across subcalculations is avoided. This multi-state extrapolation thus requires relatively low implementation effort while affording maximum flexibility in the choice of methods to be combined in the hybrid approach. The test calculations show the efficacy and robustness of this methodology in reproducing the spectrum computed for the entire molecule at a high level of theory. PMID:27179466
Testing magnetofrictional extrapolation with the Titov-Démoulin model of solar active regions
NASA Astrophysics Data System (ADS)
Valori, G.; Kliem, B.; Török, T.; Titov, V. S.
2010-09-01
We examine the nonlinear magnetofrictional extrapolation scheme using the solar active region model by Titov and Démoulin as test field. This model consists of an arched, line-tied current channel held in force-free equilibrium by the potential field of a bipolar flux distribution in the bottom boundary. A modified version with a parabolic current density profile is employed here. We find that the equilibrium is reconstructed with very high accuracy in a representative range of parameter space, using only the vector field in the bottom boundary as input. Structural features formed in the interface between the flux rope and the surrounding arcade - “hyperbolic flux tube” and “bald patch separatrix surface” - are reliably reproduced, as are the flux rope twist and the energy and helicity of the configuration. This demonstrates that force-free fields containing these basic structural elements of solar active regions can be obtained by extrapolation. The influence of the chosen initial condition on the accuracy of reconstruction is also addressed, confirming that the initial field that best matches the external potential field of the model quite naturally leads to the best reconstruction. Extrapolating the magnetogram of a Titov-Démoulin equilibrium in the unstable range of parameter space yields a sequence of two opposing evolutionary phases, which clearly indicate the unstable nature of the configuration: a partial buildup of the flux rope with rising free energy is followed by destruction of the rope, losing most of the free energy.
IMPEDANCE OF FINITE LENGTH RESISTOR
KRINSKY, S.; PODOBEDOV, B.; GLUCKSTERN, R.L.
2005-05-15
We determine the impedance of a cylindrical metal tube (resistor) of radius a, length g, and conductivity {sigma}, attached at each end to perfect conductors of semi-infinite length. Our main interest is in the asymptotic behavior of the impedance at high frequency, k >> 1/a. In the equilibrium regime, , the impedance per unit length is accurately described by the well-known result for an infinite length tube with conductivity {sigma}. In the transient regime, ka{sup 2} >> g, we derive analytic expressions for the impedance and wakefield.
López-Mondéjar, Rubén; Antón, Anabel; Raidl, Stefan; Ros, Margarita; Pascual, José Antonio
2010-04-01
The species of the genus Trichoderma are used successfully as biocontrol agents against a wide range of phytopathogenic fungi. Among them, Trichoderma harzianum is especially effective. However, to develop more effective fungal biocontrol strategies in organic substrates and soil, tools for monitoring the control agents are required. Real-time PCR is potentially an effective tool for the quantification of fungi in environmental samples. The aim of this study consisted of the development and application of a real-time PCR-based method to the quantification of T. harzianum, and the extrapolation of these data to fungal biomass values. A set of primers and a TaqMan probe for the ITS region of the fungal genome were designed and tested, and amplification was correlated to biomass measurements obtained with optical microscopy and image analysis, of the hyphal length of the mycelium of the colony. A correlation of 0.76 between ITS copies and biomass was obtained. The extrapolation of the quantity of ITS copies, calculated based on real-time PCR data, into quantities of fungal biomass provides potentially a more accurate value of the quantity of soil fungi. PMID:19897358
SU-E-J-145: Geometric Uncertainty in CBCT Extrapolation for Head and Neck Adaptive Radiotherapy
Liu, C; Kumarasiri, A; Chetvertkov, M; Gordon, J; Chetty, I; Siddiqui, F; Kim, J
2014-06-01
Purpose: One primary limitation of using CBCT images for H'N adaptive radiotherapy (ART) is the limited field of view (FOV) range. We propose a method to extrapolate the CBCT by using a deformed planning CT for the dose of the day calculations. The aim was to estimate the geometric uncertainty of our extrapolation method. Methods: Ten H'N patients, each with a planning CT (CT1) and a subsequent CT (CT2) taken, were selected. Furthermore, a small FOV CBCT (CT2short) was synthetically created by cropping CT2 to the size of a CBCT image. Then, an extrapolated CBCT (CBCTextrp) was generated by deformably registering CT1 to CT2short and resampling with a wider FOV (42mm more from the CT2short borders), where CT1 is deformed through translation, rigid, affine, and b-spline transformations in order. The geometric error is measured as the distance map ||DVF|| produced by a deformable registration between CBCTextrp and CT2. Mean errors were calculated as a function of the distance away from the CBCT borders. The quality of all the registrations was visually verified. Results: Results were collected based on the average numbers from 10 patients. The extrapolation error increased linearly as a function of the distance (at a rate of 0.7mm per 1 cm) away from the CBCT borders in the S/I direction. The errors (μ±σ) at the superior and inferior boarders were 0.8 ± 0.5mm and 3.0 ± 1.5mm respectively, and increased to 2.7 ± 2.2mm and 5.9 ± 1.9mm at 4.2cm away. The mean error within CBCT borders was 1.16 ± 0.54mm . The overall errors within 4.2cm error expansion were 2.0 ± 1.2mm (sup) and 4.5 ± 1.6mm (inf). Conclusion: The overall error in inf direction is larger due to more large unpredictable deformations in the chest. The error introduced by extrapolation is plan dependent. The mean error in the expanded region can be large, and must be considered during implementation. This work is supported in part by Varian Medical Systems, Palo Alto, CA.
Motion-based prediction explains the role of tracking in motion extrapolation.
Khoei, Mina A; Masson, Guillaume S; Perrinet, Laurent U
2013-11-01
During normal viewing, the continuous stream of visual input is regularly interrupted, for instance by blinks of the eye. Despite these frequents blanks (that is the transient absence of a raw sensory source), the visual system is most often able to maintain a continuous representation of motion. For instance, it maintains the movement of the eye such as to stabilize the image of an object. This ability suggests the existence of a generic neural mechanism of motion extrapolation to deal with fragmented inputs. In this paper, we have modeled how the visual system may extrapolate the trajectory of an object during a blank using motion-based prediction. This implies that using a prior on the coherency of motion, the system may integrate previous motion information even in the absence of a stimulus. In order to compare with experimental results, we simulated tracking velocity responses. We found that the response of the motion integration process to a blanked trajectory pauses at the onset of the blank, but that it quickly recovers the information on the trajectory after reappearance. This is compatible with behavioral and neural observations on motion extrapolation. To understand these mechanisms, we have recorded the response of the model to a noisy stimulus. Crucially, we found that motion-based prediction acted at the global level as a gain control mechanism and that we could switch from a smooth regime to a binary tracking behavior where the dot is tracked or lost. Our results imply that a local prior implementing motion-based prediction is sufficient to explain a large range of neural and behavioral results at a more global level. We show that the tracking behavior deteriorates for sensory noise levels higher than a certain value, where motion coherency and predictability fail to hold longer. In particular, we found that motion-based prediction leads to the emergence of a tracking behavior only when enough information from the trajectory has been accumulated
Line Lengths and Starch Scores.
ERIC Educational Resources Information Center
Moriarty, Sandra E.
1986-01-01
Investigates readability of different line lengths in advertising body copy, hypothesizing a normal curve with lower scores for shorter and longer lines, and scores above the mean for lines in the middle of the distribution. Finds support for lower scores for short lines and some evidence of two optimum line lengths rather than one. (SKC)
Spreading lengths of Hermite polynomials
NASA Astrophysics Data System (ADS)
Sánchez-Moreno, P.; Dehesa, J. S.; Manzano, D.; Yáñez, R. J.
2010-03-01
The Renyi, Shannon and Fisher spreading lengths of the classical or hypergeometric orthogonal polynomials, which are quantifiers of their distribution all over the orthogonality interval, are defined and investigated. These information-theoretic measures of the associated Rakhmanov probability density, which are direct measures of the polynomial spreading in the sense of having the same units as the variable, share interesting properties: invariance under translations and reflections, linear scaling and vanishing in the limit that the variable tends towards a given definite value. The expressions of the Renyi and Fisher lengths for the Hermite polynomials are computed in terms of the polynomial degree. The combinatorial multivariable Bell polynomials, which are shown to characterize the finite power of an arbitrary polynomial, play a relevant role for the computation of these information-theoretic lengths. Indeed these polynomials allow us to design an error-free computing approach for the entropic moments (weighted Lq-norms) of Hermite polynomials and subsequently for the Renyi and Tsallis entropies, as well as for the Renyi spreading lengths. Sharp bounds for the Shannon length of these polynomials are also given by means of an information-theoretic-based optimization procedure. Moreover, the existence of a linear correlation between the Shannon length (as well as the second-order Renyi length) and the standard deviation is computationally proved. Finally, the application to the most popular quantum-mechanical prototype system, the harmonic oscillator, is discussed and some relevant asymptotical open issues related to the entropic moments, mentioned previously, are posed.
Latychevskaia, Tatiana Fink, Hans-Werner; Chushkin, Yuriy; Zontone, Federico
2015-11-02
Coherent diffraction imaging is a high-resolution imaging technique whose potential can be greatly enhanced by applying the extrapolation method presented here. We demonstrate the enhancement in resolution of a non-periodical object reconstructed from an experimental X-ray diffraction record which contains about 10% missing information, including the pixels in the center of the diffraction pattern. A diffraction pattern is extrapolated beyond the detector area and as a result, the object is reconstructed at an enhanced resolution and better agreement with experimental amplitudes is achieved. The optimal parameters for the iterative routine and the limits of the extrapolation procedure are discussed.
When does length cause the word length effect?
Jalbert, Annie; Neath, Ian; Bireta, Tamra J; Surprenant, Aimée M
2011-03-01
The word length effect, the finding that lists of short words are better recalled than lists of long words, has been termed one of the benchmark findings that any theory of immediate memory must account for. Indeed, the effect led directly to the development of working memory and the phonological loop, and it is viewed as the best remaining evidence for time-based decay. However, previous studies investigating this effect have confounded length with orthographic neighborhood size. In the present study, Experiments 1A and 1B revealed typical effects of length when short and long words were equated on all relevant dimensions previously identified in the literature except for neighborhood size. In Experiment 2, consonant-vowel-consonant (CVC) words with a large orthographic neighborhood were better recalled than were CVC words with a small orthographic neighborhood. In Experiments 3 and 4, using two different sets of stimuli, we showed that when short (1-syllable) and long (3-syllable) items were equated for neighborhood size, the word length effect disappeared. Experiment 5 replicated this with spoken recall. We suggest that the word length effect may be better explained by the differences in linguistic and lexical properties of short and long words rather than by length per se. These results add to the growing literature showing problems for theories of memory that include decay offset by rehearsal as a central feature. PMID:21171805
Length dependence of thermal conductivity by approach-to-equilibrium molecular dynamics
NASA Astrophysics Data System (ADS)
Zaoui, Hayat; Palla, Pier Luca; Cleri, Fabrizio; Lampin, Evelyne
2016-08-01
The length dependence of thermal conductivity over more than two orders of magnitude has been systematically studied for a range of materials, interatomic potentials, and temperatures using the atomistic approach-to-equilibrium molecular dynamics (AEMD) method. By comparing the values of conductivity obtained for a given supercell length and maximum phonon mean free path (MFP), we find that such values are strongly correlated, demonstrating that the AEMD calculation with a supercell of finite length actually probes the thermal conductivity corresponding to a maximum phonon MFP. As a consequence, the less pronounced length dependence usually observed for poorer thermal conductors, such as amorphous silica, is physically justified by their shorter average phonon MFP. Finally, we compare different analytical extrapolations of the conductivity to infinite length and demonstrate that the frequently used Matthiessen rule is not applicable in AEMD. An alternative extrapolation more suitable for transient-time, finite-supercell simulations is derived. This approximation scheme can also be used to classify the quality of different interatomic potential models with respect to their capability of predicting the experimental thermal conductivity.
Quantification of sarcomere length distribution in whole muscle frozen sections.
O'Connor, Shawn M; Cheng, Elton J; Young, Kevin W; Ward, Samuel R; Lieber, Richard L
2016-05-15
Laser diffraction (LD) is a valuable tool for measuring sarcomere length (Ls), a major determinant of muscle function. However, this method relies on few measurements per sample that are often extrapolated to whole muscle properties. Currently it is not possible to measure Ls throughout an entire muscle and determine how Ls varies at this scale. To address this issue, we developed an actuated LD scanner for sampling large numbers of sarcomeres in thick whole muscle longitudinal sections. Sections of high optical quality and fixation were produced from tibialis anterior and extensor digitorum longus muscles of Sprague-Dawley rats (N=6). Scans produced two-dimensional Ls maps, capturing >85% of the muscle area per section. Individual Ls measures generated by automatic LD and bright-field microscopy showed excellent agreement over a large Ls range (ICC>0.93). Two-dimensional maps also revealed prominent regional Ls variations across muscles. PMID:26994184
Beresford, Nicholas A; Wood, Michael D; Vives i Batlle, Jordi; Yankovich, Tamara L; Bradshaw, Clare; Willey, Neil
2016-01-01
We will never have data to populate all of the potential radioecological modelling parameters required for wildlife assessments. Therefore, we need robust extrapolation approaches which allow us to make best use of our available knowledge. This paper reviews and, in some cases, develops, tests and validates some of the suggested extrapolation approaches. The concentration ratio (CRproduct-diet or CRwo-diet) is shown to be a generic (trans-species) parameter which should enable the more abundant data for farm animals to be applied to wild species. An allometric model for predicting the biological half-life of radionuclides in vertebrates is further tested and generally shown to perform acceptably. However, to fully exploit allometry we need to understand why some elements do not scale to expected values. For aquatic ecosystems, the relationship between log10(a) (a parameter from the allometric relationship for the organism-water concentration ratio) and log(Kd) presents a potential opportunity to estimate concentration ratios using Kd values. An alternative approach to the CRwo-media model proposed for estimating the transfer of radionuclides to freshwater fish is used to satisfactorily predict activity concentrations in fish of different species from three lakes. We recommend that this approach (REML modelling) be further investigated and developed for other radionuclides and across a wider range of organisms and ecosystems. Ecological stoichiometry shows potential as an extrapolation method in radioecology, either from one element to another or from one species to another. Although some of the approaches considered require further development and testing, we demonstrate the potential to significantly improve predictions of radionuclide transfer to wildlife by making better use of available data. PMID:25850783
Neural Extrapolation of Motion for a Ball Rolling Down an Inclined Plane
La Scaleia, Barbara; Lacquaniti, Francesco; Zago, Myrka
2014-01-01
It is known that humans tend to misjudge the kinematics of a target rolling down an inclined plane. Because visuomotor responses are often more accurate and less prone to perceptual illusions than cognitive judgments, we asked the question of how rolling motion is extrapolated for manual interception or drawing tasks. In three experiments a ball rolled down an incline with kinematics that differed as a function of the starting position (4 different positions) and slope (30°, 45° or 60°). In Experiment 1, participants had to punch the ball as it fell off the incline. In Experiment 2, the ball rolled down the incline but was stopped at the end; participants were asked to imagine that the ball kept moving and to punch it. In Experiment 3, the ball rolled down the incline and was stopped at the end; participants were asked to draw with the hand in air the trajectory that would be described by the ball if it kept moving. We found that performance was most accurate when motion of the ball was visible until interception and haptic feedback of hand-ball contact was available (Experiment 1). However, even when participants punched an imaginary moving ball (Experiment 2) or drew in air the imaginary trajectory (Experiment 3), they were able to extrapolate to some extent global aspects of the target motion, including its path, speed and arrival time. We argue that the path and kinematics of a ball rolling down an incline can be extrapolated surprisingly well by the brain using both visual information and internal models of target motion. PMID:24940874
Scotcher, Daniel; Jones, Christopher; Posada, Maria; Galetin, Aleksandra; Rostami-Hodjegan, Amin
2016-09-01
It is envisaged that application of mechanistic models will improve prediction of changes in renal disposition due to drug-drug interactions, genetic polymorphism in enzymes and transporters and/or renal impairment. However, developing and validating mechanistic kidney models is challenging due to the number of processes that may occur (filtration, secretion, reabsorption and metabolism) in this complex organ. Prediction of human renal drug disposition from preclinical species may be hampered by species differences in the expression and activity of drug metabolising enzymes and transporters. A proposed solution is bottom-up prediction of pharmacokinetic parameters based on in vitro-in vivo extrapolation (IVIVE), mediated by recent advances in in vitro experimental techniques and application of relevant scaling factors. This review is a follow-up to the Part I of the report from the 2015 AAPS Annual Meeting and Exhibition (Orlando, FL; 25th-29th October 2015) which focuses on IVIVE and mechanistic prediction of renal drug disposition. It describes the various mechanistic kidney models that may be used to investigate renal drug disposition. Particular attention is given to efforts that have attempted to incorporate elements of IVIVE. In addition, the use of mechanistic models in prediction of renal drug-drug interactions and potential for application in determining suitable adjustment of dose in kidney disease are discussed. The need for suitable clinical pharmacokinetics data for the purposes of delineating mechanistic aspects of kidney models in various scenarios is highlighted. PMID:27506526
Model of a realistic InP surface quantum dot extrapolated from atomic force microscopy results
NASA Astrophysics Data System (ADS)
Barettin, Daniele; De Angelis, Roberta; Prosposito, Paolo; Auf der Maur, Matthias; Casalboni, Mauro; Pecchia, Alessandro
2014-05-01
We report on numerical simulations of a zincblende InP surface quantum dot (QD) on \\text{I}{{\\text{n}}_{0.48}}\\text{G}{{\\text{a}}_{0.52}}\\text{P} buffer. Our model is strictly based on experimental structures, since we extrapolated a three-dimensional dot directly by atomic force microscopy results. Continuum electromechanical, \\vec{k}\\;\\cdot \\;\\vec{p} bandstructure and optical calculations are presented for this realistic structure, together with benchmark calculations for a lens-shape QD with the same radius and height of the extrapolated dot. Interesting similarities and differences are shown by comparing the results obtained with the two different structures, leading to the conclusion that the use of a more realistic structure can provide significant improvements in the modeling of QDs fact, the remarkable splitting for the electron p-like levels of the extrapolated dot seems to prove that a realistic experimental structure can reproduce the right symmetry and a correct splitting usually given by atomistic calculations even within the multiband \\vec{k}\\;\\cdot \\;\\vec{p} approach. Moreover, the energy levels and the symmetry of the holes are strongly dependent on the shape of the dot. In particular, as far as we know, their wave function symmetries do not seem to resemble to any results previously obtained with simulations of zincblende ideal structures, such as lenses or truncated pyramids. The magnitude of the oscillator strengths is also strongly dependent on the shape of the dot, showing a lower intensity for the extrapolated dot, especially for the transition between the electrons and holes ground state, as a result of a relevant reduction of the wave functions overlap. We also compare an experimental photoluminescence spectrum measured on an homogeneous sample containing about 60 dots with a numerical ensemble average derived from single dot calculations. The broader energy range of the numerical spectrum motivated us to perform further
Model of a realistic InP surface quantum dot extrapolated from atomic force microscopy results.
Barettin, Daniele; De Angelis, Roberta; Prosposito, Paolo; Auf der Maur, Matthias; Casalboni, Mauro; Pecchia, Alessandro
2014-05-16
We report on numerical simulations of a zincblende InP surface quantum dot (QD) on In₀.₄₈Ga₀.₅₂ buffer. Our model is strictly based on experimental structures, since we extrapolated a three-dimensional dot directly by atomic force microscopy results. Continuum electromechanical, [Formula: see text] bandstructure and optical calculations are presented for this realistic structure, together with benchmark calculations for a lens-shape QD with the same radius and height of the extrapolated dot. Interesting similarities and differences are shown by comparing the results obtained with the two different structures, leading to the conclusion that the use of a more realistic structure can provide significant improvements in the modeling of QDs fact, the remarkable splitting for the electron p-like levels of the extrapolated dot seems to prove that a realistic experimental structure can reproduce the right symmetry and a correct splitting usually given by atomistic calculations even within the multiband [Formula: see text] approach. Moreover, the energy levels and the symmetry of the holes are strongly dependent on the shape of the dot. In particular, as far as we know, their wave function symmetries do not seem to resemble to any results previously obtained with simulations of zincblende ideal structures, such as lenses or truncated pyramids. The magnitude of the oscillator strengths is also strongly dependent on the shape of the dot, showing a lower intensity for the extrapolated dot, especially for the transition between the electrons and holes ground state, as a result of a relevant reduction of the wave functions overlap. We also compare an experimental photoluminescence spectrum measured on an homogeneous sample containing about 60 dots with a numerical ensemble average derived from single dot calculations. The broader energy range of the numerical spectrum motivated us to perform further verifications, which have clarified some aspects of the experimental
Challenges for In vitro to in Vivo Extrapolation of Nanomaterial Dosimetry for Human Risk Assessment
Smith, Jordan N.
2013-11-01
The proliferation in types and uses of nanomaterials in consumer products has led to rapid application of conventional in vitro approaches for hazard identification. Unfortunately, assumptions pertaining to experimental design and interpretation for studies with chemicals are not generally appropriate for nanomaterials. The fate of nanomaterials in cell culture media, cellular dose to nanomaterials, cellular dose to nanomaterial byproducts, and intracellular fate of nanomaterials at the target site of toxicity all must be considered in order to accurately extrapolate in vitro results to reliable predictions of human risk.
3D Drop Size Distribution Extrapolation Algorithm Using a Single Disdrometer
NASA Technical Reports Server (NTRS)
Lane, John
2012-01-01
Determining the Z-R relationship (where Z is the radar reflectivity factor and R is rainfall rate) from disdrometer data has been and is a common goal of cloud physicists and radar meteorology researchers. The usefulness of this quantity has traditionally been limited since radar represents a volume measurement, while a disdrometer corresponds to a point measurement. To solve that problem, a 3D-DSD (drop-size distribution) method of determining an equivalent 3D Z-R was developed at the University of Central Florida and tested at the Kennedy Space Center, FL. Unfortunately, that method required a minimum of three disdrometers clustered together within a microscale network (.1-km separation). Since most commercial disdrometers used by the radar meteorology/cloud physics community are high-cost instruments, three disdrometers located within a microscale area is generally not a practical strategy due to the limitations of these kinds of research budgets. A relatively simple modification to the 3D-DSD algorithm provides an estimate of the 3D-DSD and therefore, a 3D Z-R measurement using a single disdrometer. The basis of the horizontal extrapolation is mass conservation of a drop size increment, employing the mass conservation equation. For vertical extrapolation, convolution of a drop size increment using raindrop terminal velocity is used. Together, these two independent extrapolation techniques provide a complete 3DDSD estimate in a volume around and above a single disdrometer. The estimation error is lowest along a vertical plane intersecting the disdrometer position in the direction of wind advection. This work demonstrates that multiple sensors are not required for successful implementation of the 3D interpolation/extrapolation algorithm. This is a great benefit since it is seldom that multiple sensors in the required spatial arrangement are available for this type of analysis. The original software (developed at the University of Central Florida, 1998.- 2000) has
NASA Astrophysics Data System (ADS)
Kaltenboeck, Rudolf; Kerschbaum, Markus; Hennermann, Karin; Mayer, Stefan
2013-04-01
Nowcasting of precipitation events, especially thunderstorm events or winter storms, has high impact on flight safety and efficiency for air traffic management. Future strategic planning by air traffic control will result in circumnavigation of potential hazardous areas, reduction of load around efficiency hot spots by offering alternatives, increase of handling capacity, anticipation of avoidance manoeuvres and increase of awareness before dangerous areas are entered by aircraft. To facilitate this rapid update forecasts of location, intensity, size, movement and development of local storms are necessary. Weather radar data deliver precipitation analysis of high temporal and spatial resolution close to real time by using clever scanning strategies. These data are the basis to generate rapid update forecasts in a time frame up to 2 hours and more for applications in aviation meteorological service provision, such as optimizing safety and economic impact in the context of sub-scale phenomena. On the basis of tracking radar echoes by correlation the movement vectors of successive weather radar images are calculated. For every new successive radar image a set of ensemble precipitation fields is collected by using different parameter sets like pattern match size, different time steps, filter methods and an implementation of history of tracking vectors and plausibility checks. This method considers the uncertainty in rain field displacement and different scales in time and space. By validating manually a set of case studies, the best verification method and skill score is defined and implemented into an online-verification scheme which calculates the optimized forecasts for different time steps and different areas by using different extrapolation ensemble members. To get information about the quality and reliability of the extrapolation process additional information of data quality (e.g. shielding in Alpine areas) is extrapolated and combined with an extrapolation
On the distance to Cygnus X-1. [extrapolation from nearby stars
NASA Technical Reports Server (NTRS)
Margon, B.; Bowyer, S.; Stone, R. P. S.
1973-01-01
Interstellar extinction of 50 stars in the immediate vicinity of Cyg X-1 is compared with color excess of HDE 226868. The fact that HDE 226868 has extinction drastically exceeding that of all stars in the field at less than 1 kpc is believed to be in conflict with the model of Trimble et al. (1973) in which the primary is a luminous undermassive star and with other similar models. Uniform extrapolation of the reddening yields a distance estimate of (2.5 plus or minus 0.4) kpc for Cygnus X-1, in agreement with the spectroscopic modulus for an O9 star.
RF-sheath heat flux estimates on Tore Supra and JET ICRF antennae. Extrapolation to ITER
Colas, L.; Portafaix, C.; Goniche, M.; Jacquet, Ph.
2009-11-26
RF-sheath induced heat loads are identified from infrared thermography measurements on Tore Supra ITER-like prototype and JET A2 antennae, and are quantified by fitting thermal calculations. Using a simple scaling law assessed experimentally, the estimated heat fluxes are then extrapolated to the ITER ICRF launcher delivering 20 MW RF power for several plasma scenarios. Parallel heat fluxes up to 6.7 MW/m{sup 2} are expected very locally on ITER antenna front face. The role of edge density on operation is stressed as a trade-off between easy RF coupling and reasonable heat loads. Sources of uncertainty on the results are identified.
Temperature-extrapolation method for Implicit Monte Carlo - Radiation hydrodynamics calculations
McClarren, R. G.; Urbatsch, T. J.
2013-07-01
We present a method for implementing temperature extrapolation in Implicit Monte Carlo solutions to radiation hydrodynamics problems. The method is based on a BDF-2 type integration to estimate a change in material temperature over a time step. We present results for radiation only problems in an infinite medium and for a 2-D Cartesian hohlraum problem. Additionally, radiation hydrodynamics simulations are presented for an RZ hohlraum problem and a related 3D problem. Our results indicate that improvements in noise and general behavior are possible. We present considerations for future investigations and implementations. (authors)
Wouters, Sebastian; Limacher, Peter A; Van Neck, Dimitri; Ayers, Paul W
2012-04-01
We have implemented the sweep algorithm for the variational optimization of SU(2) U(1) (spin and particle number) invariant matrix product states (MPS) for general spin and particle number invariant fermionic Hamiltonians. This class includes non-relativistic quantum chemical systems within the Born-Oppenheimer approximation. High-accuracy ab initio finite field results of the longitudinal static polarizabilities and second hyperpolarizabilities of one-dimensional hydrogen chains are presented. This allows to assess the performance of other quantum chemical methods. For small basis sets, MPS calculations in the saturation regime of the optical response properties can be performed. These results are extrapolated to the thermodynamic limit. PMID:22482543
Convergence and stability properties of minimal polynomial and reduced rank extrapolation algorithms
NASA Technical Reports Server (NTRS)
Sidi, A.
1983-01-01
The minimal polynomial and reduced rank extrapolation algorithms are two acceleration of convergence methods for sequences of vectors. In a recent survey these methods were tested and compared with the scalar, vector, topological epsilon algorithms, and were observed to be more efficient than the latter. It was also observed that the two methods have similar convergence properties. The convergence and stability properties of these methods are analyzed and the performance of the acceleration methods when applied to a class of vector sequences that includes those sequences obtained from systems of linear equations by using matrix iterative methods is discussed.
A study of alternative schemes for extrapolation of secular variation at observatories
Alldredge, L.R.
1976-01-01
The geomagnetic secular variation is not well known. This limits the useful life of geomagnetic models. The secular variation is usually assumed to be linear with time. It is found that attenative schemes that employ quasiperiodic variations from internal and external sources can improve the extrapolation of secular variation at high-quality observatories. Although the schemes discussed are not yet fully applicable in worldwide model making, they do suggest some basic ideas that may be developed into useful tools in future model work. ?? 1976.
MEGA16 - Computer program for analysis and extrapolation of stress-rupture data
NASA Technical Reports Server (NTRS)
Ensign, C. R.
1981-01-01
The computerized form of the minimum commitment method of interpolating and extrapolating stress versus time to failure data, MEGA16, is described. Examples are given of its many plots and tabular outputs for a typical set of data. The program assumes a specific model equation and then provides a family of predicted isothermals for any set of data with at least 12 stress-rupture results from three different temperatures spread over reasonable stress and time ranges. It is written in FORTRAN 4 using IBM plotting subroutines and its runs on an IBM 370 time sharing system.
CHARACTERISTICS OF THE H-MODE PEDESTAL AND EXTRAPOLATION TO ITER
OSBORNE,TH; CORDEY,JG; GROEBNER,RJ; HATAE,T; HUBBARD,A; HORTON,LD; KAMADA,Y; KRITZ,A; LAO,LL; LEONARD,AW; LOARTE,A; MAHDAVI,MA; MOSSESSIAN,D; ONJUN,T; OSSENPENKO,M; ROGNLIEN,TD; SAIBNE,G; SNYDER,PB; SUGIHARA,M; SHURYGIN,R; THOMSEN,K; WADE,MR; WILSON,HR; XU,XQ; YATSU,K
2002-11-01
A271 CHARACTERISTICS OF THE H-MODE PEDESTAL AND EXTRAPOLATION TO ITER. The peeling-ballooning mode model for edge stability along with a model for the H-mode transport barrier width is used as an approach to estimating the H-mode pedestal conditions in ITER. Scalings of the barrier width based on ion-orbit loss, neutral penetration, and turbulence suppression are examined and empirical scalings of the barrier width are presented. An empirical scaling for the pedestal {beta} is derived based on ideas from stability and the empirical width scaling. The impact of the stability model and other factors on ELM size is discussed.
Sommerfeld, Thomas; Ehara, Masahiro
2015-01-21
The energy of a temporary anion can be computed by adding a stabilizing potential to the molecular Hamiltonian, increasing the stabilization until the temporary state is turned into a bound state, and then further increasing the stabilization until enough bound state energies have been collected so that these can be extrapolated back to vanishing stabilization. The lifetime can be obtained from the same data, but only if the extrapolation is done through analytic continuation of the momentum as a function of the square root of a shifted stabilizing parameter. This method is known as analytic continuation of the coupling constant, and it requires—at least in principle—that the bound-state input data are computed with a short-range stabilizing potential. In the context of molecules and ab initio packages, long-range Coulomb stabilizing potentials are, however, far more convenient and have been used in the past with some success, although the error introduced by the long-rang nature of the stabilizing potential remains unknown. Here, we introduce a soft-Voronoi box potential that can serve as a short-range stabilizing potential. The difference between a Coulomb and the new stabilization is analyzed in detail for a one-dimensional model system as well as for the {sup 2}Π{sub u} resonance of CO{sub 2}{sup −}, and in both cases, the extrapolation results are compared to independently computed resonance parameters, from complex scaling for the model, and from complex absorbing potential calculations for CO{sub 2}{sup −}. It is important to emphasize that for both the model and for CO{sub 2}{sup −}, all three sets of results have, respectively, been obtained with the same electronic structure method and basis set so that the theoretical description of the continuum can be directly compared. The new soft-Voronoi-box-based extrapolation is then used to study the influence of the size of diffuse and the valence basis sets on the computed resonance parameters.
Increased identification of veterinary pharmaceutical contaminants in aquatic environments has raised concerns regarding potential adverse effects of these chemicals on non-target organisms. The purpose of this work was to develop a method for predictive species extrapolation ut...
Mueller, David S.
2013-01-01
proﬁles from the entire cross section and multiple transects to determine a mean proﬁle for the measurement. The use of an exponent derived from normalized data from the entire cross section is shown to be valid for application of the power velocity distribution law in the computation of the unmeasured discharge in a cross section. Selected statistics are combined with empirically derived criteria to automatically select the appropriate extrapolation methods. A graphical user interface (GUI) provides the user tools to visually evaluate the automatically selected extrapolation methods and manually change them, as necessary. The sensitivity of the total discharge to available extrapolation methods is presented in the GUI. Use of extrap by ﬁeld hydrographers has demonstrated that extrap is a more accurate and efﬁcient method of determining the appropriate extrapolation methods compared with tools currently (2012) provided in the ADCP manufacturers’ software.
Salcido, Robert; Rubin, Emily
2016-06-01
Auditors in Medicare overpayment or False Claims Act (FCA) cases often use statistical extrapolation to estimate a health-care provider's total liability from a small sample of audited claims. Courts treat statistical extrapolation differently depending on the context. They generally afford the government substantial discretion in using statistical extrapolation in overpayment cases. By contrast, courts typically more closely scrutinize the use of extrapolation in FCA cases involving multiple damages and civil penalties to ensure that the sample truly reflects the entire universe of claims and that the extrapolation rests on a sound methodological foundation. In recent cases, however, multiple courts have allowed the use of extrapolation in FCA cases. When auditors attempt to use statistical extrapolation, providers should closely inspect the sample and challenge the extrapolation when any reasonable argument exists that the sample does not constitute a reliable or accurate representation of all the provider's claims. PMID:26896700
Spackman, Peter R.; Karton, Amir
2015-05-15
Coupled cluster calculations with all single and double excitations (CCSD) converge exceedingly slowly with the size of the one-particle basis set. We assess the performance of a number of approaches for obtaining CCSD correlation energies close to the complete basis-set limit in conjunction with relatively small DZ and TZ basis sets. These include global and system-dependent extrapolations based on the A + B/L{sup α} two-point extrapolation formula, and the well-known additivity approach that uses an MP2-based basis-set-correction term. We show that the basis set convergence rate can change dramatically between different systems(e.g.it is slower for molecules with polar bonds and/or second-row elements). The system-dependent basis-set extrapolation scheme, in which unique basis-set extrapolation exponents for each system are obtained from lower-cost MP2 calculations, significantly accelerates the basis-set convergence relative to the global extrapolations. Nevertheless, we find that the simple MP2-based basis-set additivity scheme outperforms the extrapolation approaches. For example, the following root-mean-squared deviations are obtained for the 140 basis-set limit CCSD atomization energies in the W4-11 database: 9.1 (global extrapolation), 3.7 (system-dependent extrapolation), and 2.4 (additivity scheme) kJ mol{sup –1}. The CCSD energy in these approximations is obtained from basis sets of up to TZ quality and the latter two approaches require additional MP2 calculations with basis sets of up to QZ quality. We also assess the performance of the basis-set extrapolations and additivity schemes for a set of 20 basis-set limit CCSD atomization energies of larger molecules including amino acids, DNA/RNA bases, aromatic compounds, and platonic hydrocarbon cages. We obtain the following RMSDs for the above methods: 10.2 (global extrapolation), 5.7 (system-dependent extrapolation), and 2.9 (additivity scheme) kJ mol{sup –1}.
Risk extrapolation for chlorinated methanes as promoters vs initiators of multistage carcinogenesis
Bogen, K.T. )
1990-01-01
Cell-kinetic multistage (CKM) models account for clonal growth of intermediate, premalignant cell populations and thus distinguish somatic mutations and cell proliferation as separate processes that may influence observed rates of tumor formation. This paper illustrates the application of two versions of a two-stage CKM model for extrapolating cancer risk potentially associated with exposure to carbon tetrachloride, chloroform, and dichloromethane, three suspect human carcinogens commonly present in trace amounts in groundwater supplies used for domestic consumption. For each compound, the models were used to calculate a daily oral virtually safe dose' (VSD) to humans associated with a cancer risk of 10{sup {minus}6}, extrapolated from bioassay data on increased hepatocellular tumor incidence in B6C3F1 mice. Exposure-induced bioassay tumor responses were assumed first to be due solely to promotion' in accordance with the majority of available data on in vivo genotoxicity for these compounds. Available data were used to model dose response for induced hepatocellular proliferation in mice for each compound. Physiologically based pharmacokinetic models were used to predict the hepatotoxic effective dose as a function of parent compound administered dose in mice and in humans. Key issues and uncertainties in applying CKM models to risk assessment for cancer promoters are discussed.
Image extrapolation for photo stitching using nonlocal patch-based inpainting
NASA Astrophysics Data System (ADS)
Voronin, V. V.; Marchuk, V. I.; Sherstobitov, A. I.; Semenischev, E. A.; Agaian, S.; Egiazarian, K.
2014-05-01
Image alignment and mosaicing are usually performed on a set of overlapping images, using features in the area of overlap for seamless stitching. In many cases such images have different size and shape. So we need to crop panoramas or to use image extrapolation for them. This paper focuses on novel image inpainting method based on modified exemplar-based technique. The basic idea is to find an example (patch) from an image using local binary patterns, and replacing non-existed (`lost') data with it. We propose to use multiple criteria for a patch similarity search since often in practice existed exemplar-based methods produce unsatisfactory results. The criteria for searching the best matching uses several terms, including Euclidean metric for pixel brightness and Chi-squared histogram matching distance for local binary patterns. A combined use of textural geometric characteristics together with color information allows to get more informative description of the patches. In particular, we show how to apply this strategy for image extrapolation for photo stitching. Several examples considered in this paper show the effectiveness of the proposed approach on several test images.
An analytic formula for the extrapolated range of electrons in condensed materials
NASA Astrophysics Data System (ADS)
Tabata, Tatsuo; Andreo, Pedro; Shinoda, Kunihiko
1996-12-01
A single analytic formula for the extrapolated range rex of electrons in condensed materials of atomic numbers from 4 to 92 is given. It has the form of the product of the continuous-slowing-down approximation (CSDA) range r0 and a factor fd related to multiple scattering detours. The factor fd is expressed as a function of incident electron energy T0 and atomic number Z of medium. Values of adjustable parameters in fd have been optimized for data on the ratio {r ex}/{r 0}, in which the Monte Carlo evaluated values of Tabata et al. [Nucl. Instr. Meth. B 95 (1995) 289] (from 0.1 to 100 MeV) and experimental data collected from literature (from 1 keV to 0.1 MeV) for rex have been used together with NIST-database values of r0. For r0 in the extrapolated-range formula, accurate database values or an approximate analytic expression developed as a function of T0, Z, atomic weight A and mean excitation energy I of medium can be used. The maximum deviation of the resultant formula from the Monte Carlo data is about 2% for either option of r0. The determination of the expression for fd at energies below 0.1 MeV is tentative. By using an effective atomic number and atomic weight, the formula can also be applied to light compounds and mixtures.
Comparison of precipitation nowcasting by extrapolation and statistical-advection methods
NASA Astrophysics Data System (ADS)
Sokol, Zbynek; Kitzmiller, David; Pesice, Petr; Mejsnar, Jan
2013-04-01
Two models for nowcasting of 1-h, 2-h and 3-h precipitation in the warm part of the year were evaluated. The first model was based on the extrapolation of observed radar reflectivity (COTREC-IPA) and the second one combined the extrapolation with the application of a statistical model (SAMR). The accuracy of the model forecasts was evaluated on independent data using the standard measures of root-mean-squared-error, absolute error, bias and correlation coefficient as well as by spatial verification methods Fractions Skill Score and SAL technique. The results show that SAMR yields slightly better forecasts during the afternoon period. On the other hand very small or no improvement is realized at night and in the very early morning. COTREC-IPA and SAMR forecast a very similar horizontal structure of precipitation patterns but the model forecasts differ in values. SAMR, similarly as COTREC-IPA, is not able to develop new storms or significantly intensify already existing storms. This is caused by a large uncertainty regarding future development. On the other hand, the SAMR model can reliably predict decreases in precipitation intensity.
Downscaling and extrapolating dynamic seasonal marine forecasts for coastal ocean users
NASA Astrophysics Data System (ADS)
Vanhatalo, Jarno; Hobday, Alistair J.; Little, L. Richard; Spillman, Claire M.
2016-04-01
Marine weather and climate forecasts are essential in planning strategies and activities on a range of temporal and spatial scales. However, seasonal dynamical forecast models, that provide forecasts in monthly scale, often have low offshore resolution and limited information for inshore coastal areas. Hence, there is increasing demand for methods capable of fine scale seasonal forecasts covering coastal waters. Here, we have developed a method to combine observational data with dynamical forecasts from POAMA (Predictive Ocean Atmosphere Model for Australia; Australian Bureau of Meteorology) in order to produce seasonal downscaled, corrected forecasts, extrapolated to include inshore regions that POAMA does not cover. We demonstrate the method in forecasting the monthly sea surface temperature anomalies in the Great Australian Bight (GAB) region. The resolution of POAMA in the GAB is approximately 2° × 1° (lon. × lat.) and the resolution of our downscaled forecast is approximately 1° × 0.25°. We use data and model hindcasts for the period 1994-2010 for forecast validation. The predictive performance of our statistical downscaling model improves on the original POAMA forecast. Additionally, this statistical downscaling model extrapolates forecasts to coastal regions not covered by POAMA and its forecasts are probabilistic which allows straightforward assessment of uncertainty in downscaling and prediction. A range of marine users will benefit from access to downscaled and nearshore forecasts at seasonal timescales.
Molecular Dynamics/Order Parameter eXtrapolation (MD/OPX) for Bionanosystem Simulations
Miao, Yinglong; Ortoleva, Peter J.
2012-01-01
A multiscale approach, Molecular Dynamics/Order Parameter eXtrapolation (MD/OPX), to the all-atom simulation of large bionanosystems is presented. The approach starts with the introduction of a set of order parameters (OPs) automatically generated with orthogonal polynomials to characterize the nanoscale features of bionanosystems. The OPs are shown to evolve slowly via Newton’s equations and the all-atom multiscale analysis (AMA) developed earlier1 demonstrates the existence of their stochastic dynamics, which serve as the justification for our MD/OPX approach. In MD/OPX, a short MD run estimates the rate of change of the OPs, which is then used to extrapolate the state of the system over time that is much longer than the 10−14 second timescale of fast atomic vibrations and collisions. The approach is implemented in NAMD and demonstrated on cowpea chlorotic mottle virus (CCMV) capsid structural transitions (STs). It greatly accelerates the MD code and its underlying all-atom description of the nanosystems enables the use of a universal inter-atomic force field, avoiding recalibration with each new application as needed for coarse-grained models. PMID:18636559
Counter-extrapolation method for conjugate interfaces in computational heat and mass transfer.
Le, Guigao; Oulaid, Othmane; Zhang, Junfeng
2015-03-01
In this paper a conjugate interface method is developed by performing extrapolations along the normal direction. Compared to other existing conjugate models, our method has several technical advantages, including the simple and straightforward algorithm, accurate representation of the interface geometry, applicability to any interface-lattice relative orientation, and availability of the normal gradient. The model is validated by simulating the steady and unsteady convection-diffusion system with a flat interface and the steady diffusion system with a circular interface, and good agreement is observed when comparing the lattice Boltzmann results with respective analytical solutions. A more general system with unsteady convection-diffusion process and a curved interface, i.e., the cooling process of a hot cylinder in a cold flow, is also simulated as an example to illustrate the practical usefulness of our model, and the effects of the cylinder heat capacity and thermal diffusivity on the cooling process are examined. Results show that the cylinder with a larger heat capacity can release more heat energy into the fluid and the cylinder temperature cools down slower, while the enhanced heat conduction inside the cylinder can facilitate the cooling process of the system. Although these findings appear obvious from physical principles, the confirming results demonstrates the application potential of our method in more complex systems. In addition, the basic idea and algorithm of the counter-extrapolation procedure presented here can be readily extended to other lattice Boltzmann models and even other computational technologies for heat and mass transfer systems. PMID:25871245
On Extrapolating Past the Range of Observed Data When Making Statistical Predictions in Ecology.
Conn, Paul B; Johnson, Devin S; Boveng, Peter L
2015-01-01
Ecologists are increasingly using statistical models to predict animal abundance and occurrence in unsampled locations. The reliability of such predictions depends on a number of factors, including sample size, how far prediction locations are from the observed data, and similarity of predictive covariates in locations where data are gathered to locations where predictions are desired. In this paper, we propose extending Cook's notion of an independent variable hull (IVH), developed originally for application with linear regression models, to generalized regression models as a way to help assess the potential reliability of predictions in unsampled areas. Predictions occurring inside the generalized independent variable hull (gIVH) can be regarded as interpolations, while predictions occurring outside the gIVH can be regarded as extrapolations worthy of additional investigation or skepticism. We conduct a simulation study to demonstrate the usefulness of this metric for limiting the scope of spatial inference when conducting model-based abundance estimation from survey counts. In this case, limiting inference to the gIVH substantially reduces bias, especially when survey designs are spatially imbalanced. We also demonstrate the utility of the gIVH in diagnosing problematic extrapolations when estimating the relative abundance of ribbon seals in the Bering Sea as a function of predictive covariates. We suggest that ecologists routinely use diagnostics such as the gIVH to help gauge the reliability of predictions from statistical models (such as generalized linear, generalized additive, and spatio-temporal regression models). PMID:26496358
Persistence Length of Stable Microtubules
NASA Astrophysics Data System (ADS)
Hawkins, Taviare; Mirigian, Matthew; Yasar, M. Selcuk; Ross, Jennifer
2011-03-01
Microtubules are a vital component of the cytoskeleton. As the most rigid of the cytoskeleton filaments, they give shape and support to the cell. They are also essential for intracellular traffic by providing the roadways onto which organelles are transported, and they are required to reorganize during cellular division. To perform its function in the cell, the microtubule must be rigid yet dynamic. We are interested in how the mechanical properties of stable microtubules change over time. Some ``stable'' microtubules of the cell are recycled after days, such as in the axons of neurons or the cilia and flagella. We measured the persistence length of freely fluctuating taxol-stabilized microtubules over the span of a week and analyzed them via Fourier decomposition. As measured on a daily basis, the persistence length is independent of the contour length. Although measured over the span of the week, the accuracy of the measurement and the persistence length varies. We also studied how fluorescently-labeling the microtubule affects the persistence length and observed that a higher labeling ratio corresponded to greater flexibility. National Science Foundation Grant No: 0928540 to JLR.
Does length or neighborhood size cause the word length effect?
Jalbert, Annie; Neath, Ian; Surprenant, Aimée M
2011-10-01
Jalbert, Neath, Bireta, and Surprenant (2011) suggested that past demonstrations of the word length effect, the finding that words with fewer syllables are recalled better than words with more syllables, included a confound: The short words had more orthographic neighbors than the long words. The experiments reported here test two predictions that would follow if neighborhood size is a more important factor than word length. In Experiment 1, we found that concurrent articulation removed the effect of neighborhood size, just as it removes the effect of word length. Experiment 2 demonstrated that this pattern is also found with nonwords. For Experiment 3, we factorially manipulated length and neighborhood size, and found only effects of the latter. These results are problematic for any theory of memory that includes decay offset by rehearsal, but they are consistent with accounts that include a redintegrative stage that is susceptible to disruption by noise. The results also confirm the importance of lexical and linguistic factors on memory tasks thought to tap short-term memory. PMID:21461875
When Does Length Cause the Word Length Effect?
ERIC Educational Resources Information Center
Jalbert, Annie; Neath, Ian; Bireta, Tamra J.; Surprenant, Aimee M.
2011-01-01
The word length effect, the finding that lists of short words are better recalled than lists of long words, has been termed one of the benchmark findings that any theory of immediate memory must account for. Indeed, the effect led directly to the development of working memory and the phonological loop, and it is viewed as the best remaining…
Continuously variable focal length lens
Adams, Bernhard W; Chollet, Matthieu C
2013-12-17
A material preferably in crystal form having a low atomic number such as beryllium (Z=4) provides for the focusing of x-rays in a continuously variable manner. The material is provided with plural spaced curvilinear, optically matched slots and/or recesses through which an x-ray beam is directed. The focal length of the material may be decreased or increased by increasing or decreasing, respectively, the number of slots (or recesses) through which the x-ray beam is directed, while fine tuning of the focal length is accomplished by rotation of the material so as to change the path length of the x-ray beam through the aligned cylindrical slows. X-ray analysis of a fixed point in a solid material may be performed by scanning the energy of the x-ray beam while rotating the material to maintain the beam's focal point at a fixed point in the specimen undergoing analysis.
Effective Cavity Length of Gyrotrons
NASA Astrophysics Data System (ADS)
Thumm, Manfred
2014-12-01
Megawatt-class gyrotron oscillators for electron cyclotron heating and non-inductive current drive (ECH&CD) in magnetically confined thermonuclear fusion plasmas have relatively low cavity quality factors in the range of 1000 to 2000. The effective length of their cavities cannot be simply deduced from the cavity electric field profile, since this has by far not a Gaussian shape. The present paper presents a novel method to estimate the effective length of a gyrotron cavity just from the eigenvalue of the operating TEm,n mode, the cavity radius and the exact oscillation frequency which may be numerically computed or precisely measured. This effective cavity length then can be taken to calculate the Fresnel parameter in order to confirm that the cavity is not too short so that the transverse structure of any mode in the cavity is the same as that of the corresponding mode in a long circular waveguide with the same diameter.
Graduated compression stockings: knee length or thigh length.
Benkö, T; Cooke, E A; McNally, M A; Mollan, R A
2001-02-01
The mechanisms by which graduated compression stockings prevent deep venous thrombosis are not completely understood. In the current study the physiologic effect of low-pressure graduated compression stockings on the venous blood flow in the lower limb and the practical aspects of their use were assessed. Patients having elective orthopaedic surgery at a university orthopaedic department were randomized into five groups to wear two different types of graduated compression stockings in thigh and knee lengths. Patients in the fifth control group did not wear graduated compression stockings. Venous occlusion strain gauge plethysmography was used to measure venous flow. After 20-minutes bed rest there was a highly significant increase in venous capacitance and venous outflow in patients in all of the four groups wearing stockings. There was no difference in the mean of the percentage change of venous capacitance in patients in the four groups wearing stockings. The knee length Brevet stockings were less efficient in increasing the venous outflow. There was no significant change in the venous capacitance and venous outflow in patients in the control group. Visual assessment of the fit and use of stockings was done, and patients' subjective opinion of comfort was sought. The knee length graduated compression stockings wrinkled significantly less, and significantly fewer patients reported discomfort with them. All stockings were reported to be difficult to use. Thigh and knee length stockings have a significant effect on decreasing venous stasis of the lower limb. Knee length graduated compression stockings are similarly efficient in decreasing venous stasis, but they are more comfortable to wear, and they wrinkle less. PMID:11210954
Coherence length of neutron superfluids
De Blasio, F.V.; Hjorth-Jensen, M.; Lazzari, G.; Baldo, M.; Schulze, H.
1997-10-01
The coherence length of superfluid neutron matter is calculated from the microscopic BCS wave function of a Cooper pair in momentum space making use of recent nucleon-nucleon potential models and including polarization (RPA) effects. We find as our main result that the coherence length is proportional to the Fermi momentum to pairing gap ratio, in good agreement with simple estimates used in the literature, with a nearly interaction independent constant of proportionality. Our calculations can be applied to the problem of inhomogeneous superfluidity of hadronic matter in the crust of a neutron star. {copyright} {ital 1997} {ital The American Physical Society}
Overview of bunch length measurements.
Lumpkin, A. H.
1999-02-19
An overview of particle and photon beam bunch length measurements is presented in the context of free-electron laser (FEL) challenges. Particle-beam peak current is a critical factor in obtaining adequate FEL gain for both oscillators and self-amplified spontaneous emission (SASE) devices. Since measurement of charge is a standard measurement, the bunch length becomes the key issue for ultrashort bunches. Both time-domain and frequency-domain techniques are presented in the context of using electromagnetic radiation over eight orders of magnitude in wavelength. In addition, the measurement of microbunching in a micropulse is addressed.
CT image construction of a totally deflated lung using deformable model extrapolation
Sadeghi Naini, Ali; Pierce, Greg; Lee, Ting-Yim; and others
2011-02-15
Purpose: A novel technique is proposed to construct CT image of a totally deflated lung from a free-breathing 4D-CT image sequence acquired preoperatively. Such a constructed CT image is very useful in performing tumor ablative procedures such as lung brachytherapy. Tumor ablative procedures are frequently performed while the lung is totally deflated. Deflating the lung during such procedures renders preoperative images ineffective for targeting the tumor. Furthermore, the problem cannot be solved using intraoperative ultrasound (U.S.) images because U.S. images are very sensitive to small residual amount of air remaining in the deflated lung. One possible solution to address these issues is to register high quality preoperative CT images of the deflated lung with their corresponding low quality intraoperative U.S. images. However, given that such preoperative images correspond to an inflated lung, such CT images need to be processed to construct CT images pertaining to the lung's deflated state. Methods: To obtain the CT images of deflated lung, we present a novel image construction technique using extrapolated deformable registration to predict the deformation the lung undergoes during full deflation. The proposed construction technique involves estimating the lung's air volume in each preoperative image automatically in order to track the respiration phase of each 4D-CT image throughout a respiratory cycle; i.e., the technique does not need any external marker to form a respiratory signal in the process of curve fitting and extrapolation. The extrapolated deformation field is then applied on a preoperative reference image in order to construct the totally deflated lung's CT image. The technique was evaluated experimentally using ex vivo porcine lung. Results: The ex vivo lung experiments led to very encouraging results. In comparison with the CT image of the deflated lung we acquired for the purpose of validation, the constructed CT image was very similar. The
On shrinkage and model extrapolation in the evaluation of clinical center performance
Varewyck, Machteld; Goetghebeur, Els; Eriksson, Marie; Vansteelandt, Stijn
2014-01-01
We consider statistical methods for benchmarking clinical centers based on a dichotomous outcome indicator. Borrowing ideas from the causal inference literature, we aim to reveal how the entire study population would have fared under the current care level of each center. To this end, we evaluate direct standardization based on fixed versus random center effects outcome models that incorporate patient-specific baseline covariates to adjust for differential case-mix. We explore fixed effects (FE) regression with Firth correction and normal mixed effects (ME) regression to maintain convergence in the presence of very small centers. Moreover, we study doubly robust FE regression to avoid outcome model extrapolation. Simulation studies show that shrinkage following standard ME modeling can result in substantial power loss relative to the considered alternatives, especially for small centers. Results are consistent with findings in the analysis of 30-day mortality risk following acute stroke across 90 centers in the Swedish Stroke Register. PMID:24812420
Vector Extrapolation-Based Acceleration of Regularized Richardson Lucy Image Deblurring
NASA Astrophysics Data System (ADS)
Remmele, Steffen; Hesser, Jürgen
Confocal fluorescence microscopy has become an important tool in biological and medical sciences for imaging thin specimen, even living ones. Due to out-of-focus blurring and noise the acquired images are degraded and thus it is necessary to restore them. One of the most popular methods is an iterative Richardson-Lucy algorithm with total variation regularization. This algorithm while improving the image quality is converging slowly whereas with a constantly increasing amount of image data fast methods are required. In this paper, we present an accelerated version of the algorithm and investigate the achieved speed up. The acceleration method is based on a vector extrapolation technique and avoids a computational intensive evaluation of the underlying cost function. To evaluate the acceleration two synthetic test images are used. The accelerated algorithm reaches an acceptable result within 30% to 40% less computational time.
Verloock, Leen; Joseph, Wout; Gati, Azeddine; Varsier, Nadège; Flach, Björn; Wiart, Joe; Martens, Luc
2013-06-01
An experimental validation of a low-cost method for extrapolation and estimation of the maximal electromagnetic-field exposure from long-term evolution (LTE) radio base station installations are presented. No knowledge on downlink band occupation or service characteristics is required for the low-cost method. The method is applicable in situ. It only requires a basic spectrum analyser with appropriate field probes without the need of expensive dedicated LTE decoders. The method is validated both in laboratory and in situ, for a single-input single-output antenna LTE system and a 2×2 multiple-input multiple-output system, with low deviations in comparison with signals measured using dedicated LTE decoders. PMID:23179190
Age of Eocene/Oligocene boundary based on extrapolation from North American microtektite layer
Glass, B.P.; Crosbie, J.R.
1982-04-01
Microtektites believed to belong to the North American tektite strewn field have been found in upper Eocene sediments in cores from nine Deep Sea Drilling Project sites in the Caribbean Sea, Gulf of Mexico, equatorial Pacific, and eastern equatorial Indian Ocean. The microtektite layer has an age of 34.2 +- 0.6 m.y. based on fission-track dating of the microtektites and K-Ar and fission-track dating of the North American tektites. Extrapolation from the microtektite layer to the overlying Eocene/Oligocene boundary indicates an age of 32.3 +- 0.9 m.y. for the Eocene/Oligocene boundary as defined at each site in the Initial Reports of the Deep Sea Drilling Project. This age is approximately 5 m.y. younger than the age of 37.5 m.y. that is generally assigned to the boundary based on recently published Cenozoic time scales. 3 figures, 5 tables.
NASA Astrophysics Data System (ADS)
Fernandes, Ryan I.; Fairweather, Graeme
2012-08-01
An alternating direction implicit (ADI) orthogonal spline collocation (OSC) method is described for the approximate solution of a class of nonlinear reaction-diffusion systems. Its efficacy is demonstrated on the solution of well-known examples of such systems, specifically the Brusselator, Gray-Scott, Gierer-Meinhardt and Schnakenberg models, and comparisons are made with other numerical techniques considered in the literature. The new ADI method is based on an extrapolated Crank-Nicolson OSC method and is algebraically linear. It is efficient, requiring at each time level only O(N) operations where N is the number of unknowns. Moreover, it is shown to produce approximations which are of optimal global accuracy in various norms, and to possess superconvergence properties.
Influence of experimental scatter on the analysis and extrapolation of stress-rupture data
Booker, M.K.
1986-01-01
A great deal of research has been performed in the area of analysis and extrapolation of stress rupture data for more than three decades. Most of this work has concentrated on development of clever model forms for describing behavior in a simple mathematical way. However, little direct consideration has been given to the importance of inherent data scatter for such analyses. This paper presents a direct discussion of that topic, including an illustration of the impact of data scatter for several computer simulated data sets for which ''true'' behavior and the ''true'' amount of scatter are known by definition. The results clearly indicate that less attention should be paid to development of clever model forms for fitting data, and more should be paid to the use of sound, valid statistical techniques for the fitting of whatever models are used.
The use of extrapolation concepts to augment the Frequency Separation Technique
NASA Astrophysics Data System (ADS)
Alexiou, Spiros
2015-03-01
The Frequency Separation Technique (FST) is a general method formulated to improve the speed and/or accuracy of lineshape calculations, including strong overlapping collisions, as is the case for ion dynamics. It should be most useful when combined with ultrafast methods, that, however have significant difficulties when the impact regime is approached. These difficulties are addressed by the Frequency Separation Technique, in which the impact limit is correctly recovered. The present work examines the possibility of combining the Frequency Separation Technique with the addition of extrapolation to improve results and minimize errors resulting from the neglect of fast-slow coupling and thus obtain the exact result with a minimum of extra effort. To this end the adequacy of one such ultrafast method, the Frequency Fluctuation Method (FFM) for treating the nonimpact part is examined. It is found that although the FFM is unable to reproduce the nonimpact profile correctly, its coupling with the FST correctly reproduces the total profile.
NASA Astrophysics Data System (ADS)
Florez, W. F.; Portapila, M.; Hill, A. F.; Power, H.; Orsini, P.; Bustamante, C. A.
2015-03-01
The aim of this paper is to present how to implement a control volume approach improved by Hermite radial basis functions (CV-RBF) for geochemical problems. A multi-step strategy based on Richardson extrapolation is proposed as an alternative to the conventional dual step sequential non-iterative approach (SNIA) for coupling the transport equations with the chemical model. Additionally, this paper illustrates how to use PHREEQC to add geochemical reaction capabilities to CV-RBF transport methods. Several problems with different degrees of complexity were solved including cases of cation exchange, dissolution, dissociation, equilibrium and kinetics at different rates for mineral species. The results show that the solution and strategies presented here are effective and in good agreement with other methods presented in the literature for the same cases.
Dowding, Kevin J.; Hills, Richard Guy
2005-04-01
Numerical models of complex phenomena often contain approximations due to our inability to fully model the underlying physics, the excessive computational resources required to fully resolve the physics, the need to calibrate constitutive models, or in some cases, our ability to only bound behavior. Here we illustrate the relationship between approximation, calibration, extrapolation, and model validation through a series of examples that use the linear transient convective/dispersion equation to represent the nonlinear behavior of Burgers equation. While the use of these models represents a simplification relative to the types of systems we normally address in engineering and science, the present examples do support the tutorial nature of this document without obscuring the basic issues presented with unnecessarily complex models.
A two-grid method with Richardson extrapolation for a semilinear convection-diffusion problem
NASA Astrophysics Data System (ADS)
Tikhovskaya, S. V.; Zadorin, A. I.
2015-10-01
A boundary value problem for a second-order semilinear singularly perturbed ordinary differential equation is considered. We use Newton and Picard iterations for a linearization. To solve the problem at each iteration we apply the difference scheme with the property of uniform with respect to the singular perturbation parameter convergence. A modified Samarskii and central difference schemes on Shishkin mesh are considered. It is known that these schemes are almost second order accuracy uniformly with respect to the singular perturbation parameter. To decrease the required number of arithmetical operations for resolving the difference scheme, a two-grid method is proposed. To increase the accuracy of difference scheme, we investigate the possibility to apply Richardson extrapolation using known solutions of the difference scheme on both meshes. The comparison of modified Samarskii and central difference schemes is carried out. The results of some numerical experiments are discussed.
Extrapolation of Group Proximity from Member Relations Using Embedding and Distribution Mapping
NASA Astrophysics Data System (ADS)
Misawa, Hideaki; Horio, Keiichi; Morotomi, Nobuo; Fukuda, Kazumasa; Taniguchi, Hatsumi
In the present paper, we address the problem of extrapolating group proximities from member relations, which we refer to as the group proximity problem. We assume that a relational dataset consists of several groups and that pairwise relations of all members can be measured. Under these assumptions, the goal is to estimate group proximities from pairwise relations. In order to solve the group proximity problem, we present a method based on embedding and distribution mapping, in which all relational data, which consist of pairwise dissimilarities or dissimilarities between members, are transformed into vectorial data by embedding methods. After this process, the distributions of the groups are obtained. Group proximities are estimated as distances between distributions by distribution mapping methods, which generate a map of distributions. As an example, we apply the proposed method to document and bacterial flora datasets. Finally, we confirm the feasibility of using the proposed method to solve the group proximity problem.
Ordering of metal-ion toxicities in different species--extrapolation to man
England, M.W.; Turner, J.E.; Hingerty, B.E.; Jacobson, K.B. )
1989-01-01
Our previous attempts to predict the toxicities of 24 metal ions for a given species, using physicochemical parameters associated with the ions, are summarized. In our current attempt we have chosen indicators of toxicity for biological systems of increasing levels of complexity--starting with individual biological molecules and ascending to mice as representative of higher-order animals. The numerical values for these indicators have been normalized to a scale of 100 for Mg{sup 2+} (essentially nontoxic) and 0 for Cd{sup 2+} (very toxic). To give predicted toxicities to humans, extrapolations across biological species have been made for each of the metal ions considered. The predicted values are then compared with threshold limit values (TLV) from the literature. Both methods for predicting toxicities have their advantages and disadvantages, and both have limited success for metal ions. However, the second approach suggests that the TLV for Cu{sup 2+} should be lower than that currently recommended.
King, A W
1991-12-31
A general procedure for quantifying regional carbon dynamics by spatial extrapolation of local ecosystem models is presented Monte Carlo simulation to calculate the expected value of one or more local models, explicitly integrating the spatial heterogeneity of variables that influence ecosystem carbon flux and storage. These variables are described by empirically derived probability distributions that are input to the Monte Carlo process. The procedure provides large-scale regional estimates based explicitly on information and understanding acquired at smaller and more accessible scales.Results are presented from an earlier application to seasonal atmosphere-biosphere CO{sub 2} exchange for circumpolar ``subarctic`` latitudes (64{degree}N-90{degree}N). Results suggest that, under certain climatic conditions, these high northern ecosystems could collectively release 0.2 Gt of carbon per year to the atmosphere. I interpret these results with respect to questions about global biospheric sinks for atmospheric CO{sub 2} .
Harding, M. E.; Vazquez, J.; Ruscic, B.; Wilson, A. K.; Gauss, J.; Stanton, J. F.; Chemical Sciences and Engineering Division; Univ. t Mainz; The Univ. of Texas; Univ. of North Texas
2008-01-01
Effects of increased basis-set size as well as a correlated treatment of the diagonal Born-Oppenheimer approximation are studied within the context of the high-accuracy extrapolated ab initio thermochemistry (HEAT) theoretical model chemistry. It is found that the addition of these ostensible improvements does little to increase the overall accuracy of HEAT for the determination of molecular atomization energies. Fortuitous cancellation of high-level effects is shown to give the overall HEAT strategy an accuracy that is, in fact, higher than most of its individual components. In addition, the issue of core-valence electron correlation separation is explored; it is found that approximate additive treatments of the two effects have limitations that are significant in the realm of <1 kJ mol{sup -1} theoretical thermochemistry.
Risk extrapolation for chlorinated methanes as promoters vs initiators of multistage carcinogenesis.
Bogen, K T
1990-10-01
"Cell-kinetic multistage" (CKM) models account for clonal growth of intermediate, premalignant cell populations and thus distinguish somatic mutations and cell proliferation as separate processes that may influence observed rates of tumor formation. This paper illustrates the application of two versions of a two-stage CKM model (one assuming exponential and the other geometric proliferation of intermediate cells) for extrapolating cancer risk potentially associated with exposure to carbon tetrachloride, chloroform, and dichloromethane, three suspect human carcinogens commonly present in trace amounts in groundwater supplies used for domestic consumption. For each compound, the models were used to calculate a daily oral "virtually safe dose" (VSD) to humans associated with a cancer risk of 10(-6), extrapolated from bioassay data on increased hepatocellular tumor incidence in B6C3F1 mice. Exposure-induced bioassay tumor responses were assumed first to be due solely to "promotion" (enhanced proliferation of premalignant cells, here associated with cytotoxicity), in accordance with the majority of available data on in vivo genotoxicity for these compounds. Available data were used to model dose response for induced hepatocellular proliferation in mice for each compound. Physiologically based pharmacokinetic models were used to predict the hepatotoxic effect (metabolized) dose as a function of parent compound administered dose in mice and in humans. Resulting calculated VSDs are shown to be from three to five orders of magnitude greater than corresponding values obtained assuming each of the compounds is carcinogenic only through induced somatic mutations within the CKM framework. Key issues and uncertainties in applying CKM models to risk assessment for cancer promoters are discussed. PMID:2258018
Jiang Chaowei; Feng Xueshang E-mail: fengx@spaceweather.ac.cn
2012-04-20
The magnetic field in the solar corona is usually extrapolated from a photospheric vector magnetogram using a nonlinear force-free field (NLFFF) model. NLFFF extrapolation needs considerable effort to be devoted to its numerical realization. In this paper, we present a new implementation of the magnetohydrodynamics (MHD) relaxation method for NLFFF extrapolation. The magnetofrictional approach, which is introduced for speeding the relaxation of the MHD system, is realized for the first time by the spacetime conservation-element and solution-element scheme. A magnetic field splitting method is used to further improve the computational accuracy. The bottom boundary condition is prescribed by incrementally changing the transverse field to match the magnetogram, and all other artificial boundaries of the computational box are simply fixed. We examine the code using two types of NLFFF benchmark tests, the Low and Lou semi-analytic force-free solutions and a more realistic solar-like case constructed by van Ballegooijen et al. The results show that our implementation is successful and versatile for extrapolations of either the relatively simple cases or the rather complex cases that need significant rebuilding of the magnetic topology, e.g., a flux rope. We also compute a suite of metrics to quantitatively analyze the results and demonstrate that the performance of our code in extrapolation accuracy basically reaches the same level of the present best-performing code, i.e., that developed by Wiegelmann.
Pelkonen, O; Turpeinen, M
2007-01-01
Although the measurement of metabolite formation or substrate depletion in in vitro systems, from recombinant enzymes to tissue slices, is a relatively routine task, there are a number of more or less unresolved issues in the extrapolation of the enzymatic intrinsic clearance into hepatic metabolic clearance. Nominal concentrations of the drug added to the incubation system are not necessarily the concentration the transporter or the metabolizing enzyme sees. In addition, peculiarities of incubation set-ups should be assessed. Unbound drug fractions (concentrations) in the in vitro system itself should be measured or estimated for the appropriate assessment of enzymatic intrinsic clearance. In addition, blood and/or plasma concentrations to be encountered in the in vivo situation should be measured or estimated for the extrapolation. Extrapolation always means making a number of assumptions and the most important of these, such as scaling factors from recombinant enzymes, microsomes or hepatocytes to the mass unit of the liver, liver weight, blood flow, and distribution volume amongst others, and so on, should be explicitly stated and included in the extrapolation process. Despite all the above-mentioned reservations the in vitro-in vivo extrapolation of metabolic clearance seems to be a useful and mostly fairly precise tool for predicting the important pharmacokinetic processes of a drug. PMID:17968737
Long-Period Tidal Variations in the Length of Day
NASA Technical Reports Server (NTRS)
Ray, Richard D.; Erofeeva, Svetlana Y.
2014-01-01
A new model of long-period tidal variations in length of day is developed. The model comprises 80 spectral lines with periods between 18.6 years and 4.7 days, and it consistently includes effects of mantle anelasticity and dynamic ocean tides for all lines. The anelastic properties followWahr and Bergen; experimental confirmation for their results now exists at the fortnightly period, but there remains uncertainty when extrapolating to the longest periods. The ocean modeling builds on recent work with the fortnightly constituent, which suggests that oceanic tidal angular momentum can be reliably predicted at these periods without data assimilation. This is a critical property when modeling most long-period tides, for which little observational data exist. Dynamic ocean effects are quite pronounced at shortest periods as out-of-phase rotation components become nearly as large as in-phase components. The model is tested against a 20 year time series of space geodetic measurements of length of day. The current international standard model is shown to leave significant residual tidal energy, and the new model is found to mostly eliminate that energy, with especially large variance reduction for constituents Sa, Ssa, Mf, and Mt.
Welding arc length control system
NASA Technical Reports Server (NTRS)
Iceland, William F. (Inventor)
1993-01-01
The present invention is a welding arc length control system. The system includes, in its broadest aspects, a power source for providing welding current, a power amplification system, a motorized welding torch assembly connected to the power amplification system, a computer, and current pick up means. The computer is connected to the power amplification system for storing and processing arc weld current parameters and non-linear voltage-ampere characteristics. The current pick up means is connected to the power source and to the welding torch assembly for providing weld current data to the computer. Thus, the desired arc length is maintained as the welding current is varied during operation, maintaining consistent weld penetration.
Softness Correlations Across Length Scales
NASA Astrophysics Data System (ADS)
Ivancic, Robert; Shavit, Amit; Rieser, Jennifer; Schoenholz, Samuel; Cubuk, Ekin; Durian, Douglas; Liu, Andrea; Riggleman, Robert
In disordered systems, it is believed that mechanical failure begins with localized particle rearrangements. Recently, a machine learning method has been introduced to identify how likely a particle is to rearrange given its local structural environment, quantified by softness. We calculate the softness of particles in simulations of atomic Lennard-Jones mixtures, molecular Lennard-Jones oligomers, colloidal systems and granular systems. In each case, we find that the length scale characterizing spatial correlations of softness is approximately a particle diameter. These results provide a rationale for why localized rearrangements--whose size is presumably set by the scale of softness correlations--might occur in disordered systems across many length scales. Supported by DOE DE-FG02-05ER46199.
Variable focal length deformable mirror
Headley, Daniel; Ramsey, Marc; Schwarz, Jens
2007-06-12
A variable focal length deformable mirror has an inner ring and an outer ring that simply support and push axially on opposite sides of a mirror plate. The resulting variable clamping force deforms the mirror plate to provide a parabolic mirror shape. The rings are parallel planar sections of a single paraboloid and can provide an on-axis focus, if the rings are circular, or an off-axis focus, if the rings are elliptical. The focal length of the deformable mirror can be varied by changing the variable clamping force. The deformable mirror can generally be used in any application requiring the focusing or defocusing of light, including with both coherent and incoherent light sources.
Critical Length Limiting Superlow Friction
NASA Astrophysics Data System (ADS)
Ma, Ming; Benassi, Andrea; Vanossi, Andrea; Urbakh, Michael
2015-02-01
Since the demonstration of superlow friction (superlubricity) in graphite at nanoscale, one of the main challenges in the field of nano- and micromechanics was to scale this phenomenon up. A key question to be addressed is to what extent superlubricity could persist, and what mechanisms could lead to its failure. Here, using an edge-driven Frenkel-Kontorova model, we establish a connection between the critical length above which superlubricity disappears and both intrinsic material properties and experimental parameters. A striking boost in dissipated energy with chain length emerges abruptly due to a high-friction stick-slip mechanism caused by deformation of the slider leading to a local commensuration with the substrate lattice. We derived a parameter-free analytical model for the critical length that is in excellent agreement with our numerical simulations. Our results provide a new perspective on friction and nanomanipulation and can serve as a theoretical basis for designing nanodevices with superlow friction, such as carbon nanotubes.
NASA Astrophysics Data System (ADS)
Antonio, Patrícia L.; Xavier, Marcos; Caldas, Linda V. E.
2014-11-01
The Calibration Laboratory (LCI) at the Instituto de Pesquisas Energéticas e Nucleares (IPEN) is going to establish a Böhm extrapolation chamber as a primary standard system for the dosimetry and calibration of beta radiation sources and detectors. This chamber was already tested in beta radiation beams with an aluminized Mylar entrance window, and now, it was characterized with an original Hostaphan entrance window. A comparison between the results of the extrapolation chamber with the two entrance windows was performed. The results showed that this extrapolation chamber presents the same effectiveness in beta radiation fields as a primary standard system with both entrance windows, showing that any one of them may be utilized.
NASA Astrophysics Data System (ADS)
Wohlers, M.; Huguenin, R.; Weinberg, M.; Huffman, R. E.; Eastes, R. W.; Delgreco, F. P.
1989-08-01
There is continuing need for information about the earth background clutter at ultraviolet wavelengths. The methodology and the results obtained at 1304 are described. A wavelength from an analysis of th AFGL Polar Bear Experiment. The basic measurement equipment provided data of a spatial resolution of 20 km over a large portion of the earth. The instrumentation also provided sampled outputs as the footprint scanned along the measurement track. The combination of the fine scanning and large area coverage provided opportunity for a spatial power spectral analysis that in turn provided a means for extrapolation to finer-spatial scale; a companion paper discusses the physical basis for this extrapolation.
New extrapolation method for low-lying states of nuclei in the sd and the pf shells
Shen, J. J.; Zhao, Y. M.; Arima, A.; Yoshinaga, N.
2011-04-15
We study extrapolation approaches to evaluate energies of low-lying states for nuclei in the sd and pf shells, by sorting the diagonal matrix elements of the nuclear shell-model Hamiltonian. We introduce an extrapolation method with perturbation and apply our new method to predict both low-lying state energies and E2 transition rates between low-lying states. Our predicted results arrive at an accuracy of the root-mean-squared deviations {approx}40-60 keV for low-lying states of these nuclei.
Method and apparatus for determining minority carrier diffusion length in semiconductors
Goldstein, Bernard; Dresner, Joseph; Szostak, Daniel J.
1983-07-12
Method and apparatus are provided for determining the diffusion length of minority carriers in semiconductor material, particularly amorphous silicon which has a significantly small minority carrier diffusion length using the constant-magnitude surface-photovoltage (SPV) method. An unmodulated illumination provides the light excitation on the surface of the material to generate the SPV. A manually controlled or automatic servo system maintains a constant predetermined value of the SPV. A vibrating Kelvin method-type probe electrode couples the SPV to a measurement system. The operating optical wavelength of an adjustable monochromator to compensate for the wavelength dependent sensitivity of a photodetector is selected to measure the illumination intensity (photon flux) on the silicon. Measurements of the relative photon flux for a plurality of wavelengths are plotted against the reciprocal of the optical absorption coefficient of the material. A linear plot of the data points is extrapolated to zero intensity. The negative intercept value on the reciprocal optical coefficient axis of the extrapolated linear plot is the diffusion length of the minority carriers.
Hazards in determination and extrapolation of depositional rates of recent sediments
Isphording, W.C. . Dept. of Geology-Geography); Jackson, R.B. )
1992-01-01
Calculation of depositional rates for the past 250 years in estuarine sediments at sites in the Gulf of Mexico have been carried out by measuring changes that have taken place on bathymetric charts. Depositional rates during the past 50 to 100 years can similarly be estimated by this method and may be often confirmed by relatively abrupt changes at depth in the content of certain heavy metals in core samples. Analysis of bathymetric charts of Mobile Bay, Alabama, dating back to 1858, disclosed an essentially constant sedimentation rate of 3.9 mm/year. Apalachicola Bay, Florida, similarly, was found to have a rate of 5.4 mm/year. Though, in theory, these rates should provide reliable estimates of the influx of sediment into the estuaries, considerable caution must be used in attempting to extrapolate them to any depth in core samples. The passage of hurricanes in the Gulf of Mexico is a common event and can rapidly, and markedly, alter the bathymetry of an estuary. The passage of Hurricane Elena near Apalachicola Bay in 1985, for example, removed over 84 million tons of sediment from the bay and caused an average deepening of nearly 50 cm. The impact of Hurricane Frederick on Mobile Bay in 1979 was more dramatic. During the approximate 7 hour period when winds from this storm impacted the estuary, nearly 290 million tons of sediment was driven out of the bay and an average deepening of 46 cm was observed. With such weather events common in the Gulf Coast, it is not surprising that when radioactive age dating methods were used to obtain dates of approximately 7,500 years for organic remains in cores from Apalachicola Bay, that the depths at which the dated materials were obtained in the cores corresponded to depositional rates of only 0.4 mm/year, or one-tenth that obtained from historic bathymetric data. Because storm scour effects are a common occurrence in the Gulf, no attempt should be made to extrapolate bathymetric-derived rates to beyond the age of the charts.