Benchmarking of Improved DPAC Transient Deflagration Analysis Code
Laurinat, James E.; Hensel, Steve J.
2017-09-27
The deflagration pressure analysis code (DPAC) has been upgraded for use in modeling hydrogen deflagration transients. The upgraded code is benchmarked using data from vented hydrogen deflagration tests conducted at the HYDRO-SC Test Facility at the University of Pisa. DPAC originally was written to calculate peak pressures for deflagrations in radioactive waste storage tanks and process facilities at the Savannah River Site. Upgrades include the addition of a laminar flame speed correlation for hydrogen deflagrations and a mechanistic model for turbulent flame propagation, incorporation of inertial effects during venting, and inclusion of the effect of water vapor condensation on vesselmore » walls. In addition, DPAC has been coupled with chemical equilibrium with applications (CEA), a NASA combustion chemistry code. The deflagration tests are modeled as end-to-end deflagrations. As a result, the improved DPAC code successfully predicts both the peak pressures during the deflagration tests and the times at which the pressure peaks.« less
Benchmarking of Improved DPAC Transient Deflagration Analysis Code
DOE Office of Scientific and Technical Information (OSTI.GOV)
Laurinat, James E.; Hensel, Steve J.
The deflagration pressure analysis code (DPAC) has been upgraded for use in modeling hydrogen deflagration transients. The upgraded code is benchmarked using data from vented hydrogen deflagration tests conducted at the HYDRO-SC Test Facility at the University of Pisa. DPAC originally was written to calculate peak pressures for deflagrations in radioactive waste storage tanks and process facilities at the Savannah River Site. Upgrades include the addition of a laminar flame speed correlation for hydrogen deflagrations and a mechanistic model for turbulent flame propagation, incorporation of inertial effects during venting, and inclusion of the effect of water vapor condensation on vesselmore » walls. In addition, DPAC has been coupled with chemical equilibrium with applications (CEA), a NASA combustion chemistry code. The deflagration tests are modeled as end-to-end deflagrations. As a result, the improved DPAC code successfully predicts both the peak pressures during the deflagration tests and the times at which the pressure peaks.« less
2016-04-01
the Deflagration of Ammonium Perchlorate— Hydroxyl-Terminated Polybutadiene Composite Propellants by Chiung-Chu Chen and Michael McQuaid...for Modeling the Deflagration of Ammonium Perchlorate— Hydroxyl-Terminated Polybutadiene Composite Propellants by Chiung-Chu Chen and Michael...Ammonium Perchlorate—Hydroxyl-Terminated Polybutadiene Composite Propellants 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6
Modeling deflagration waves out of hot spots
NASA Astrophysics Data System (ADS)
Partom, Yehuda
2017-01-01
It is widely accepted that shock initiation and detonation of heterogeneous explosives comes about by a two-step process known as ignition and growth. In the first step a shock sweeping through an explosive cell (control volume) creates hot spots that become ignition sites. In the second step, deflagration waves (or burn waves) propagate out of those hot spots and transform the reactant in the cell into reaction products. The macroscopic (or average) reaction rate of the reactant in the cell depends on the speed of those deflagration waves and on the average distance between neighboring hot spots. Here we simulate the propagation of deflagration waves out of hot spots on the mesoscale in axial symmetry using a 2D hydrocode, to which we add heat conduction and bulk reaction. The propagation speed of the deflagration waves may depend on both pressure and temperature. It depends on pressure for quasistatic loading near ambient temperature, and on temperature at high temperatures resulting from shock loading. From the simulation we obtain deflagration fronts emanating out of the hot spots. For 8 to 13 GPa shocks, the emanating fronts propagate as deflagration waves to consume the explosive between hot spots. For higher shock levels deflagration waves may interact with the sweeping shock to become detonation waves on the mesoscale. From the simulation results we extract average deflagration wave speeds.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Menikoff, Ralph
2012-04-03
Shock initiation in a plastic-bonded explosives (PBX) is due to hot spots. Current reactive burn models are based, at least heuristically, on the ignition and growth concept. The ignition phase occurs when a small localized region of high temperature (or hot spot) burns on a fast time scale. This is followed by a growth phase in which a reactive front spreads out from the hot spot. Propagating reactive fronts are deflagration waves. A key question is the deflagration speed in a PBX compressed and heated by a shock wave that generated the hot spot. Here, the ODEs for a steadymore » deflagration wave profile in a compressible fluid are derived, along with the needed thermodynamic quantities of realistic equations of state corresponding to the reactants and products of a PBX. The properties of the wave profile equations are analyzed and an algorithm is derived for computing the deflagration speed. As an illustrative example, the algorithm is applied to compute the deflagration speed in shock compressed PBX 9501 as a function of shock pressure. The calculated deflagration speed, even at the CJ pressure, is low compared to the detonation speed. The implication of this are briefly discussed.« less
Mesoscale Modeling of Deflagration-Induced Deconsolidation in Polymer-Bonded Explosives
NASA Astrophysics Data System (ADS)
Springer, H. Keo; Reaugh, J. E.; Glascoe, E. A.; Kercher, J. R.; Friedman, G.
2011-06-01
Initially intact polymer-bonded explosives can transition from conductive burning to more violent convective burning via rapid deconsolidation at higher pressures. The pressure-dependent infiltration of cracks and pores, i.e., damage, by product gases at the burn-front is a key step in the transition to convective burning. However, the relative influence of pre-existing damage and deflagration-induced damage on the transition to convective burning is not well understood. The objective of this study is to investigate the role of explosive constituent properties, microstructure, and deflagration velocity on deconsolidation. We performed simulations using the multi-physics hydrocode, ALE3D. HMX was used as the model energetic grain. We used a JWL form for the unreacted and reacted equation-of-state of the HMX. Simplified strength and failure models were used for the HMX and the binder. The propensity for deconsolidation increased with increasing grain volume fraction, increasing porosity, decreasing binder strength, and increasing deflagration velocity. These studies are important because they enable the development of deflagration-induced damage models, as well as the design of inherently safer explosives. This work performed under the auspices of the U.S. DOE by LLNL under Contract DE-AC52-07NA27344. This work was funded by the Joint DoD/DOE Munitions Technology Development Program.
Small-scale deflagration cylinder test with velocimetry wall-motion diagnostics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hooks, Daniel E; Hill, Larry G; Pierce, Timothy H
Predicting the likelihood and effects of outcomes resultant from thermal initiation of explosives remains a significant challenge. For certain explosive formulations, the general outcome can be broadly predicted given knowledge of certain conditions. However, there remain unexplained violent events, and increased statistical understanding of outcomes as a function of many variables, or 'violence categorization,' is needed. Additionally, the development of an equation of state equivalent for deflagration would be very useful in predicting possible detailed event consequences using traditional hydrodynamic detonation moders. For violence categorization, it is desirable that testing be efficient, such that it is possible to statistically definemore » outcomes reliant on the processes of initiation of deflagration, steady state deflagration, and deflagration to detonation transitions. If the test simultaneously acquires information to inform models of violent deflagration events, overall predictive capabilities for event likelihood and consequence might improve remarkably. In this paper we describe an economical scaled deflagration cylinder test. The cyclotetramethylene tetranitramine (HMX) based explosive formu1lation PBX 9501 was tested using different temperature profiles in a thick-walled copper cylindrical confiner. This test is a scaled version of a recently demonstrated deflagration cylinder test, and is similar to several other thermal explosion tests. The primary difference is the passive velocimetry diagnostic, which enables measurement of confinement vessel wall velocities at failure, regardless of the timing and location of ignition.« less
Properties of Deflagration Fronts and Models for Type IA Supernovae
NASA Astrophysics Data System (ADS)
Domínguez, I.; Höflich, P.
2000-01-01
Detailed models of the explosion of a white dwarf that include self-consistent calculations of the light curve and spectra provide a link between observational quantities and the underlying explosion model. These calculations assume spherical geometry and are based on parameterized descriptions of the burning front. Recently, the first multidimensional calculations for nuclear burning fronts have been performed. Although a fully consistent treatment of the burning fronts is beyond the current state of the art, these calculations provide a new and better understanding of the physics. Several new descriptions for flame propagation have been proposed by Khokhlov et al. and Niemeyer et al. Using various descriptions for the propagation of a nuclear deflagration front, we have studied the influence on the results of previous analyses of Type Ia supernovae, namely, the nucleosynthesis and structure of the expanding envelope. Our calculations are based on a set of delayed detonation models with parameters that give a good account of the optical and infrared light curves and of the spectral evolution. In this scenario, the burning front first propagates in a deflagration mode and subsequently turns into a detonation. The explosions and light curves are calculated using a one-dimensional Lagrangian radiation-hydro code including a detailed nuclear network. We find that the results of the explosion are rather insensitive to details of the description of the deflagration front, even if its speed and the time from the transition to detonation differ almost by a factor of 2. For a given white dwarf (WD) and a fixed transition density, the total production of elements changes by less than 10%, and the distribution in the velocity space changes by less than 7%. Qualitatively, this insensitivity of the final outcome of the explosion to the details of the flame propagation during the (slow) deflagration phase can be understood as follows: for plausible variations in the speed of the turbulent deflagration, the duration of this phase is several times longer than the sound crossing time in the initial WD. Therefore, the energy produced during the early nuclear burning can be redistributed over the entire WD, causing a slow preexpansion. In this intermediate state, the WD is still bound but its binding energy is reduced by the amount of nuclear energy. The expansion ratio depends mainly on the total amount of burning during the deflagration phase. Consequently, the conditions are very similar under which nuclear burning takes place during the subsequent detonation phase. In our example, the density and temperature at the burning front changes by less than 3%, and the expansion velocity changes by less than 10%. The burning conditions are very close to previous calculations which used a constant deflagration velocity. Based on a comparison with observations, those required low deflagration speeds (~2%-3% of the speed of sound). Exceptions to the similarity are the innermost layers of ~0.03-0.05 Msolar. Still, nuclear burning is in nuclear statistical equilibrium, but the rate of electron capture is larger for the new descriptions of the flame propagation. Consequently, the production of very neutron-rich isotopes is increased. In our example, close to the center Ye is about 0.44, compared to 0.46 in the model with constant deflagration speed. This increases the 48Ca production by more than a factor of 100 to 3.E-6 Msolar. Conclusions from previous analyses of light curves and spectra on the properties of the WD and the explosions will not change, and even with the new descriptions, the delayed detonation scenario is consistent with the observations. Namely, the central density results with respect to the chemical structure of the progenitor and the transition density from deflagration to detonation do not change. The reason for this similarity is the fact that the total amount of burning during the long deflagration phase determines the restructuring of the WD prior to the detonation. Therefore, we do not expect that the precise, microphysical prescription for the speed of a subsonic burning front has a significant effect on the outcome. However, at the current level of uncertainties for the burning front, the relation between properties of the burning front and of the initial white dwarf cannot be obtained from a comparison between observation and theoretical predictions by one-dimensional models. Multidimensional calculations are needed (1) to get inside the relations between model parameters such as central density and properties of the deflagration front and its relation to the transition density between deflagration and detonation and (2) to make use of information on asphericity that is provided by polarization measurements. These questions are essential to test, estimate, and predict some of the evolutionary effects of SNe Ia and their use as cosmological yardsticks.
Nucleosynthesis of Iron-Peak Elements in Type-Ia Supernovae
NASA Astrophysics Data System (ADS)
Leung, Shing-Chi; Nomoto, Ken'ichi
The observed features of typical Type Ia supernovae are well-modeled as the explosions of carbon-oxygen white dwarfs both near Chandrasekhar mass and sub-Chandrasekhar mass. However, observations in the last decade have shown that Type Ia supernovae exhibit a wide diversity, which implies models for wider range of parameters are necessary. Based on the hydrodynamics code we developed, we carry out a parameter study of Chandrasekhar mass models for Type Ia supernovae. We conduct a series of two-dimensional hydrodynamics simulations of the explosion phase using the turbulent flame model with the deflagration-detonation-transition (DDT). To reconstruct the nucleosynthesis history, we use the particle tracer scheme. We examine the role of model parameters by examining their influences on the final product of nucleosynthesis. The parameters include the initial density, metallicity, initial flame structure, detonation criteria and so on. We show that the observed chemical evolution of galaxies can help constrain these model parameters.
The delayed-detonation model of a type Ia supernovae. 1: The deflagration phase
NASA Technical Reports Server (NTRS)
Arnett, David; Livne, Eli
1994-01-01
The nature of the 'delayed detonation' mechanism of Khokhlov for the explosion of Type Ia supernovae is investigated by using two-dimensional numerical hydrodynamics simulations. A new algorithm is used to treat the deflagration front. Assuming that it propagates locally at the laminar flame speed, the deflagration is insufficient to unbind the star. Expansion shuts of the flame; much of this small production of iron group nuclei occurs at lower densities, which reduces the electron-capture problem. The burning front does become wrinkled, but the wavelength of the instability is much larger than the computational grid size and is resolved; this is consistent with previous analysis. Because the degenerate star has an adiabatic exponent only slightly above 4/3, the energy released by deflagration drives a pulsation of large amplitude. During the first expansion phase, adiabatic cooling shuts off the burning, and a Rayleigh-Taylor instability then gives mixing of high-entropy ashes with low-entropy fuel. During the first contraction phase, compressional heating reignites the material. This paper deals with the deflagration phase, from the onset of burning, through expansion and quenching of the flame, to the first contraction.
Wood, Mitchell A.; Cherukara, Mathew J.; Kober, Edward M.; ...
2015-06-13
We use molecular dynamics simulations to describe the chemical reactions following shock-induced collapse of cylindrical pores in the high-energy density material RDX. For shocks with particle velocities of 2 km/s we find that the collapse of a 40 nm diameter pore leads to a deflagration wave. Molecular collisions during the collapse lead to ultrafast, multistep chemical reactions that occur under nonequilibrium conditions. WE found that exothermic products formed during these first few picoseconds prevent the nanoscale hotspot from quenching. Within 30 ps, a local deflagration wave develops. It propagates at 0.25 km/s and consists of an ultrathin reaction zone ofmore » only ~5 nm, thus involving large temperature and composition gradients. Contrary to the assumptions in current models, a static thermal hotspot matching the dynamical one in size and thermodynamic conditions fails to produce a deflagration wave indicating the importance of nonequilibrium loading in the criticality of nanoscale hot spots. These results provide insight into the initiation of reactive decomposition.« less
Detonation Initiation of Heterogeneous Melt-Cast High Explosives
NASA Astrophysics Data System (ADS)
Chuzeville, Vincent; Baudin, Gerard; Lefrancois, Alexandre; Boulanger, Remi; Catoire, Laurent
2015-06-01
The melt-cast explosives' shock initiation mechanisms are less investigated than pressed and cast-cured ones. If the existence of hot-spots is widely recognized, their formation mechanism is not yet established. We study here two melt-cast explosives, NTO-TNT 60:40 and RDX-TNT 60:40 in order to establish a relation between the microstructure and the reaction rate using a two-phase model based on a ZND approach. Such a model requires the reaction rate, the equations of state of the unreacted phase and of the detonation products and an interaction model between the two phases to describe the reaction zone thermodynamics. The reaction rate law can be written in a factorized form including the number of initiation sites, the explosive's deflagration velocity around hot spots and a function depending on gas volume fraction produced by the deflagration front propagation. The deflagration velocity mainly depends on pressure and is determined from pop-plot tests using the hypothesis of the single curve build-up. This hypothesis has been verified for our two melt-cast explosives. The function depending on gas volume fraction is deduced from microstructural observations and from an analogy with the solid nucleation and growth theory. It has been established for deflagration fronts growing from grain's surface and a given initial grain size distribution. The model requires only a few parameters, calibrated thanks to an inversion method. A good agreement is obtained between experiments and numerical simulations.
Non-equilibrium dynamics due to moving deflagration front at RDX/HTPB interface
NASA Astrophysics Data System (ADS)
Chaudhuri, Santanu; Joshi, Kaushik; Lacevic, Naida
Reactive dissipative particle dynamics (DPD-RX), a promising tool in characterizing the sensitivity and performance of heterogeneous solid propellants like polymer bonded explosives (PSXs), requires further testing for non-equilibrium dynamics. It is important to understand detailed atomistic chemistry for developing coarse grain reactive models needed for the DPD-RX. In order to obtain insights into combustion chemistry of RDX/HTPB binder, we used reactive molecular dynamics (RMD) to obtain energy up-pumping and reaction mechanisms at RDX/HTPB interface when exposed to a self-sustaining deflagration front. Hot spots are ignited near and away from the heterogeneous interface using the thermal pulse. The results show that the hot spot near interface significantly delays the transition from ignition to deflagration. We will present the mechanical response and the combustion chemistry of HTPB when the propagating deflagration front hits the polymer binder. We will discuss our efforts to incorporate this RMD based chemistry into the DPD-RX which will enable us to perform such non-equilibrium dynamics simulations on large-length scale with microstructural heterogeneities. Funding from DTRA Grant Number HDTRA1-15-1-0034 is acknowledged.
NASA Astrophysics Data System (ADS)
Dave, Pranav; Kashyap, Rahul; Fisher, Robert; Timmes, Frank; Townsley, Dean; Byrohl, Chris
2017-05-01
Recent Suzaku X-ray spectra of supernova remnant (SNR) 3C 397 indicate enhanced stable iron group element abundances of Ni, Mn, Cr, and Fe. Seeking to address key questions about the progenitor and explosion mechanism of 3C 397, we compute nucleosynthetic yields from a suite of multidimensional hydrodynamics models in the near-Chandrasekhar-mass, single-degenerate paradigm for Type Ia supernovae (SNe Ia). Varying the progenitor white dwarf (WD) internal structure, composition, ignition, and explosion mechanism, we find that the best match to the observed iron peak elements of 3C 397 are dense (central density ≥6 × 109 g cm-3), low-carbon WDs that undergo a weak, centrally ignited deflagration, followed by a subsequent detonation. The amount of 56Ni produced is consistent with a normal or bright normal SNe Ia. A pure deflagration of a centrally ignited, low central density (≃2 × 109 g cm-3) progenitor WD, frequently considered in the literature, is also found to produce good agreement with 3C 397 nucleosynthetic yields, but leads to a subluminous SN Ia event, in conflict with X-ray line width data. Additionally, in contrast to prior work that suggested a large supersolar metallicity for the WD progenitor for SNR 3C 397, we find satisfactory agreement for solar- and subsolar-metallicity progenitors. We discuss a range of implications our results have for the single-degenerate channel.
Early Emission from Type Ia Supernovae
NASA Astrophysics Data System (ADS)
Rabinak, Itay; Livne, Eli; Waxman, Eli
2012-09-01
A unique feature of deflagration-to-detonation (DDT) white dwarf explosion models of supernovae of type Ia is the presence of a strong shock wave propagating through the outer envelope. We consider the early emission expected in such models, which is produced by the expanding shock-heated outer part of the ejecta and precedes the emission driven by radioactive decay. We expand on earlier analyses by considering the modification of the pre-detonation density profile by the weak shocks generated during the deflagration phase, the time evolution of the opacity, and the deviation of the post-shock equation of state from that obtained for radiation pressure domination. A simple analytic model is presented and shown to provide an acceptable approximation to the results of one-dimensional numerical DDT simulations. Our analysis predicts a ~103 s long UV/optical flash with a luminosity of ~1 to ~3 × 1039 erg s-1. Lower luminosity corresponds to faster (turbulent) deflagration velocity. The luminosity of the UV flash is predicted to be strongly suppressed at t > t drop ~ 1 hr due to the deviation from pure radiation domination.
Performance Impact of Deflagration to Detonation Transition Enhancing Obstacles
NASA Technical Reports Server (NTRS)
Paxson, Daniel E.; Schauer, Frederick; Hopper, David
2012-01-01
A sub-model is developed to account for the drag and heat transfer enhancement resulting from deflagration-to-detonation (DDT) inducing obstacles commonly used in pulse detonation engines (PDE). The sub-model is incorporated as a source term in a time-accurate, quasi-onedimensional, CFD-based PDE simulation. The simulation and sub-model are then validated through comparison with a particular experiment in which limited DDT obstacle parameters were varied. The simulation is then used to examine the relative contributions from drag and heat transfer to the reduced thrust which is observed. It is found that heat transfer is far more significant than aerodynamic drag in this particular experiment.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Townsley, Dean M.; Miles, Broxton J.; Timmes, F. X.
2016-07-01
We refine our previously introduced parameterized model for explosive carbon–oxygen fusion during thermonuclear Type Ia supernovae (SNe Ia) by adding corrections to post-processing of recorded Lagrangian fluid-element histories to obtain more accurate isotopic yields. Deflagration and detonation products are verified for propagation in a medium of uniform density. A new method is introduced for reconstructing the temperature–density history within the artificially thick model deflagration front. We obtain better than 5% consistency between the electron capture computed by the burning model and yields from post-processing. For detonations, we compare to a benchmark calculation of the structure of driven steady-state planar detonationsmore » performed with a large nuclear reaction network and error-controlled integration. We verify that, for steady-state planar detonations down to a density of 5 × 10{sup 6} g cm{sup −3}, our post-processing matches the major abundances in the benchmark solution typically to better than 10% for times greater than 0.01 s after the passage of the shock front. As a test case to demonstrate the method, presented here with post-processing for the first time, we perform a two-dimensional simulation of a SN Ia in the scenario of a Chandrasekhar-mass deflagration–detonation transition (DDT). We find that reconstruction of deflagration tracks leads to slightly more complete silicon burning than without reconstruction. The resulting abundance structure of the ejecta is consistent with inferences from spectroscopic studies of observed SNe Ia. We confirm the absence of a central region of stable Fe-group material for the multi-dimensional DDT scenario. Detailed isotopic yields are tabulated and change only modestly when using deflagration reconstruction.« less
Self-ignition of S.I. engine model fuels: A shock tube investigation at high pressure
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fieweger, K.; Blumenthal, R.; Adomeit, G.
1997-06-01
The self-ignition of several spark-ignition (SI) engine fuels (iso-octane, methanol, methyl tert-butyl ether and three different mixtures of iso-octane and n-heptane), mixed with air, was investigated experimentally under relevant engine conditions by the shock tube technique. Typical modes of the self-ignition process were registered cinematographically. For temperatures relevant to piston engine combustion, the self-ignition process always starts as an inhomogeneous, deflagrative mild ignition. This instant is defined by the ignition delay time, {tau}{sub defl}. The deflagration process in most cases is followed by a secondary explosion (DDT). This transition defines a second ignition delay time, {tau}{sub DDT}, which is amore » suitable approximation for the chemical ignition delay time, if the change of the thermodynamic conditions of the unburned test gas due to deflagration is taken into account. For iso-octane at p = 40 bar, a NTC (negative temperature coefficient), behavior connected with a two step (cool flame) self-ignition at low temperatures was observed. This process was very pronounced for rich and less pronounced for stoichiometric mixtures. The results of the {tau}{sub DDT} delays of the stoichiometric mixtures were shortened by the primary deflagration process in the temperature range between 800 and 1,000 K. Various mixtures of iso-octane and n-heptane were investigated. The results show a strong influence of the n-heptane fraction in the mixture, both on the ignition delay time and on the mode of self-ignition. The self-ignition of methanol and MTBE (methyl tert-butyl ether) is characterized by a very pronounced initial deflagration. For temperatures below 900 K (methanol: 800 K), no secondary explosion occurs. Taking into account the pressure increase due to deflagration, the measured delays {tau}{sub DDT} of the secondary explosion are shortened by up to one order of magnitude.« less
Combustion of Han-Based Monopropellant Droplets in Reduced Gravity
NASA Technical Reports Server (NTRS)
Shaw, B. D.
1999-01-01
The objective of this research is to study combustion of monopropellant droplets and monopropellant droplet components in reduced-gravity environments so that spherical symmetry is strongly promoted. The experiments will use hydroxylammonium nitrate (HAN, chemical formula NH3OHNO3) based monopropellants. This class of monopropellant is selected for study because of its current relevance and also because it is relatively benign and safe to work with. The experimental studies will allow for accurate determination of fundamental data on deflagration rates, gas-phase temperature profiles, transient gas-phase flame behaviors, the onset of bubbling in droplets at lower pressures, and the low-pressure deflagration limit. The theoretical studies will provide rational models of deflagration mechanisms of HAN-based liquid propellants. Besides advancing fundamental knowledge, the proposed research should aid in applications (e.g., spacecraft thrusters and liquid propellant guns) of this unique class of monopropellants.
Radial magnetic compression in the expelled jet of a plasma deflagration accelerator
DOE Office of Scientific and Technical Information (OSTI.GOV)
Loebner, Keith T. K., E-mail: kloebner@stanford.edu; Underwood, Thomas C.; Mouratidis, Theodore
2016-02-29
A spectroscopic study of a pulsed plasma deflagration accelerator is carried out that confirms the existence of a strong compression in the emerging jet at the exit plane of the device. An imaging spectrometer is used to collect broadened Hα emission from a transaxial slice of the emerging jet at high spatial resolution, and the radial plasma density profile is computed from Voigt fits of the Abel inverted emissivity profiles. The plasma temperature, determined via Doppler broadening of impurity line emission, is compared against the temperature predictions of a radial magnetohydrodynamic equilibrium model applied to the measured density profiles. Empiricalmore » scaling laws developed for the plasma density, combined with the measured and predicted temperatures, indicate that a radially equilibrated Z-pinch is formed within the expelled plasma jet at the exit plane during the deflagration process.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dave, Pranav; Kashyap, Rahul; Fisher, Robert
Recent Suzaku X-ray spectra of supernova remnant (SNR) 3C 397 indicate enhanced stable iron group element abundances of Ni, Mn, Cr, and Fe. Seeking to address key questions about the progenitor and explosion mechanism of 3C 397, we compute nucleosynthetic yields from a suite of multidimensional hydrodynamics models in the near-Chandrasekhar-mass, single-degenerate paradigm for Type Ia supernovae (SNe Ia). Varying the progenitor white dwarf (WD) internal structure, composition, ignition, and explosion mechanism, we find that the best match to the observed iron peak elements of 3C 397 are dense (central density ≥6 × 10{sup 9} g cm{sup −3}), low-carbon WDsmore » that undergo a weak, centrally ignited deflagration, followed by a subsequent detonation. The amount of {sup 56}Ni produced is consistent with a normal or bright normal SNe Ia. A pure deflagration of a centrally ignited, low central density (≃2 × 10{sup 9} g cm{sup −3}) progenitor WD, frequently considered in the literature, is also found to produce good agreement with 3C 397 nucleosynthetic yields, but leads to a subluminous SN Ia event, in conflict with X-ray line width data. Additionally, in contrast to prior work that suggested a large supersolar metallicity for the WD progenitor for SNR 3C 397, we find satisfactory agreement for solar- and subsolar-metallicity progenitors. We discuss a range of implications our results have for the single-degenerate channel.« less
Determination of the combustion behavior for pure components and mixtures using a 20-liter sphere
NASA Astrophysics Data System (ADS)
Mashuga, Chad Victor
1999-11-01
The safest method to prevent fires and explosions of flammable vapors is to prevent the existence of flammable mixtures in the first place. This methodology requires detailed knowledge of the flammability region as a function of the fuel, oxygen, and nitrogen concentrations. A triangular flammability diagram is the most useful tool to display the flammability region, and to determine if a flammable mixture is present during plant operations. An automated apparatus for assessing the flammability region and for determining the potential effect of confined fuel-air explosions is described. Data derived from the apparatus included the limits of combustion, maximum combustion pressure, and the deflagration index, or KG. Accurate measurement of these parameters can be influenced by numerous experimental conditions, including igniter energy, humidity and gas composition. Gas humidity had a substantial effect on the deflagration index, but had little effect on the maximum combustion pressure. Small changes in gas compositions had a greater effect on the deflagration index than the maximum combustion pressure. Both the deflagration indices and the maximum combustion pressure proved insensitive to the range of igniter energies examined. Estimation of flammability limits using a calculated adiabatic flame temperature (CAFT) method is demonstrated. The CAFT model is compared with the extensive experimental data from this work for methane, ethylene and a 50/50 mixture of methane and ethylene. The CAFT model compares well to methane and ethylene throughout the flammability zone when using a 1200K threshold temperature. Deviations between the method and the experimental data occurs in the fuel rich region. For the 50/50 fuel mixture the CAFT deviates only in the fuel rich region---the inclusion of carbonaceous soot as one of the equilibrium products improved the fit. Determination of burning velocities from a spherical flame model utilizing the extensive pressure---time data was also completed. The burning velocities determined compare well to other investigators using this method. The data collected for the methane/ethylene mixture was used to evaluate mixing rules for the flammability limits, maximum combustion pressure, deflagration index, and burning velocity. These rules attempt to predict the behavior of fuel mixtures from pure component data. Le Chatelier's law and averaging both work well for predicting the flammability boundary in the fuel lean region and for mixtures of inerted fuel and air. Both methods underestimate the flammability boundary in the fuel rich region. For a mixture of methane and ethylene, we were unable to identify mixing rules for estimating the maximum combustion pressure and the burning velocity from pure component data. Averaging the deflagration indices for fuel air mixtures did provide a adequate estimation of the mixture behavior. Le Chatelier's method overestimated the maximum deflagration index in air but provided a satisfactory estimation in the extreme fuel lean and rich regions.
Reactive Behavior of Explosive Billets in Deflagration Tube of Varied Confinements
NASA Astrophysics Data System (ADS)
Hu, Haibo; Guo, Yingwen; Li, Tao; Fu, Hua; Shang, Hailin; Wen, Shanggang; Qiu, Tian; LaboratoryShock Wave; Detonation Physics Research Team
2017-06-01
The deflagration process of small size cylinder billets of pressed HMX-based explosive JO-9159 and the deflagration tube wall deformation is recorded by combined pressure velocity-meter high-speed frame photographic and radiographic diagnostic system. The influence of confinement structure strength on deflagration evolution behavior is compared with analysis of convective flame propagation along the slot between explosive billet and confinement wall.The follow-up reaction inside the cracks on the initiation site end surface on the side surfaces and between the end surfaces of explosive billets is restored with the analysis results of post experimental explosive billet remains.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Maienschein, J L; Wardell, J F; Weese, R K
The violence of thermal explosions with energetic materials is affected by many material properties, including mechanical and thermal properties, thermal ignition kinetics, and deflagration behavior. These properties must be characterized for heated samples as well as pristine materials. We present available data for these properties for two HMX-based formulations--LX-04 and PBX-9501, and two RDX-based formulations--Composition B and PBXN-109. We draw upon separately published data on the thermal explosion violence with these materials to compare the material properties with the observed violence. We have the most extensive data on deflagration behavior of these four formulations, and we discuss the correlation ofmore » the deflagration data with the violence results. The data reported here may also be used to develop models for application in simulation codes such as ALE3D to calculate and Dredict thermal explosion violence.« less
Yang, Kun; Wu, Yanqing; Huang, Fenglei; Li, Ming
2017-09-05
An effective computational model is required to accurately predict the dynamic responses in accidental initiations of explosives. The present work uses a series of two-dimensional mechanical-chemical simulations performed via a hydrodynamic-code, DREXH-2D, to efficiently describe the mechanical and ignition-deflagration responses of cased cylindrical polymer-bonded explosives (PBXs) undergoing a low-to-medium-level impact (70-350m/s) in longitudinal direction. The ignition response was predicted based on an ignition criterion of effective plastic work. Slow burning and its growth to deflagration were described through a pressure-dependent reaction rate equation. The extreme value of effective plastic work was found to be useful to determine the ignition threshold velocity for PBXs. For low-level velocity impact, the incident stress wave reflection from lateral surfaces contributed to the formation of ignition regions. After the ignition, the deflagration was induced in the medium-level impact, and its violence was related to the shock strength. However, the low-strength stress wave only induced reaction at local regions, and sequent burning was no longer sensitive to the strength of incident wave. The predicted pressure and temperature results of PBXs were consistent with the medium-level impact tests performed by China Academy of Engineering Physics. Copyright © 2017 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Subramaniam, Vivek; Underwood, Thomas C.; Raja, Laxminarayan L.; Cappelli, Mark A.
2018-02-01
We present a magnetohydrodynamic (MHD) numerical simulation to study the physical mechanisms underlying plasma acceleration in a coaxial plasma gun. Coaxial plasma accelerators are known to exhibit two distinct modes of operation depending on the delay between gas loading and capacitor discharging. Shorter delays lead to a high velocity plasma deflagration jet and longer delays produce detonation shocks. During a single operational cycle that typically consists of two discharge events, the plasma acceleration exhibits a behavior characterized by a mode transition from deflagration to detonation. The first of the discharge events, a deflagration that occurs when the discharge expands into an initially evacuated domain, requires a modification of the standard MHD algorithm to account for rarefied regions of the simulation domain. The conventional approach of using a low background density gas to mimic the vacuum background results in the formation of an artificial shock, inconsistent with the physics of free expansion. To this end, we present a plasma-vacuum interface tracking framework with the objective of predicting a physically consistent free expansion, devoid of the spurious shock obtained with the low background density approach. The interface tracking formulation is integrated within the MHD framework to simulate the plasma deflagration and the second discharge event, a plasma detonation, formed due to its initiation in a background prefilled with gas remnant from the deflagration. The mode transition behavior obtained in the simulations is qualitatively compared to that observed in the experiments using high framing rate Schlieren videography. The deflagration mode is further investigated to understand the jet formation process and the axial velocities obtained are compared against experimentally obtained deflagration plasma front velocities. The simulations are also used to provide insight into the conditions responsible for the generation and sustenance of the magnetic pinch. The pinch width and number density distribution are compared to experimentally obtained data to calibrate the inlet boundary conditions used to set up the plasma acceleration problem.
Diagnostic techniques in deflagration and detonation studies.
Proud, William G; Williamson, David M; Field, John E; Walley, Stephen M
2015-12-01
Advances in experimental, high-speed techniques can be used to explore the processes occurring within energetic materials. This review describes techniques used to study a wide range of processes: hot-spot formation, ignition thresholds, deflagration, sensitivity and finally the detonation process. As this is a wide field the focus will be on small-scale experiments and quantitative studies. It is important that such studies are linked to predictive models, which inform the experimental design process. The stimuli range includes, thermal ignition, drop-weight, Hopkinson Bar and Plate Impact studies. Studies made with inert simulants are also included as these are important in differentiating between reactive response and purely mechanical behaviour.
Pre-ignition confinement and deflagration violence in LX-10 and PBX 9501
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tringe, J. W., E-mail: tringe2@llnl.gov; Glascoe, E. A.; McClelland, M. A.
In thermal explosions of the nitramine octahydro-1,3,5,7-tetranitro-1,3,5,7-tetrazocine (HMX)-based explosives LX-10 and PBX-9501, the pre-ignition spatial and temporal heating profile defines the ignition location. The ignition location then determines the extent of inertial confinement and the violence of the resulting deflagration. In this work, we present results of experiments in which ∼23 g cylinders of LX-10 and PBX 9501 in thin-walled aluminum confinement vessels were subjected to identical heating profiles but which presented starkly different energy release signatures. Post-explosion LX-10 containment vessels were completely fragmented, while the PBX 9501 vessels were merely ruptured. Flash x-ray radiography images show that the initiation locationmore » for the LX-10 is a few mm farther from the end caps of the vessel relative to the initiation location of PBX 9501. This difference increases deflagration confinement for LX-10 at the time of ignition and extends the pressurization time during which the deflagration front propagates in the explosive. The variation in the initiation location, in turn, is determined by the thermal boundary conditions, which differ for these two explosives because of the larger coefficient of thermal expansion and greater thermal stability of the Viton binder in LX-10 relative to the estane and bis(2,2-dinitropropyl) acetal/formal binder of the PBX 9501. The thermal profile and initiation location were modeled for LX-10 using the hydrodynamics and structures code ALE3D; results indicate temperatures in the vicinity of the ignition location in excess of 274 °C near the time of ignition. The conductive burn rates for these two explosives, as determined by flash x-ray radiography, are comparable in the range 0.1–0.2 mm/μs, somewhat faster than rates observed by strand burner experiments for explosives in the temperature range 150–180 °C and pressures up to 100 MPa. The thinnest-wall aluminum containment vessels presented here rupture at lower pressures, in the range 10 MPa, suggesting that moderately higher temperatures and pressures are present near the deflagration front. For these explosives, however the most important property for determining deflagration violence is the degree of inertial confinement.« less
IMPACT OF NEW GAMOW–TELLER STRENGTHS ON EXPLOSIVE TYPE IA SUPERNOVA NUCLEOSYNTHESIS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mori, Kanji; Famiano, Michael A.; Kajino, Toshitaka
2016-12-20
Recent experimental results have confirmed a possible reduction in the Gamow–Teller (GT{sub +}) strengths of pf-shell nuclei. These proton-rich nuclei are of relevance in the deflagration and explosive burning phases of SNe Ia. While prior GT strengths result in nucleosynthesis predictions with a lower-than-expected electron fraction, a reduction in the GT{sub +} strength can result in a slightly increased electron fraction compared to previous shell model predictions, though the enhancement is not as large as previous enhancements in going from rates computed by Fuller, Fowler, and Newman based on an independent particle model. A shell model parametrization has been developed thatmore » more closely matches experimental GT strengths. The resultant electron-capture rates are used in nucleosynthesis calculations for carbon deflagration and explosion phases of SNe Ia, and the final mass fractions are compared to those obtained using more commonly used rates.« less
SN 1991T - Gamma-Ray Observatory's first supernova?
NASA Technical Reports Server (NTRS)
Burrows, Adam; Shankar, Anurag; Van Riper, Kenneth A.
1991-01-01
Consideration is given to the explosion of the Type Ia supernova SN 1991T in the nearby galaxy NGC 4527 detected in gamma-ray lines by the recently launched GRO. The dominant gamma-line and continuum features of the new 'delayed detonation' model FDEFA1 are calculated and compared to those for standard deflagration models W7 and cdtg7. It is shown that there are many useful hard photon discriminants of the Type Ia explosion mechanism that can, in principle, be detected by the OSSE and COMPTEL instruments on the GRO. Either SN 1991T, if bright enough, or one of the several Type Ia supernovae expected to be within the GRO's range during its active life, may make it possible to settle the detonation/deflagration debate, verify the generic thermonuclear white dwarf model of Type Ia explosions, and calibrate the Type Ia B(max)/847 keV line flux ratio.
Impact of New Gamow-Teller Strengths on Explosive Type Ia Supernova Nucleosynthesis
NASA Astrophysics Data System (ADS)
Mori, Kanji; Famiano, Michael A.; Kajino, Toshitaka; Suzuki, Toshio; Hidaka, Jun; Honma, Michio; Iwamoto, Koichi; Nomoto, Ken'ichi; Otsuka, Takaharu
2016-12-01
Recent experimental results have confirmed a possible reduction in the Gamow-Teller (GT+) strengths of pf-shell nuclei. These proton-rich nuclei are of relevance in the deflagration and explosive burning phases of SNe Ia. While prior GT strengths result in nucleosynthesis predictions with a lower-than-expected electron fraction, a reduction in the GT+ strength can result in a slightly increased electron fraction compared to previous shell model predictions, though the enhancement is not as large as previous enhancements in going from rates computed by Fuller, Fowler, and Newman based on an independent particle model. A shell model parametrization has been developed that more closely matches experimental GT strengths. The resultant electron-capture rates are used in nucleosynthesis calculations for carbon deflagration and explosion phases of SNe Ia, and the final mass fractions are compared to those obtained using more commonly used rates.
NASA Technical Reports Server (NTRS)
Fassnacht, P. O.
1971-01-01
A heat flux study of deflagrating pyrotechnic munitions is presented. Three tests were authorized to investigate whether heat flux measurements may be used as effective hazards evaluation criteria to determine safe quantity distances for pyrotechnics. A passive sensor study was conducted simultaneously to investigate their usefulness in recording events and conditions. It was concluded that heat flux measurements can effectively be used to evaluate hazards criteria and that passive sensors are an inexpensive tool to record certain events in the vicinity of deflagrating pyrotechnic stacks.
Shock implosion of a small homogeneous pellet
NASA Astrophysics Data System (ADS)
Fujimoto, Yasuichi; Mishkin, Eli A.; Alejaldre, Carlos
1985-10-01
A small spherical, or cylindrical, pellet is imploded by an intensive, evenly distributed, short energy pulse. At the surface of the pellet the matter ionizes, its temperature and pressure rapidly rise, and the ablated plasma, by reaction, implodes the inner nucleus of the pellet. The involved structure of the energy absorbing zone is idealized and a sharp deflagration front is considered. With an almost square energy pulse, slightly dropping with time, the solution of the mass, momentum, and energy conservation equations of the compressed matter, is self-similar. The differential equation of the nondimensional position of the deflagration front, its integral, and the magnitude and shape of the outside energy pulse are derived. The process of ablation is shown to depend solely on the nondimensional velocity of the gas just ahead of the deflagration front, minus the speed of sound, or the ratio of the gas densities across the deflagration front.
On the photometric homogeneity of Type IA supernovae
NASA Astrophysics Data System (ADS)
Bravo, E.; Dominguez, I.; Isern, J.; Canal, R.; Hoeflich, P.; Labay, J.
1993-03-01
The dependence of the characteristics of the light curves of Type Ia supernovae on the ignition density of the progenitor white dwarf is studied with the aid of two models of propagation of the thermonuclear burning front: as a deflagration and as a delayed detonation. The light curve is computed from opacities which take into account the velocity gradients. The results show that in all cases the resulting light curves roughly agree with observations and that they are not sensitive to the ignition density of the white dwarf. Only the model corresponding to a deflagration starting at a density of 8 x 10 exp 9 g/cu cm shows a deviation from the general behavior, having a significantly lower luminosity at maximum. A dispersion of about 1000 km/s is found in the computed expansion velocities at maximum, which compares well with that found in the observations.
Combustion of Nitramine Propellants
1983-03-01
through development of a comprehensive analytical model. The ultimate goals are to enable prediction of deflagration rate over a wide pressure range...superior in burn rate prediction , both simple models fail in correlating existing temperature- sensitivity data. (2) In the second part, a...auxiliary condition to enable independent burn rate prediction ; improved melt phase model including decomposition-gas bubbles; model for far-field
Chain-Thermal Explosions and the Transition from Deflagration Combustion to Detonation
NASA Astrophysics Data System (ADS)
Prokopenko, V. M.; Azatyan, V. V.
2018-01-01
The transition from combustion to a chain-thermal explosion, a necessary step in the transition from deflagration combustion into detonation, is studied using the example of hydrogen oxidation. Differences between the kinetic modes of ignition and a chain-thermal explosion are discussed.
Observation and modeling of deflagration-to-detonation transition (DDT) in low-density HMX
NASA Astrophysics Data System (ADS)
Tringe, Joseph W.; Vandersall, Kevin S.; Reaugh, John E.; Levie, Harold W.; Henson, Bryan F.; Smilowitz, Laura B.; Parker, Gary R.
2017-01-01
We employ simultaneous flash x-ray radiography and streak imaging, together with a multi-phase finite element model, to understand deflagration-to-detonation transition (DDT) phenomena in low-density (˜1.2 gm/cm3) powder of the explosive cyclotetramethylene-tetranitramine (HMX). HMX powder was lightly hand-tamped in a 12.7 mm diameter column, relatively lightly-confined in an optically-transparent polycarbonate cylinder with wall thickness 25.4 mm. We observe apparent compaction of the powder in advance of the detonation transition by the motion of small steel spheres pre-emplaced throughout the length of explosive. High-speed imaging along the explosive cylinder length provides a more temporally continuous record of the transition that is correlated with the high-resolution x-ray image record. Preliminary simulation of these experiments with the HERMES model implemented in the ALE3D code enables improved understanding of the explosive particle burning, compaction and detonation phenomena which are implied by the observed reaction rate and transition location within the cylinder.
NASA Technical Reports Server (NTRS)
Boggs, T. L.; Price, C. F.; Zurn, D. E.; Atwood, A. I.; Eisel, J. L.
1980-01-01
The thermal stability and resistance to impact was investigated for the ingredient TABA. Particular attention was given to determining the use of TABA as a possible alternative ingredient or substitute for HMX in explosives and high energy propellants. The burn rate of TABA was investigated as a function of pressure. It was concluded that the self deflagration rate of TABA is an order of magnitude lower than HMX over the range 2000-15000 psi; TABA will not sustain self deflagration at low pressures (less than or equal to 1500 psi) in the sample configuration and apparatus used.
On the Initiation Mechanism in Exploding Bridgewire and Laser Detonators
NASA Astrophysics Data System (ADS)
Stewart, D. Scott; Thomas, K.; Saenz, J.
2005-07-01
Since its invention by Los Alamos during the Manhattan Project era the exploding bridgewire detonator (EBW) has seen tremendous use and study. Recent development of a laser-powered device with detonation properties similar to an EBW is reviving interest in the basic physics of the Deflagration-to-Detonation (DDT) process in both of these devices,[1]. Cutback experiments using both laser interferometry and streak camera observations are providing new insight into the initiation mechanism in EBWs. These measurements are being correlated to a DDT model of compaction to detonation and shock to detonation developed previously by Xu and Stewart, [2]. The DDT model is incorporated into a high-resolution, multi-material model code for simulating the complete process. Model formulation and predictions against the test data will be discussed. REFS. [1] A. Munger, J. Kennedy, A. Akinci, and K. Thomas, "Dev. of a Laser Detonator" 30th Int. Pyrotechnics Seminar, Fort Collins, CO, (2004). [2] Xu, S. and Stewart, D. S. Deflagration to detonation transition in porous energetic materials: A model study. J. Eng. Math., 31, 143-172 (1997)
MESOSCALE MODELING OF DEFLAGRATION-INDUCED DECONSOLIDATION IN POLYMER-BONDED EXPLOSIVES
DOE Office of Scientific and Technical Information (OSTI.GOV)
Springer, H K; Glascoe, E A; Reaugh, J E
Initially undamaged polymer-bonded explosives can transition from conductive burning to more violent convective burning via rapid deconsolidation at higher pressures. The pressure-dependent infiltration of cracks and pores, i.e., damage, by product gases at the burn-front is a key step in the transition to convective burning. However, the relative influence of pre-existing damage and the evolution of deflagration-induced damage during the transition to convective burning is not well understood. The objective of this study is to investigate the role of microstructure and initial pressurization on deconsolidation. We performed simulations using the multi-physics hydrocode, ALE3D. HMX-Viton A served as our model explosive.more » A Prout-Tompkins chemical kinetic model, Vielle's Law pressure-dependent burning, Gruneisen equation-of-state, and simplified strength model were used for the HMX. The propensity for deconsolidation increased with increasing defect size and decreasing initial pressurization, as measured by the increase in burning surface area. These studies are important because they enable the development of continuum-scale damage models and the design of inherently safer explosives.« less
JANNAF workshop on hazards due to pre-combustion behavior of high energy propellants
NASA Technical Reports Server (NTRS)
Boggs, T. L.
1980-01-01
Experimental results and hazards associated with the pre-combustion behavior of high energy propellants are discussed. Special attention was given to future needs for modeling and the problem of deflagration to detonation transition.
NASA Astrophysics Data System (ADS)
Ranc-Darbord, Isabelle; Baudin, Gérard; Genetier, Marc; Ramel, David; Vasseur, Pierre; Legrand, Julien; Pina, Vincent
2018-03-01
Emission of gas and Al2O3 smoke within the deflagration of H2{-}O2-{N2{-}CO2}-Al particles has been studied in a closed combustion chamber at pressures of up to 18 bar and at gas temperatures of up to 3700 K. Measurements of radiance intensity were taken using a five wavelength pyrometer (0.660 μ m, 0.850 μ m, 1.083 μ m, 1.260 μ m, 1.481 μ m) and a grating spectrometer in the range (4.10 μ m to 4.30 μ m). In order to characterize the aluminum oxide smoke size and temperature, an inversion method has been developed based on the radiation transfer equation and using pyrometer measurements and thermochemical calculations of Al2O3 smoke volume fractions. Temperatures in combustion gas have been determined using a method based on the assumed blackbody head of the 4.26 μ m CO2 emission line and on its spectral shift with pressure and temperature. For validation purpose, this method has been applied to measurements obtained when calibrated alumina particles are injected in a combustion chamber prior to gaseous deflagrations. This mathematical inversion method was developed to investigate explosive fireballs.
NASA Technical Reports Server (NTRS)
Boggs, T. L.; Price, C. F.; Atwood, A. I.; Zurn, D. E.; Eisel, J. L.; Derr, R. L.
1980-01-01
The inadequacies of the two commonly used assumptions are shown, along with the need for considering gas phase reactions. Kinetic parameters that describe the gas phase reactions for several ingredients are provided, and the first steps in convective combustion leading to deflagration to detonation transition are described.
Modelling of Deflagration to Detonation Transition in Porous PETN of Density 1.4 g / cc with HERMES
NASA Astrophysics Data System (ADS)
Reaugh, John; Curtis, John; Maheswaran, Mary-Ann
2017-06-01
The modelling of Deflagration to Detonation Transition in explosives is a severe challenge for reactive burn models because of the complexity of the physics; there is mechanical and thermal interaction of the gaseous burn products with the burning porous matrix, with resulting compaction, shock formation and subsequent detonation. Experiments on the explosive PETN show a strong dependence of run distance to detonation on porosity. The minimum run distance appears to occur when the density is approximately 1.4 g / cc. Recent research on the High Explosive Response to Mechanical Stimulation (HERMES) model for High Explosive Violent Reaction has included the development of a model for PETN at 1.4 g / cc., which allows the prediction of the run distance in the experiments for PETN at this density. Detonation and retonation waves as seen in the experiment are evident. The HERMES simulations are analysed to help illuminate the physics occurring in the experiments. JER's work was performed under the auspices of the US DOE by LLNL under Contract DE-AC52-07NA27344 and partially funded by the Joint US DoD/DOE Munitions Technology Development Program. LLNL-ABS-723537.
Deflagration-to-Detonation Transition Control by Nanosecond Gas Discharges
2008-04-07
Report 3. DATES COVERED (From – To) 1 April 2007 - 18 August 09 4. TITLE AND SUBTITLE Deflagration-To- Detonation Transition Control By Nanosecond...SUPPLEMENTARY NOTES 14. ABSTRACT During the current project, an extensive experimental study of detonation initiation by high{voltage...nanosecond gas discharges has been performed in a smooth detonation tube with different discharge chambers and various discharge cell numbers. The chambers
Simulations of Flame Acceleration and Deflagration-to-Detonation Transitions in Methane-Air Systems
2010-03-17
are neglected. 3. Model parameter calibration The one-step Arrhenius kinetics used in this model cannot ex- actly reproduce all properties of laminar...with obstacles are compared to previ- ously reported experimental data. The results obtained using the simple reaction model qualitatively, and in...have taken in developing a multidimensional numerical model to study explosions in large-scale systems containing mixtures of nat- ural gas and air
Detonation initiation of heterogeneous melt-cast high explosives
NASA Astrophysics Data System (ADS)
Chuzeville, V.; Baudin, G.; Lefrançois, A.; Genetier, M.; Barbarin, Y.; Jacquet, L.; Lhopitault, J.-L.; Peix, J.; Boulanger, R.; Catoire, L.
2017-01-01
2,4,6-trinitrotoluene (TNT) is widely used in conventional and insensitive munitions as a fusible binder, commonly melt-cast with other explosives such as 1,3,5-trinitroperhydro-1,3,5-triazine (RDX) or 3-nitro-1,2,4-triazol-one (NTO). In this paper, we study the shock-to-detonation transition phenomenon in two melt-cast high explosives (HE). We have performed plate impact tests on wedge samples to measure run-distance and time-to-detonation in order to establish the Pop-plot relation for several melt-cast HE. Highlighting the existence of the single curve buildup, we propose a two phase model based on a Zeldovich, Von-Neumann, Döring (ZND) approach where the deflagration fronts grow from the explosive grain boundaries. Knowing the grain size distribution, we calculate the deflagration velocities of the explosive charges as a function of shock pressure and explore the possible grain fragmentation.
Deflagrations, Detonations, and the Deflagration-to-Detonation Transition in Methane-Air Mixtures
2011-04-27
we attempt to answer the question: Given a large enough volume of flammable mixture of NG and air, can a weak spark ignition develop into a...detonation? Large -scale numerical simulations, in conjunction with experimental work conducted at the National Institute for Occupational Safety and...12 2.3.3. Flame Acceleration and DDT in Channels with Obstacles . . . . . . . . . . . . . 14 2.3.4. DDT in Large Spaces
Propagation of Avalanches in Mn12-Acetate: Magnetic Deflagration
NASA Astrophysics Data System (ADS)
Suzuki, Yoko; Sarachik, M. P.; Chudnovsky, E. M.; McHugh, S.; Gonzalez-Rubio, R.; Avraham, Nurit; Myasoedov, Y.; Zeldov, E.; Shtrikman, H.; Chakov, N. E.; Christou, G.
2005-09-01
Local time-resolved measurements of fast reversal of the magnetization of single crystals of Mn12-acetate indicate that the magnetization avalanche spreads as a narrow interface that propagates through the crystal at a constant velocity that is roughly 2 orders of magnitude smaller than the speed of sound. We argue that this phenomenon is closely analogous to the propagation of a flame front (deflagration) through a flammable chemical substance.
Razus, D; Brinzea, V; Mitu, M; Movileanu, C; Oancea, D
2011-06-15
The maximum rates of pressure rise during closed vessel explosions of propane-air mixtures are reported, for systems with various initial concentrations, pressures and temperatures ([C(3)H(8)]=2.50-6.20 vol.%, p(0)=0.3-1.3 bar; T(0)=298-423 K). Experiments were performed in a spherical vessel (Φ=10 cm) with central ignition. The deflagration (severity) index K(G), calculated from experimental values of maximum rates of pressure rise is examined against the adiabatic deflagration index, K(G, ad), computed from normal burning velocities and peak explosion pressures. At constant temperature and fuel/oxygen ratio, both the maximum rates of pressure rise and the deflagration indices are linear functions of total initial pressure, as reported for other fuel-air mixtures. At constant initial pressure and composition, the maximum rates of pressure rise and deflagration indices are slightly influenced by the initial temperature; some influence of the initial temperature on maximum rates of pressure rise is observed only for propane-air mixtures far from stoichiometric composition. The differentiated temperature influence on the normal burning velocities and the peak explosion pressures might explain this behaviour. Copyright © 2011 Elsevier B.V. All rights reserved.
Cars Spectroscopy of Propellant Flames
1983-11-01
applicability of CARS in studies of the combustion of propellants and other reactive systems. Broadband CARS spectra were obtained from both the reaction zone...ref 12). When ienited vith a flame, propellant burned in air with a luainous flame. A-e Ignittou with i hot wire resulted in flameless burning (fizz...ester). Current models of nitramine propellant combustion are essentially models of HMX (cyclotetranithylene tetranitramine) and RDX deflagration. The
Deflagration to Detonation Transition Processes in Pulsed Detonation Engines
2002-08-03
which subsequently leads to DDT. The modelling approach taken here is as outlined by Arntzen et al. [9] and features a fractal based eddy-breakup... Arntzen , B.J., Hjertager, B., Lindstedt, R.P., Mercx, W.P.M. and Popat, N. “Investigations to Improve and Assess the Accuracy of Computational Fluid
DOE Office of Scientific and Technical Information (OSTI.GOV)
Parker, Gary R. Jr.; Holmes, Matthew D.; Dickson, Peter
Conventional high explosives (e.g. PBX 9501, LX-07) have been observed to react violently following thermal insult: (1) Fast convective and compressive burns (HEVR); (2) Thermal explosions (HEVR); and (3) Deflagration-to-detonation transition (DDT). No models exist that sufficiently capture/predict these complex multiphase and multiscale behaviors. For now, research is focused on identifying vulnerabilities and factors that control this behavior.
Observation and modeling of deflagration-to-detonation (DDT) transition in low-density HMX
NASA Astrophysics Data System (ADS)
Tringe, Joseph; Vandersall, Kevin; Reaugh, Jack; Levie, Harold; Henson, Bryan; Smilowitz, Laura; Parker, Gary
2015-06-01
We employ simultaneous flash x-ray radiography and streak imaging, together with a multi-phase finite element model, to understand deflagration-to-detonation transition (DDT) phenomena in low-density (~ 1.2 gm/cm3) powder of the explosive cyclotetramethylene-tetranitramine (HMX). HMX powder was lightly hand-tamped in a 12.7 mm diameter column, relatively lightly-confined in an optically-transparent polycarbonate cylinder with wall thickness 25.4 mm. We observe apparent compaction of the powder in advance of the detonation transition, both by x-ray contrast and by the motion of small steel spheres pre-emplaced throughout the length of explosive. High-speed imaging along the explosive cylinder length provides a temporally continuous record of the transition that is correlated with the high-resolution x-ray image record. Preliminary simulation of these experiments with the HERMES model implemented in the ALE3D code enables improved understanding of the explosive particle burning, compaction and detonation phenomena which are implied by the observed reaction rate and transition location within the cylinder. This work performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Acceleration Modes and Transitions in Pulsed Plasma Accelerators
NASA Technical Reports Server (NTRS)
Polzin, Kurt A.; Greve, Christine M.
2018-01-01
Pulsed plasma accelerators typically operate by storing energy in a capacitor bank and then discharging this energy through a gas, ionizing and accelerating it through the Lorentz body force. Two plasma accelerator types employing this general scheme have typically been studied: the gas-fed pulsed plasma thruster and the quasi-steady magnetoplasmadynamic (MPD) accelerator. The gas-fed pulsed plasma accelerator is generally represented as a completely transient device discharging in approximately 1-10 microseconds. When the capacitor bank is discharged through the gas, a current sheet forms at the breech of the thruster and propagates forward under a j (current density) by B (magnetic field) body force, entraining propellant it encounters. This process is sometimes referred to as detonation-mode acceleration because the current sheet representation approximates that of a strong shock propagating through the gas. Acceleration of the initial current sheet ceases when either the current sheet reaches the end of the device and is ejected or when the current in the circuit reverses, striking a new current sheet at the breech and depriving the initial sheet of additional acceleration. In the quasi-steady MPD accelerator, the pulse is lengthened to approximately 1 millisecond or longer and maintained at an approximately constant level during discharge. The time over which the transient phenomena experienced during startup typically occur is short relative to the overall discharge time, which is now long enough for the plasma to assume a relatively steady-state configuration. The ionized gas flows through a stationary current channel in a manner that is sometimes referred to as the deflagration-mode of operation. The plasma experiences electromagnetic acceleration as it flows through the current channel towards the exit of the device. A device that had a short pulse length but appeared to operate in a plasma acceleration regime different from the gas-fed pulsed plasma accelerators was developed by Cheng, et al. The Coaxial High ENerGy (CHENG) thruster operated on the 10-microseconds timescales of pulsed plasma thrusters, but claimed high thrust density, high efficiency and low electrode erosion rates, which are more consistent with the deflagration mode of acceleration. Separate work on gas-fed pulsed plasma thrusters (PPTs) by Ziemer, et al. identified two separate regimes of performance. The regime at higher mass bits (termed Mode I in that work) possessed relatively constant thrust efficiency (ratio of jet kinetic energy to input electrical energy) as a function of mass bit. In the second regime at very low mass bits (termed Mode II), the efficiency increased with decreasing mass bit. Work by Poehlmann et al. and by Sitaraman and Raja sought to understand the performance of the CHENG thruster and the Mode I / Mode II performance in PPTs by modeling the acceleration using the Hugoniot Relation, with the detonation and deflagration modes representing two distinct sets of solutions to the relevant conservation laws. These works studied the proposal that, depending upon the values of the various controllable parameters, the accelerator would operate in either the detonation or deflagration mode. In the present work, we propose a variation on the explanation for the differences in performance between the various pulsed plasma accelerators. Instead of treating the accelerator as if it were only operating in one mode or the other during a pulse, we model the initial stage of the discharge in all cases as an accelerating current sheet (detonation mode). If the current sheet reaches the exit of the accelerator before the discharge is completed, the acceleration mode transitions to the deflagration mode type found in the quasi-steady MPD thrusters. This modeling method is used to demonstrate that standard gas-fed pulsed plasma accelerators, the CHENG thruster, and the quasi-steady MPD accelerator are variations of the same device, with the overall acceleration of the plasma depending upon the behavior of the plasma discharge during initial transient phase and the relative lengths of the detonation and deflagration modes of operation.
Two-Dimensional Failure Waves and Ignition Fronts in Premixed Combustion
NASA Technical Reports Server (NTRS)
Vedarajan, T. G.; Buckmaster J.; Ronney, P.
1998-01-01
This paper is a continuation of our work on edge-flames in premixed combustion. An edge-flame is a two-dimensional structure constructed from a one-dimensional configuration that has two stable solutions (bistable equilibrium). Edge-flames can display wavelike behavior, advancing as ignition fronts or retreating as failure waves. Here we consider two one-dimensional configurations: twin deflagrations in a straining flow generated by the counterflow of fresh streams of mixture: and a single deflagration subject to radiation losses. The edge-flames constructed from the first configuration have positive or negative speeds, according to the value of the strain rate. But our numerical solutions strongly suggest that only positive speeds (corresponding to ignition fronts) can exist for the second configuration. We show that this phenomenon can also occur in diffusion flames when the Lewis numbers are small. And we discuss the asymptotics of the one-dimensional twin deflagration configuration. an overlooked problem from the 70s.
Deflagration-to-detonation transition in gases in tubes with cavities
NASA Astrophysics Data System (ADS)
Smirnov, N. N.; Nikitin, V. F.; Phylippov, Yu. G.
2010-12-01
The existence of a supersonic second combustion mode — detonation — discovered by Mallard and Le Chatelier and by Berthélot and Vieille in 1881 posed the question of mechanisms for transition from one mode to the other. In the period 1959-1969, experiments by Salamandra, Soloukhin, Oppenheim, and their coworkers provided insights into this complex phenomenon. Since then, among all the phenomena related to combustion processes, deflagration-to-detonation transition is, undoubtedly, the most intriguing one. Deflagration-to-detonation transition (DDT) in gases is connected with gas and vapor explosion safety issues. Knowing mechanisms of detonation onset control is of major importance for creating effective mitigation measures addressing two major goals: to prevent DDT in the case of mixture ignition, or to arrest the detonation wave in the case where it has been initiated. A new impetus to the increase in interest in deflagration-to-detonation transition processes was given by the recent development of pulse detonation devices. The probable application of these principles to creation of a new generation of engines put the problem of effectiveness of pulse detonating devices at the top of current research needs. The effectiveness of the pulse detonation cycle turned out to be the key factor characterizing the Pulse Detonation Engine (PDE), whose operation modes were shown to be closely related to periodical onset and degeneration of a detonation wave. Those unsteady-state regimes should be self-sustained to guarantee a reliable operation of devices using the detonation mode of burning fuels as a constitutive part of their working cycle. Thus deflagration-to-detonation transition processes are of major importance for the issue. Minimizing the predetonation length and ensuring stability of the onset of detonation enable one to increase the effectiveness of a PDE. The DDT turned out to be the key factor characterizing the PDE operating cycle. Thus, the problem of DDT control in gaseous fuel-air mixtures became very acute. This paper contains results of theoretical and experimental investigations of DDT processes in combustible gaseous mixtures. In particular, the paper investigates the effect of cavities incorporated in detonation tubes at the onset of detonation in gases. Extensive numerical modeling and simulations allowed studying the features of deflagration-to-detonation transition in gases in tubes incorporating cavities of a wider cross section. The presence of cavities substantially affects the combustion modes being established in the device and their dependence on the governing parameters of the problem. The influence of geometrical characteristics of the confinement and flow turbulization on the onset of detonation and the influence of temperature and fuel concentration in the unburned mixture are discussed. It was demonstrated both experimentally and theoretically that the presence of cavities of wider cross section in the ignition part of the tube promotes DDT and shortens the predetonation length. At the same time, cavities incorporated along the whole length or in the far-end section inhibit detonation and bring about the onset of low-velocity galloping detonation or galloping combustion modes. The presence of cavities in the ignition section turns an increase in the initial mixture temperature into a DDT-promoting factor instead of a DDT-inhibiting factor.
Nitramine Monopropellant Deflagration and General Nonsteady Reacting Rocket Chamber Flows.
1980-01-01
experimentally in RDX crystals, with varying degrees of molecular symmetry (concerning the N o groups ) and one is a boat conformer. One of these chair...regarding deflagration control by condensed phase decomposition at lower pressures is unwarranted, although their results follow the experimental data, for...two groups of experimental data were correlated by use of Eq. (III.19b) . The low pressure (0.5 to 6.6 MPa) mass burning rate measurements of Bobolev
Development and Execution of a Large-scale DDT Tube Test for IHE Material Qualification
DOE Office of Scientific and Technical Information (OSTI.GOV)
Parker, Gary Robert; Broilo, Robert M.; Lopez-Pulliam, Ian Daniel
Insensitive High Explosive (IHE) Materials are defined in Chapter IX of the DOE Explosive Safety Standard (DOE-STD-1212-2012) as being materials that are massdetonable explosives that are so insensitive that the probability of accidental initiation or transition from burning to detonation is negligible1. There are currently a number of tests included in the standard that are required to qualify a material as IHE, however, none of the tests directly evaluate for the transition from burning to detonation (aka deflagration-to-detonation transition, DDT). Currently, there is a DOE complex-wide effort to revisit the IHE definition in DOE-STD-1212-2012 and change the qualification requirements. Themore » proposal lays out a new approach, requiring fewer, but more appropriate tests, for IHE Material qualification. One of these new tests is the Deflagration-to-Detonation Test. According to the redefinition proposal, the purpose of the new deflagration-todetonation test is “to demonstrate that an IHE material will not undergo deflagration-to-detonation under stockpile relevant conditions of scale, confinement, and material condition. Inherent in this test design is the assumption that ignition does occur, with onset of deflagration. The test design will incorporate large margins and replicates to account for the stochastic nature of DDT events.” In short, the philosophy behind this approach is that if a material fails to undergo DDT in a significant over-test, then it is extremely unlikely to do so in realistic conditions. This effort will be valuable for the B61 LEP to satisfy their need qualify the new production lots of PBX 9502. The work described in this report is intended as a preliminary investigation to support the proposed design of an overly conservative, easily fielded DDT test for updated IHE Material Qualification standard. Specifically, we evaluated the aspects of confinement, geometry, material morphology and temperature. We also developed and tested a thermally robust igniter system.« less
NASA Technical Reports Server (NTRS)
Cheng, D. Y.; Wang, P.
1972-01-01
The injection of dense plasmas into a B sub z long magnetic field from both ends of the field coil was investigated. Deflagration plasma guns and continuous flow Z-pinch are discussed along with the possibility of a continuous flow Z-pinch fusion reactor. The injection experiments are described with emphasis on the synchronization of the two plasma deflagration guns, the collision of the two plasma beams, and the determination of plasma density.
NASA Astrophysics Data System (ADS)
Shevchenko, A. A.; Dolgoborodov, A. Yu; Kirilenko, V. G.; Brazhnikov, M. A.
2016-11-01
Deflagration-to-detonation transition (DDT) in aluminum-ammonium perchlorate (Al/AP) loose-packed charges (80% porosity) has been studied. The charges were manufactured from preliminary mechanoactivated mixtures. The mixtures placed in steel tubes 10 mm in diameter were ignited by Nichrome wire. It was found that it is possible to distinguish three parts corresponding to different stages of DDT process development. Steady-state detonation velocity reached the level of 2500 m/s at the distance of 90 mm from the ignition point.
The role and importance of porosity in the deflagration rates of HMX-based materials
DOE Office of Scientific and Technical Information (OSTI.GOV)
Glascoe, E A; Hsu, P C; Springer, H K
The deflagration behavior of thermally damaged HMX-based materials will be discussed. Strands of material were burned at pressures ranging from 10-300 MPa using the LLNL high pressure strand burner. Strands were heated in-situ and burned while still hot; temperatures range from 90-200 C and were chosen in order to allow for thermal damage of the material without significant decomposition of the HMX. The results indicate that multiple variables affect the burn rate but the most important are the polymorph of HMX and the nature and thermal stability of the non-HE portion of the material. Characterization of the strands indicate thatmore » the thermal soak produces significant porosity and permeability in the sample allowing for significantly faster burning due to the increased surface area and new pathways for flame spread into the material. Specifically, the deflagration rates of heated PBXN-9, LX-10, and PBX-9501 will be discussed and compared.« less
Experimental and Numerical Study of Ammonium Perchlorate Counterflow Diffusion Flames
NASA Technical Reports Server (NTRS)
Smooke, M. D.; Yetter, R. A.; Parr, T. P.; Hanson-Parr, D. M.; Tanoff, M. A.
1999-01-01
Many solid rocket propellants are based on a composite mixture of ammonium perchlorate (AP) oxidizer and polymeric binder fuels. In these propellants, complex three-dimensional diffusion flame structures between the AP and binder decomposition products, dependent upon the length scales of the heterogeneous mixture, drive the combustion via heat transfer back to the surface. Changing the AP crystal size changes the burn rate of such propellants. Large AP crystals are governed by the cooler AP self-deflagration flame and burn slowly, while small AP crystals are governed more by the hot diffusion flame with the binder and burn faster. This allows control of composite propellant ballistic properties via particle size variation. Previous measurements on these diffusion flames in the planar two-dimensional sandwich configuration yielded insight into controlling flame structure, but there are several drawbacks that make comparison with modeling difficult. First, the flames are two-dimensional and this makes modeling much more complex computationally than with one-dimensional problems, such as RDX self- and laser-supported deflagration. In addition, little is known about the nature, concentration, and evolution rates of the gaseous chemical species produced by the various binders as they decompose. This makes comparison with models quite difficult. Alternatively, counterflow flames provide an excellent geometric configuration within which AP/binder diffusion flames can be studied both experimentally and computationally.
The Chemical Signature of SNIax in the Stars of Ursa Minor?
NASA Astrophysics Data System (ADS)
Cescutti, Gabriele; Kobayashi, Chiaki
2018-06-01
Recently, a new class of supernovae Ia was discovered: the supernovae Iax; the increasing sample of these objects share common features as lower maximum-light velocities and typically lower peak magnitudes.In our scenario, the progenitors of the SNe Iax are very massive white dwarfs, possibly hybrid C+O+Ne white dwarfs; due to the accretion from a binary companion, they reach the Chandrasekhar mass and undergo a central carbon deflagration, but the deflagration is quenched when it reaches the outer O+Ne layer. This class of SNe Ia are expected to be rarer than standard SNe Ia and do not affect the chemical evolution in the solar neighbourhood; however, they have a short delay time and they could influence the evolution of metal-poor systems. Therefore, we have included in a stochastic chemical evolution model for the dwarf spheroidal galaxy Ursa minor the contribution of SNe Iax.The model predicts a spread in [Mn/Fe] in the ISM medium at low metallicity and - at the same time - a decrease of the [alpha/Fe] elements, as in the classical time delay model. This is in surprising agreement with the observed abundances in stars of Ursa minor and provide a strong indication to the origin of this new classes of SNIa.
NASA Technical Reports Server (NTRS)
Baron, E.; Cooperstein, J.; Kahana, S.; Nomoto, K.
1987-01-01
The results of the hydrodynamic collapse of an accreting C + O white dwarf are presented. Collapse is induced by electron captures in the iron core behind a conductive deflagration front. The shock wave produced by the hydrodynamic bounce of the iron core stalls at about 115 km, and thus a neutron star formed in such a model would be formed as an optically quiet event.
A Pressure-Dependent Damage Model for Energetic Materials
2013-04-01
appropriate damage nucleation and evolution laws, and the equation of state ) with its reactive response. 15. SUBJECT TERMS pressure-dependent...evolution laws, and the equation of state ) with its reactive response. INTRODUCTION Explosions and deflagrations are classifications of sub-detonative...energetic material’s mechanical response (through the yield criterion, damage evolution and equation of state ) with its reactive response. DAMAGE-FREE
Numerical modeling of deflagration mode in coaxial plasma guns
NASA Astrophysics Data System (ADS)
Sitaraman, Hariswaran; Raja, Laxminarayan
2012-10-01
Pulsed coaxial plasma guns have been used in several applications in the field of space propulsion, nuclear fusion and materials processing. These devices operate in two modes based on the delay between gas injection and breakdown initiation. Larger delay led to the plasma detonation mode where a compression wave in the form of a luminous front propagates from the breech to the muzzle. Shorter delay led to the more efficient deflagration mode characterized by a relatively diffuse plasma with higher resistivity. The overall physics of the discharge in the two modes of operation and in particular the latter remain relatively unexplored. Here we perform a computational modeling study by solving the non-ideal Magneto-hydrodynamics equations for the quasi-neutral plasma in the coaxial plasma gun. A finite volume formulation on an unstructured mesh framework with an implicit scheme is used to do stable computations. The final work will present details of important species in the plasma, particle energies and Mach number at the muzzle. A comparison of the plasma parameters will be made with the experiments reported in ref. [1]. [4pt] [1] F. R. Poehlmann et al., Phys. Plasmas 17, 123508 (2010)
NASA Technical Reports Server (NTRS)
Nomoto, K.
1981-01-01
As a plausible explosion model for a Type I supernova, the evolution of carbon-oxygen white dwarfs accreting helium in binary systems was investigated from the onset of accretion up to the point at which a thermonuclear explosion occurs. The relationship between the conditions in the binary system and the triggering mechanism for the supernova explosion is discussed, especially for the cases with relatively slow accretion rate. It is found that the growth of a helium zone on the carbon-oxygen core leads to a supernova explosion which is triggered either by the off-center helium detonation for slow and intermediate accretion rates or by the carbon deflagration for slow and rapid accretion rates. Both helium detonation and carbon deflagration are possible for the case of slow accretion, since in this case the initial mass of the white dwarf is an important parameter for determining the mode of ignition. Finally, various modes of building up the helium zone on the white dwarf, namely, direct transfer of helium from the companion star and the various types and strength of the hydrogen shell flashes are discussed in some detail.
Funsten, Herbert O.; McComas, David J.
1997-01-01
Apparatus and method for rapid detection of explosives residue from the deflagration signature thereof. A property inherent to most explosives is their stickiness, resulting in a strong tendency of explosive particulate to contaminate the environment of a bulk explosive. An apparatus for collection of residue particulate, burning the collected particulate, and measurement of the optical emission produced thereby is described. The present invention can be utilized for real-time screening of personnel, cars, packages, suspected devices, etc., and provides an inexpensive, portable, and noninvasive means for detecting explosives.
Funsten, Herbert O.; McComas, David J.
1999-01-01
Apparatus and method for rapid detection of explosives residue from the deflagration signature thereof. A property inherent to most explosives is their stickiness, resulting in a strong tendency of explosive particulate to contaminate the environment of a bulk explosive. An apparatus for collection of residue particulate, burning the collected particulate, and measurement of the ultraviolet emission produced thereby, is described. The present invention can be utilized for real-time screening of personnel, cars, packages, suspected devices, etc., and provides an inexpensive, portable, and noninvasive means for detecting explosives.
Effect of Nano-Aluminum and Fumed Silica Particles on Deflagration and Detonation of Nitromethane
2009-01-01
0178 Justin L. Sabourin a, Richard A. Yetter a, Blaine W. Asay b, Joseph M. Lloyd b, Victor E. Sanders b, Grant A. Risha c, Steven F. Son d a The...Fumed Silica Particles on Deflagration and Detonation of Nitromethane Justin L. Sabourin *, Richard A. Yetter The Pennsylvania State University...containing nAl, which was also found by other workers. 386 J. L. Sabourin , R. A. Yetter, B. W. Asay, J. M. Lloyd, V. E. Sanders, G. A. Risha, S. F. Son
Low voltage nonprimary explosive detonator
Dinegar, Robert H.; Kirkham, John
1982-01-01
A low voltage, electrically actuated, nonprimary explosive detonator is disclosed wherein said detonation is achieved by means of an explosive train in which a deflagration-to-detonation transition is made to occur. The explosive train is confined within a cylindrical body and positioned adjacent to low voltage ignition means have electrical leads extending outwardly from the cylindrical confining body. Application of a low voltage current to the electrical leads ignites a self-sustained deflagration in a donor portion of the explosive train which then is made to undergo a transition to detonation further down the train.
White dwarf models for type 1 supernovae and quiet supernovae, and presupernova evolution
NASA Technical Reports Server (NTRS)
Nomoto, K.
1980-01-01
Supernova mechanisms in accreting white dwarfs are considered with emphasis on deflagration as a plausible mechanism for producing Type I supernovae and electron captures to form quiet supernovae leaving neutron stars. These outcomes depend on accretion rate of helium, initial mass and composition of the white dwarf. The various types of hydrogen shell burning in the presupernova stage are also discussed.
Comustion of HAN-Based Monopropellant Droplets in Reduced Gravity
NASA Technical Reports Server (NTRS)
Shaw, B. D.
2001-01-01
Hydroxylammonium nitrate (HAN) is a major constituent in a class of liquid monopropellants that have many attractive characteristics and which display phenomena that differ significantly from other liquid monopropellants. They are composed primarily of HAN, H2O and a fuel species, often triethanolammonium nitrate (TEAN). HAN-based propellants have attracted attention as liquid gun propellants, and are attractive for NASA spacecraft propulsion applications. A representative propellant is XM46. This mixture is 60.8% HAN, 19.2% TEAN and 20% H2O by weight. Other HAN-based propellant mixtures are also of interest. For example, methanol and glycine have been investigated as potential fuel species for HAN-based monopropellants for thruster applications. In the present research, experimental and theoretical studies are performed on combustion of HAN-based monopropellant droplets. The fuel species considered are TEAN, methanol and glycine. Droplets initially in the mm size range are studied at pressures up to 30 atm. These pressures are applicable to spacecraft thruster applications. The droplets are placed in environments with various amounts of Ar, N2, O2, NO2 and N2O. Reduced gravity is employed to enable observations of burning rates and flame structures to be made without the complicating effects of buoyant and forced convection. Normal gravity experiments are also performed in this research program. The experiment goals are to provide accurate fundamental data on deflagration rates, gasphase temperature profiles, transient gas-phase flame behaviors, the onset of bubbling in droplets at lower pressures, and the low-pressure deflagration limit. Theoretical studies are performed to provide rational models of deflagration mechanisms of HAN-based liquid propellants. Besides advancing fundamental knowledge, this research should aid in applications (e.g., spacecraft thrusters and liquid propellant guns) of this unique class of monopropellants.
A pocket model for aluminum agglomeration in composite propellants
NASA Technical Reports Server (NTRS)
Cohen, N. S.
1981-01-01
This paper presents a model for the purpose of estimating the fraction of aluminum powder that will form agglomerates at the surface of deflagrating composite propellants. The basic idea is that the fraction agglomerated depends upon the amount of aluminum that melts within effective binder pocket volumes framed by oxidizer particles. The effective pocket depends upon the ability of ammonium perchlorate modals to encapsulate the aluminum and provide a local temperature sufficient to ignite the aluminum. Model results are discussed in the light of data showing effects of propellant formulation variables and pressure.
Fast Hydrogen-Air Flames for Turbulence Driven Deflagration to Detonation Transition
NASA Astrophysics Data System (ADS)
Chambers, Jessica; Ahmed, Kareem
2016-11-01
Flame acceleration to Detonation produces several combustion modes as the Deflagration-to-Detonation Transition (DDT) is initiated, including fast deflagration, auto-ignition, and quasi-detonation. Shock flame interactions and turbulence levels in the reactant mixture drive rapid flame expansion, formation of a leading shockwave and post-shock conditions. An experimental study to characterize the developing shock and flame front behavior of propagating premixed hydrogen-air flames in a square channel is presented. To produce each flame regime, turbulence levels and flame propagation velocity are controlled using perforated plates in several configurations within the experimental facility. High speed optical diagnostics including Schlieren and Particle Image Velocimetry are used to capture the flow field. In-flow pressure measurements acquired post-shock, detail the dynamic changes that occur in the compressed gas directly ahead of the propagating flame. Emphasis on characterizing the turbulent post-shock environment of the various flame regimes helps identify the optimum conditions to initiate the DDT process. The study aims to further the understanding of complex physical mechanisms that drive transient flame conditions for detonation initiation. American Chemical Society.
Funsten, H.O.; McComas, D.J.
1999-06-15
Apparatus and method are disclosed for rapid detection of explosives residue from the deflagration signature thereof. A property inherent to most explosives is their stickiness, resulting in a strong tendency of explosive particulate to contaminate the environment of a bulk explosive. An apparatus for collection of residue particulate, burning the collected particulate, and measurement of the ultraviolet emission produced thereby, is described. The present invention can be utilized for real-time screening of personnel, cars, packages, suspected devices, etc., and provides an inexpensive, portable, and noninvasive means for detecting explosives. 4 figs.
The delayed-detonation model of Type Ia supernovae. 2: The detonation phase
NASA Technical Reports Server (NTRS)
Arnett, David; Livne, Eli
1994-01-01
The investigation, by use of two-dimensional numerical hydrodynamics simulations, of the 'delayed detonation' mechanism of Khokhlov for the explosion of Type Ia supernovae is continued. Previously we found that the deflagration is insufficient to unbind the star. Expansion shuts off the flame; much of this small production of iron group nuclei occurs at lower densities, which reduces the electron-capture problem. Because the degenerate star has an adiabatic exponent only slightly above 4/3, the energy released by deflagration drives a pulsation of large amplitude. During the first expansion phase, adiabatic cooling shuts off the burning, and a Rayleigh-Taylor instability then gives mixing of high-entropy ashes with low-entropy fuel. During the first contraction phase, compressional heating reignites the material. The burning was allowed to develop into a detonation in these nonspherical models. The detonation grows toward spherical symmetry at late times. At these densities (rho approx. 10(exp 7) to 10(exp 8) g cm(exp -3)), either Ni-56 or nuclei of the Si-Ca group are the dominant products of the burning. The bulk yields are sensitive to the density of the star when the transition to detonation occurs. The relevance of the abundances, velocities, mixing, and total energy release to the theory and interpretation of Type Ia supernovae is discussed.
Non-Local Thermodynamic Equilibrium Spectrum Synthesis of Type IA Supernovae
NASA Astrophysics Data System (ADS)
Nugent, Peter Edward
1997-09-01
Type Ia supernovae (SNe Ia) are valuable distance indicators for cosmology and the elements they eject are are important for nucleosynthesis. They appear to be thermonuclear disruptions of carbon-oxygen white dwarfs that accrete from companion stars until they approach the Chandrasekbar mass, and there is a suspicion that the propagation of the nuclear burning front involves a transition from a deflagration to a detonation. Detailed modeling of the atmospheres and spectra of SNe Ia is needed to advance our understanding of SNe Ia. Comparison of synthetic and observed spectra provides information on the temperature, density, velocity, and composition of the ejected matter and thus constrain hydrodynamical models. In addition, the expanding photosphere method yields distances to individual events that are independent of distances based on the decay of 56Ni in SNe Ia and of Cepheid variable stars in the parent galaxies. This thesis is broken down into 4 major sections, each highlighting a different way with which to use spectrum synthesis to analyze SNe Ia. Chapters 2 and 3 look at normal SNe Ia and their potential use as distance indicators using SEAM. Chapter 4 examines spectral correlations with luminosity in SNe Ia and provides a plausible explanation for these correlations via spectrum synthesis. In Chapter 5 the spectra of various hydrodynamical models are calculated in an effort to answer the question of which current progenitor/explosion model is the most plausible for a SN Ia. Finally, we look at the importance of NLTE calculations and line identifications in Chapter 6. Also included are two appendices which contain more technical information concerning γ-ray deposition and the thermalization parameter.
Deflagration-to-detonation transition in spiral channels
NASA Astrophysics Data System (ADS)
Golovastov, S. V.; Mikushkin, A. Yu.; Golub, V. V.
2017-10-01
The deflagration-to-detonation transition in hydrogen-air mixtures that fill spiral channels has been studied. A spiral channel has been produced in a cylindrical detonation tube with a twisted ribbon inside. The gas mixture has been ignited by means of a spark gap switch. The predetonation distance versus the twisted ribbon configuration and molar ratio between the gas mixture components has been determined. A pulling force exerted by the detonation tube after a single event of hydrogen-air mixture burnout has been found for four configurations of the twisted ribbon. Conditions under which the use of a spiral tube can be more effective (increase the pulling force) have been formulated.
Deflagration Rates and Molecular Bonding Trends of Statically Compressed Secondary Explosives
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zaug, J M; Foltz, M F; Hart, E
2010-03-09
We discuss our measurements of the chemical reaction propagation rate as a function of pressure. Materials investigated have included CL-20, HMX, TATB, and RDX crystalline powders, LX-04, Comp B, and nitromethane. The anomalous correspondence between crystal structure, including in some instances isostructural phase transitions, on pressure-dependant RPRs of TATB, HMX, Nitromethane, CL-20, and PETN have been elucidated using micro-IR and -Raman spectroscopies. Here we specifically highlight pressure-dependent physicochemical mechanisms affecting the deflagration rate of nitromethane and epsilon-CL-20. We find that pressure induced splitting of symmetric stretch NO{sub 2} vibrations can signal the onset of increasingly more rapid combustion reactions.
NASA Astrophysics Data System (ADS)
Chambers, Jessica; McGarry, Joseph; Ahmed, Kareem
2015-11-01
Detonation is a high energetic mode of pressure gain combustion. Detonation combustion exploits the pressure rise to augment high flow momentum and thermodynamic cycle efficiencies. The driving mechanism of deflagrated flame acceleration to detonation is turbulence generation and induction. A fluidic jet is an innovative method for the production of turbulence intensities and flame acceleration. Compared to traditional obstacles, the jet reduces the pressure losses and heat soak effects while providing turbulence generation control. The investigation characterizes the turbulent flame-flow interactions. The focus of the study is on classifying the turbulent flame dynamics and the temporal evolution of turbulent flame regime. The turbulent flame-flow interactions are experimentally studied using a LEGO Detonation facility. Advanced high-speed laser diagnostics, particle image velocimetry (PIV), planar laser induced florescence (PLIF), and Schlieren imaging are used in analyzing the physics of the interaction and flame acceleration. Higher turbulence induction is observed within the turbulent flame after contact with the jet, leading to increased flame burning rates. The interaction with the fluidic jet results in turbulent flame transition from the thin reaction zones to the broken reaction regime.
Chemical Energy Release in Several Recently Discovered Detonation and Deflagration Flows
NASA Astrophysics Data System (ADS)
Tarver, Craig M.
2010-10-01
Several recent experiments on complex detonation and deflagration flows are analyzed in terms of the chemical energy release required to sustain these flows. The observed double cellular structures in detonating gaseous nitromethane-oxygen and NO2-fuel (H2, CH4, and C2H6) mixtures are explained by the amplification of two distinct pressure wave frequencies by two exothermic reactions, the faster reaction forming vibrationally excited NO* and the slower reaction forming highly vibrationally excited N2**. The establishment of a Chapman-Jouguet (C-J) deflagration behind a weak shock wave, the C-J detonation established after a head-on collision with a shock front, and the C-J detonation conditions established in reactive supersonic flows are quantitatively calculated using the chemical energy release of a H2 + Cl2 mixture. For these three reactive flows, these calculations illustrate that different fractions of the exothermic chemical energy are used to sustain steady-state propagation. C-J detonation calculations on the various initial states using the CHEETAH chemical equilibrium code are shown to be in good agreement with experimental detonation velocity measurements for the head-on collision and supersonic flow detonations.
Bille, Todd W; Cromartie, Carter; Farr, Matthew
2009-09-01
This study investigated the effects of time, cyanoacrylate fuming, and location of the biological material on DNA analysis of post-blast pipe bomb fragments. Multiple aliquots of a cell suspension (prepared by soaking buccal swabs in water) were deposited on components of the devices prior to assembly. The pipe bombs were then deflagrated and the fragments recovered. Fragments from half of the devices were cyanoacrylate fumed. The cell spots on the fragments were swabbed and polymerase chain reaction/short tandem repeat analysis was performed 1 week and 3 months after deflagration. A significant decrease in the amount of DNA recovered was observed between samples collected and analyzed within 1 week compared with the samples collected and analyzed 3 months after deflagration. Cyanoacrylate fuming did not have a measurable effect on the success of the DNA analysis at either time point. Greater quantities of DNA were recovered from the pipe nipples than the end caps. Undeflagrated controls showed that the majority (>95%) of the DNA deposited on the devices was not recovered at a week or 3 months.
High Pressure Burn Rate Measurements on an Ammonium Perchlorate Propellant
DOE Office of Scientific and Technical Information (OSTI.GOV)
Glascoe, E A; Tan, N
2010-04-21
High pressure deflagration rate measurements of a unique ammonium perchlorate (AP) based propellant are required to design the base burn motor for a Raytheon weapon system. The results of these deflagration rate measurements will be key in assessing safety and performance of the system. In particular, the system may experience transient pressures on the order of 100's of MPa (10's kPSI). Previous studies on similar AP based materials demonstrate that low pressure (e.g. P < 10 MPa or 1500 PSI) burn rates can be quite different than the elevated pressure deflagration rate measurements (see References and HPP results discussed herein),more » hence elevated pressure measurements are necessary in order understand the deflagration behavior under relevant conditions. Previous work on explosives have shown that at 100's of MPa some explosives will transition from a laminar burn mechanism to a convective burn mechanism in a process termed deconsolidative burning. The resulting burn rates that are orders-of-magnitude faster than the laminar burn rates. Materials that transition to the deconsolidative-convective burn mechanism at elevated pressures have been shown to be considerably more violent in confined heating experiments (i.e. cook-off scenarios). The mechanisms of propellant and explosive deflagration are extremely complex and include both chemical, and mechanical processes, hence predicting the behavior and rate of a novel material or formulation is difficult if not impossible. In this work, the AP/HTPB based material, TAL-1503 (B-2049), was burned in a constant volume apparatus in argon up to 300 MPa (ca. 44 kPSI). The burn rate and pressure were measured in-situ and used to calculate a pressure dependent burn rate. In general, the material appears to burn in a laminar fashion at these elevated pressures. The experiment was reproduced multiple times and the burn rate law using the best data is B = (0.6 {+-} 0.1) x P{sup (1.05{+-}0.02)} where B is the burn rate in mm/s and P is the pressure in units of MPa. Details of the experimental method, results and data analysis are discussed herein and briefly compared to other AP based materials that have been measured in this apparatus.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
García-Senz, D.; Cabezón, R. M.; Thielemann, F. K.
Currently the number of models aimed at explaining the phenomena of type Ia supernovae is high and distinguishing between them is a must. In this work we explore the influence of rotation on the evolution of the nuclear flame that drives the explosion in the so-called gravitational confined detonation models. Assuming that the flame starts in a pointlike region slightly above the center of the white dwarf (WD) and adding a moderate amount of angular velocity to the star we follow the evolution of the deflagration using a smoothed particle hydrodynamics code. We find that the results are very dependentmore » on the angle between the rotational axis and the line connecting the initial bubble of burned material with the center of the WD at the moment of ignition. The impact of rotation is larger for angles close to 90° because the Coriolis force on a floating element of fluid is maximum and its principal effect is to break the symmetry of the deflagration. Such symmetry breaking weakens the convergence of the nuclear flame at the antipodes of the initial ignition volume, changing the environmental conditions around the convergence region with respect to non-rotating models. These changes seem to disfavor the emergence of a detonation in the compressed volume at the antipodes and may compromise the viability of the so-called gravitational confined detonation mechanism.« less
Extending atomistic scale chemistry to mesoscale model of condensed-phase deflagration
NASA Astrophysics Data System (ADS)
Joshi, Kaushik; Chaudhuri, Santanu
2017-01-01
Predictive simulations connecting chemistry that follow the shock or thermal initiation of energetic materials to subsequent deflagration or detonation events is currently outside the realm of possibilities. Molecular dynamics and first-principles based dynamics have made progress in understanding reactions in picosecond to nanosecond time scale. Results from thermal ignition of different phases of RDX show a complex reaction network and emergence of a deterministic behavior for critical temperature before ignition and hot spot growth rates. The kinetics observed is dependent on the hot spot temperature, system size and thermal conductivity. For cases where ignition is observed, the incubation period is dominated by intermolecular and intramolecular hydrogen transfer reactions. The gradual temperature and pressure increase in the incubation period is accompanied by accumulation of heavier polyradicals. The challenge of connecting such chemistry in mesoscale simulations remain in reducing the complexity of chemistry. The hot spot growth kinetics in RDX grains and interfaces is an important challenge for reactive simulations aiming to fill in the gaps in our knowledge in the nanoseconds to microseconds time scale. The results discussed indicate that the mesoscale chemistry may include large polyradical molecules in dense reactive mix reaching an instability point at certain temperatures and pressures.
2006-12-01
models attempted to bracket the extremes of the conditions of interest. These conditions were Mach 2 and Mach 3 shocks , with initial medium...later, but all traces have been expanded to the area of interest. Pressure readings were primarily used to measure shock speeds, and initially used...results for the clean tube configuration. The characteristics of the initial shock are similar, and are comparable for all configurations tested
Deflagration rates of secondary explosives under static MPa - GPa pressure
NASA Astrophysics Data System (ADS)
Zaug, Joseph; Young, Christopher; Glascoe, Elizabeth; Maienschein, Jon; Hart, Elaine; Long, Gregory; Black, Collin; Sykora, Gregory; Wardell, Jeffrey
2009-06-01
We discuss our measurements of the chemical reaction propagation rate (RPR) as a function of pressure using diamond anvil cell (DAC) and strand burner technologies. Materials investigated include HMX and RDX crystalline powders, LX-04 (85% HMX and 15% Viton A), and Comp B (63% RDX, 36% TNT, 1% wax). The anomalous correspondence between crystal structure, including in some instances isostructural phase transitions, on pressure dependant RPRs of TATB, HMX, Nitromethane, and Viton are elucidated using micro -IR and -Raman spectroscopies. The contrast between DAC GPa and strand burner MPa regime measurements yields insight into explosive material burn phenomena. Here we highlight pressure dependent physicochemical mechanisms that appear to affect the deflagration rate of precompressed energetic materials.
NASA Astrophysics Data System (ADS)
Underwood, Thomas; Loebner, Keith; Cappelli, Mark
2016-10-01
In this work, the suitability of a pulsed deflagration accelerator to simulate the interaction of edge-localized modes with plasma first wall materials is investigated. Experimental measurements derived from a suite of diagnostics are presented that focus on the both the properties of the plasma jet and the manner in which such jets couple with material interfaces. Detailed measurements of the thermodynamic plasma state variables within the jet are presented using a quadruple Langmuir probe operating in current-saturation mode. This data in conjunction with spectroscopic measurements of H α Stark broadening via a fast-framing, intensified CCD camera provide spatial and temporal measurements of how the plasma density and temperature scale as a function of input energy. Using these measurements, estimates for the energy flux associated with the deflagration accelerator are found to be completely tunable over a range spanning 150 MW m-2 - 30 GW m-2. The plasma-material interface is investigated using tungsten tokens exposed to the plasma plume under variable conditions. Visualizations of resulting shock structures are achieved through Schlieren cinematography and energy transfer dynamics are discussed by presenting temperature measurements of exposed materials. This work is supported by the U.S. Department of Energy Stewardship Science Academic Program in addition to the National Defense Science Engineering Graduate Fellowship.
Propagation of Reactions in Thermally-damaged PBX-9501
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tringe, J W; Glascoe, E A; Kercher, J R
A thermally-initiated explosion in PBX-9501 (octahydro-1,3,5,7-tetranitro-1,3,5,7-tetrazocine) is observed in situ by flash x-ray imaging, and modeled with the LLNL multi-physics arbitrary-Lagrangian-Eulerian code ALE3D. The containment vessel deformation provides a useful estimate of the reaction pressure at the time of the explosion, which we calculate to be in the range 0.8-1.4 GPa. Closely-coupled ALE3D simulations of these experiments, utilizing the multi-phase convective burn model, provide detailed predictions of the reacted mass fraction and deflagration front acceleration. During the preinitiation heating phase of these experiments, the solid HMX portion of the PBX-9501 undergoes a {beta}-phase to {delta}-phase transition which damages the explosivemore » and induces porosity. The multi-phase convective burn model results demonstrate that damaged particle size and pressure are critical for predicting reaction speed and violence. In the model, energetic parameters are taken from LLNL's thermochemical-kinetics code Cheetah and burn rate parameters from Son et al. (2000). Model predictions of an accelerating deflagration front are in qualitative agreement with the experimental images assuming a mode particle diameter in the range 300-400 {micro}m. There is uncertainty in the initial porosity caused by thermal damage of PBX-9501 and, thus, the effective surface area for burning. To better understand these structures, we employ x-ray computed tomography (XRCT) to examine the microstructure of PBX-9501 before and after thermal damage. Although lack of contrast between grains and binder prevents the determination of full grain size distribution in this material, there are many domains visible in thermally damaged PBX-9501 with diameters in the 300-400 {micro}m range.« less
Numerical studies of nonspherical carbon combustion models
NASA Astrophysics Data System (ADS)
Mueller, E.; Arnett, W. D.
1982-10-01
First results of axisymmetric numerical studies of the final evolution of degenerate C + O cores are reported. The two-dimensional convective flow is treated without a phenomenological theory of convection. The computations show that, in the beginning, the nuclear burning propagates slowly outward from the center of the star in a spherical combustion front. Small-scale eddies form, giving rise to bumps in the front. The bumps grow into blobs and eventually into fingers, which steadily elongate relative to the rest of the combustion front. This behavior is not well described by either the detonation or deflagration models, being more complex than either.
Deflagration-to-detonation characteristics of a laser exploding bridge detonator
NASA Astrophysics Data System (ADS)
Welle, E. J.; Fleming, K. J.; Marley, S. K.
2006-08-01
Evaluation of laser initiated explosive trains has been an area of extreme interest due to the safety benefits of these systems relative to traditional electro-explosive devices. A particularly important difference is these devices are inherently less electro-static discharge (ESD) sensitive relative to traditional explosive devices due to the isolation of electrical power and associated materials from the explosive interface. This paper will report work conducted at Sandia National Laboratories' Explosive Components Facility, which evaluated the initiation and deflagration-to-detonation characteristics of a Laser Driven Exploding Bridgewire detonator. This paper will report and discuss characteristics of Laser Exploding Bridgewire devices loaded with hexanitrohexaazaisowurtzitane (CL-20) and tetraammine-cis-bis-(5-nitro-2H-tetrazolato-N2) cobalt (III) perchlorate (BNCP).
Three-dimensional Simulations of Pure Deflagration Models for Thermonuclear Supernovae
NASA Astrophysics Data System (ADS)
Long, Min; Jordan, George C., IV; van Rossum, Daniel R.; Diemer, Benedikt; Graziani, Carlo; Kessler, Richard; Meyer, Bradley; Rich, Paul; Lamb, Don Q.
2014-07-01
We present a systematic study of the pure deflagration model of Type Ia supernovae (SNe Ia) using three-dimensional, high-resolution, full-star hydrodynamical simulations, nucleosynthetic yields calculated using Lagrangian tracer particles, and light curves calculated using radiation transport. We evaluate the simulations by comparing their predicted light curves with many observed SNe Ia using the SALT2 data-driven model and find that the simulations may correspond to under-luminous SNe Iax. We explore the effects of the initial conditions on our results by varying the number of randomly selected ignition points from 63 to 3500, and the radius of the centered sphere they are confined in from 128 to 384 km. We find that the rate of nuclear burning depends on the number of ignition points at early times, the density of ignition points at intermediate times, and the radius of the confining sphere at late times. The results depend primarily on the number of ignition points, but we do not expect this to be the case in general. The simulations with few ignition points release more nuclear energy E nuc, have larger kinetic energies E K, and produce more 56Ni than those with many ignition points, and differ in the distribution of 56Ni, Si, and C/O in the ejecta. For these reasons, the simulations with few ignition points exhibit higher peak B-band absolute magnitudes M B and light curves that rise and decline more quickly; their M B and light curves resemble those of under-luminous SNe Iax, while those for simulations with many ignition points are not.
NASA Astrophysics Data System (ADS)
Tringe, J. W.; Létant, S. E.; Dugan, L. C.; Levie, H. W.; Kuhl, A. L.; Murphy, G. A.; Alves, S. W.; Vandersall, K. S.; Pantoya, M. L.
2013-12-01
Energetic materials are being considered for the neutralization of spore-forming bacteria. In this study, the neutralization effects of a monomolecular explosive were compared to the effects of halogen-containing thermites. Bacillus atrophaeus spores were exposed to the post-detonation environment of a 100 g charge of the military explosive C-4 at a range of 50 cm. These tests were performed in the thermodynamically closed environment of a 506-l barometric calorimeter. Associated temperatures were calculated using a thermodynamic model informed by calculations with the Cheetah thermochemical code. Temperatures in the range of 2300-2800 K were calculated to persist for nearly the full 4 ms pressure observation time. After the detonation event, spores were characterized using optical microscopy and the number of viable spores was assessed. Results showed live spore survival rates in the range of 0.01%-1%. For the thermite tests, a similar, smaller-scale configuration was employed that examined the spore neutralization effects of two thermites: aluminum with iodine pentoxide and aluminum with potassium chlorate. Only the former mixture resulted in spore neutralization. These results indicate that the detonation environment produced by an explosive with no chemical biocides may provide effective spore neutralization similar to a deflagrating thermite containing iodine.
Influence of obstacle disturbance in a duct on explosion characteristics of coal gas
NASA Astrophysics Data System (ADS)
Wang, Cheng; Ma, Tianbao; Lu, Jie
2010-02-01
In combination with experimental research, numerical simulation is performed to investigate the influence law of the obstacles in a duct on the explosion flame of premixed coal gas and air. The numerical method uses upwind WENO scheme and two-step chemical reaction model. The interaction mechanism is addressed between the compression wave from reflection on the right end of the duct and flame propagation. The reflected wave is found to result in the decrease of flame velocity. On this basis, we analyze the mechanism of the obstacles on flame as well as the law of flow field variation thus caused. The results suggest that, due to the obstacles, deflagration wave is repeatedly reflected, combustible gas mixture is fully compressed, temperature and pressure rise, chemical reaction speed increases, and hence flame intensity is strengthened. At the same time, a tripe point forms as a result of wall reflection of the deflagration wave from the obstacles and furthermore local flame speed increases. As the triple point propagates forward, the flame speed gradually decreases due to dissipation of energy. These conclusions provide a valuable theoretical foundation for the prediction of explosion field, prevention of fire and explosion and effective control of the combustion speed and flame propagation speed in detonation propulsion.
Schlieren Cinematography of Current Driven Plasma Jet Dynamics
NASA Astrophysics Data System (ADS)
Loebner, Keith; Underwood, Thomas; Cappelli, Mark
2016-10-01
Schlieren cinematography of a pulsed plasma deflagration jet is presented and analyzed. An ultra-high frame rate CMOS camera coupled to a Z-type laser Schlieren apparatus is used to obtain flow-field refractometry data for the continuous flow Z-pinch formed within the plasma deflagration jet. The 10 MHz frame rate for 256 consecutive frames provides high temporal resolution, enabling turbulent fluctuations and plasma instabilities to be visualized over the course of a single pulse (20 μs). The Schlieren signal is radiometrically calibrated to obtain a two dimensional mapping of the refraction angle of the axisymmetric pinch plasma, and this mapping is then Abel inverted to derive the plasma density distribution as a function radius, axial coordinate, and time. Analyses of previously unknown discharge characteristics and comparisons with prior work are discussed.
Numerical Modeling of Pulse Detonation Rocket Engine Gasdynamics and Performance
NASA Technical Reports Server (NTRS)
Morris, C. I.
2003-01-01
Pulse detonation engines (PDB) have generated considerable research interest in recent years as a chemical propulsion system potentially offering improved performance and reduced complexity compared to conventional gas turbines and rocket engines. The detonative mode of combustion employed by these devices offers a theoretical thermodynamic advantage over the constant-pressure deflagrative combustion mode used in conventional engines. However, the unsteady blowdown process intrinsic to all pulse detonation devices has made realistic estimates of the actual propulsive performance of PDES problematic. The recent review article by Kailasanath highlights some of the progress that has been made in comparing the available experimental measurements with analytical and numerical models.
49 CFR 173.59 - Description of terms for explosives.
Code of Federal Regulations, 2011 CFR
2011-10-01
... deflagration produce inflation, linear or rotary motion; activate diaphragms, valves or switches, or project.... Articles whose functioning depends of physico-chemical reaction of their contents with water. Cord...
Pulsed-Laser, High Speed Photography of Rocket Propellant Surface Deflagration.
1986-05-01
Investigator was Dr Roger J. Becker. AFRPL Project Manager was Mr Gary L. Vogt. This technical report has been reviewed and is approved for publication...8217;YMlB)OI (/P’I I la . i tJ .o C ’ Gary L. Vogt (805) 277-5258 AFPLIDYCR DD FORM 1473,83 APR EDITION OF 1 JAN 73 IS OBSOLETE. Unclass i fied" SECURl iY...84-1236. 4. G. A. Flandro , "A Simple Conceptual Model for the Nonlinear Transient Combustion of a Solid Rocket Propellant," AIAA Paper No. 82-1222
Numerical study of nonequilibrium plasma assisted detonation initiation in detonation tube
NASA Astrophysics Data System (ADS)
Zhou, Siyin; Wang, Fang; Che, Xueke; Nie, Wansheng
2016-12-01
Nonequilibrium plasma has shown great merits in ignition and combustion nowadays, which should be especially useful for hypersonic propulsion. A coaxial electrodes configuration was established to investigate the effect of alternating current (AC) dielectric barrier discharge nonequilibrium plasma on the detonation initiation process in a hydrogen-oxygen mixture. A discharge simulation-combustion simulation loosely coupled method was used to simulate plasma assisted detonation initiation. First, the dielectric barrier discharge in the hydrogen-oxygen mixture driven by an AC voltage was simulated, which takes 17 kinds of particles (including positively charged particles, negatively charged particles, and neutral particles) and 47 reactions into account. The temporal and spatial characteristics of the discharge products were obtained. Then, the discharge products were incorporated into the combustion model of a detonation combustor as the initial conditions for the later detonation initiation simulation. Results showed that the number density distributions of plasma species are different in space and time, and develop highly nonuniformly from high voltage electrode to grounded electrode at certain times. All the active species reach their highest concentration at approximately 0.6T (T denotes a discharge cycle). Compared with the no plasma case, the differences of flowfield shape mainly appear in the early stage of the deflagration to detonation transition process. None of the sub-processes (including the very slow combustion, deflagration, over-driven detonation, detonation decay, and propagation of a self-sustained stable detonation wave) have been removed by the plasma. After the formation of a C-J detonation wave, the whole flowfield remains unchanged. With the help of plasma, the deflagration to detonation transition (DDT) time and distance are reduced by about 11.6% and 12.9%, respectively, which should be attributed to the active particles effect of nonequilibrium plasma and the local turbulent enhancing effect by the spatial characteristics of discharge. In addition, as the duration of forming a shock wave in the combustor is shortened by approximately 8.1%, it can be inferred that the plasma accelerates the DDT process more significantly before the flow becomes supersonic.
Onset of a Propagating Self-Sustained Spin Reversal Front in a Magnetic System
NASA Astrophysics Data System (ADS)
Kent, Andrew D.
2014-03-01
The energy released in a magnetic material by reversing spins as they relax toward equilibrium can lead to a dynamical magnetic instability in which all the spins in a sample rapidly reverse in a run-away process known as magnetic deflagration. A well-defined front separating reversed and un-reversed spins develops that propagates at a constant speed. This process is akin to a chemical reaction in which a flammable substance ignites and the resulting exothermic reaction leads via thermal conduction to increases in the temperature of an adjacent unburned substance that ignites it. In a magnetic system the reaction is the reversal of spins that releases Zeeman energy and the magnetic anisotropy barrier is the reaction's activation energy. An interesting aspect of magnetic systems is that these key energies-the activation energy and the energy released-can be independently controlled by applied magnetic fields enabling systematic studies of these magnetic instabilities. We have studied the instability that leads to the ignition of magnetic deflagration in a thermally driven Mn12-Ac molecular magnet single crystal. Each Mn12-ac molecule is a uniaxial nanomagnet with spin 10 and energy barrier of 60 K. We use a longitudinal field (a field parallel to the easy axis) to set the energy released and a transverse field to control the activation energy. A heat pulse is applied to one end of the crystal to initiate the process. We study the crossover between slow magnetic relaxation and rapid, self-sustained magnetic deflagration as a function of these fields at low temperature (0.5 K). An array of Hall sensors adjacent to a single crystal is used to detect and measure the speed of the spin-reversal front. I will describe a simple model we developed based on a reaction-diffusion process that describes our experimental findings. I will also discuss prospects for observing spin-fronts driven by magnetic dipole interactions between molecules that can be sonic, i.e. travel near the speed of sound (~ 1000 m/s). In collaboration with P. Subedi, S. Velez, F. Macià, S. Li, M. P. Sarachik, J. Tejada, S. Mukherjee and G. Christou. Supported by NSF-DMR-1006575.
On the Initiation Mechanism in Exploding Bridgewire and Laser Detonators
NASA Astrophysics Data System (ADS)
Stewart, D. Scott; Thomas, Keith A.; Clarke, S.; Mallett, H.; Martin, E.; Martinez, M.; Munger, A.; Saenz, Juan
2006-07-01
Since its invention by Los Alamos during the Manhattan Project era the exploding bridgewire detonator (EBW) has seen tremendous use and study. Recent development of a laser-powered device with detonation properties similar to an EBW is reviving interest in the basic physics of the deflagration-to-detonation (DDT) process in both of these devices. Cutback experiments using both laser interferometry and streak camera observations are providing new insight into the initiation mechanism in EBWs. These measurements are being correlated to a DDT model of compaction to detonation and shock to detonation developed previously by Xu and Stewart. The DDT model is incorporated into a high-resolution, multi-material model code for simulating the complete process. Model formulation and the modeling issues required to describe the test data will be discussed.
Stability of cosmological deflagration fronts
NASA Astrophysics Data System (ADS)
Mégevand, Ariel; Membiela, Federico Agustín
2014-05-01
In a cosmological first-order phase transition, bubbles of the stable phase nucleate and expand in the supercooled metastable phase. In many cases, the growth of bubbles reaches a stationary state, with bubble walls propagating as detonations or deflagrations. However, these hydrodynamical solutions may be unstable under corrugation of the interface. Such instability may drastically alter some of the cosmological consequences of the phase transition. Here, we study the hydrodynamical stability of deflagration fronts. We improve upon previous studies by making a more careful and detailed analysis. In particular, we take into account the fact that the equation of motion for the phase interface depends separately on the temperature and fluid velocity on each side of the wall. Fluid variables on each side of the wall are similar for weakly first-order phase transitions, but differ significantly for stronger phase transitions. As a consequence, we find that, for large enough supercooling, any subsonic wall velocity becomes unstable. Moreover, as the velocity approaches the speed of sound, perturbations become unstable on all wavelengths. For smaller supercooling and small wall velocities, our results agree with those of previous works. Essentially, perturbations on large wavelengths are unstable, unless the wall velocity is higher than a critical value. We also find a previously unobserved range of marginally unstable wavelengths. We analyze the dynamical relevance of the instabilities, and we estimate the characteristic time and length scales associated with their growth. We discuss the implications for the electroweak phase transition and its cosmological consequences.
Neutrinos from type Ia supernovae: The deflagration-to-detonation transition scenario
Wright, Warren P.; Nagaraj, Gautam; Kneller, James P.; ...
2016-07-19
It has long been recognized that the neutrinos detected from the next core-collapse supernova in the Galaxy have the potential to reveal important information about the dynamics of the explosion and the nucleosynthesis conditions as well as allowing us to probe the properties of the neutrino itself. The neutrinos emitted from thermonuclear—type Ia—supernovae also possess the same potential, although these supernovae are dimmer neutrino sources. For the first time, we calculate the time, energy, line of sight, and neutrino-flavor-dependent features of the neutrino signal expected from a three-dimensional delayed-detonation explosion simulation, where a deflagration-to-detonation transition triggers the complete disruption ofmore » a near-Chandrasekhar mass carbon-oxygen white dwarf. We also calculate the neutrino flavor evolution along eight lines of sight through the simulation as a function of time and energy using an exact three-flavor transformation code. We identify a characteristic spectral peak at ˜10 MeV as a signature of electron captures on copper. This peak is a potentially distinguishing feature of explosion models since it reflects the nucleosynthesis conditions early in the explosion. We simulate the event rates in the Super-K, Hyper-K, JUNO, and DUNE neutrino detectors with the SNOwGLoBES event rate calculation software and also compute the IceCube signal. Hyper-K will be able to detect neutrinos from our model out to a distance of ˜10 kpc. Here, at 1 kpc, JUNO, Super-K, and DUNE would register a few events while IceCube and Hyper-K would register several tens of events.« less
Nebular Phase Observations of the Type Ia Supernova 2014J in the Near Infrared
NASA Astrophysics Data System (ADS)
Diamond, Tiara
2018-01-01
Late-time spectra of SNe Ia show numerous strong emission features of iron and cobalt throughout the near infrared region. As the spectrum ages, the cobalt features fade as is expected from the decay of 56Co to 56Fe. The strong 1.6440 μm [Fe II] feature is sensitive to the central density of the white dwarf just prior to the runaway because of electron capture in the early stages of burning, hence the line profile width and evolution can be used to probe possible progenitor scenarios. The line profile is dependent on the extent of mixing during any deflagration burning in addition to asymmetries in the distribution of burning products or an off-center ignition. We present observations of SN 2014J from 300–500 days post-explosion. The data are consistent with spherical models of a MCh explosion with a deflagration-to-detonation transition, central density of 0.7×109 g/cm3, and limited mixing. An asymmetry in the line profile of the last spectrum could indicate an off-center ignition or burning products that are not centered on the kinetic center of the explosion. These and other late-time spectroscopic observations in the infrared of a significant sample of SNe Ia will provide insight into the natural variety of these objects, improving our understanding of the underlying physical processes and their usability in cosmology.
Tringe, J. W.; Letant, S. E.; Dugan, L. C.; ...
2013-12-17
We found that energetic materials are being considered for the neutralization of spore-forming bacteria. In this study, the neutralization effects of a monomolecular explosive were compared to the effects of halogen-containing thermites. Bacillus atrophaeus spores were exposed to the post-detonation environment of a 100 g charge of the military explosive C-4 at a range of 50 cm. These tests were performed in the thermodynamically closed environment of a 506-l barometric calorimeter. Associated temperatures were calculated using a thermodynamic model informed by calculations with the Cheetah thermochemicalcode. Temperatures in the range of 2300–2800 K were calculated to persist for nearly themore » full 4 ms pressure observation time. After the detonation event, spores were characterized using optical microscopy and the number of viable spores was assessed. These results showed live spore survival rates in the range of 0.01%–1%. For the thermite tests, a similar, smaller-scale configuration was employed that examined the spore neutralization effects of two thermites: aluminum with iodine pentoxide andaluminum with potassium chlorate. Only the former mixture resulted in spore neutralization. Our results indicate that the detonation environment produced by an explosive with no chemical biocides may provide effective spore neutralization similar to a deflagrating thermite containing iodine.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tringe, J. W.; Létant, S. E.; Dugan, L. C.
2013-12-21
Energetic materials are being considered for the neutralization of spore-forming bacteria. In this study, the neutralization effects of a monomolecular explosive were compared to the effects of halogen-containing thermites. Bacillus atrophaeus spores were exposed to the post-detonation environment of a 100 g charge of the military explosive C-4 at a range of 50 cm. These tests were performed in the thermodynamically closed environment of a 506-l barometric calorimeter. Associated temperatures were calculated using a thermodynamic model informed by calculations with the Cheetah thermochemical code. Temperatures in the range of 2300–2800 K were calculated to persist for nearly the full 4 ms pressure observation time.more » After the detonation event, spores were characterized using optical microscopy and the number of viable spores was assessed. Results showed live spore survival rates in the range of 0.01%–1%. For the thermite tests, a similar, smaller-scale configuration was employed that examined the spore neutralization effects of two thermites: aluminum with iodine pentoxide and aluminum with potassium chlorate. Only the former mixture resulted in spore neutralization. These results indicate that the detonation environment produced by an explosive with no chemical biocides may provide effective spore neutralization similar to a deflagrating thermite containing iodine.« less
NASA Astrophysics Data System (ADS)
Loebner, Keith; Wang, Benjamin; Cappelli, Mark
2014-10-01
The formation and propagation of high velocity plasma jets in a pulsed, coaxial, deflagration-type discharge is examined experimentally. A sensitive, miniaturized, immersed probe array is used to map out magnetic flux density and associated radial current density as a function of time and axial position. This array is also used to probe the magnetic field gradient across the exit of the accelerator and in the jet formation region. Sensitive interferometry via a continuous-wave helium-neon laser source is used to probe the structure of the plasma jet over multiple chords and axial locations. A two dimensional plasma density gradient profile at an instant in time during jet formation is compiled via Shack-Hartmann wavefront sensor analysis. The qualitative characteristics of rarefaction and/or shock wave formation as a function of chamber back-pressure is examined via fast-framing ICCD imaging. These measurements are compared to existing resistive MHD simulations of the coaxial deflagration accelerator and the ensuing rarefaction jet that is expelled from the electrode assembly. The physical mechanisms governing the behavior of the discharge and the formation of these high energy density plasma jets are proposed and validated against both theoretical models and numerically simulated behavior. This research was conducted with Government support under and awarded by DoD, Air Force Office of Scientific Research, National Defense Science and Engineering Graduate (NDSEG) Fellowship, 32 CFR 168a.
NASA Technical Reports Server (NTRS)
Cheng, D. Y.
1971-01-01
Converging, coaxial accelerator electrode configuration operates in vacuum as plasma gun. Plasma forms by periodic injections of high pressure gas that is ionized by electrical discharges. Deflagration mode of discharge provides acceleration, and converging contours of plasma gun provide focusing.
Simulating the Thermal Response of High Explosives on Time Scales of Days to Microseconds
NASA Astrophysics Data System (ADS)
Yoh, Jack J.; McClelland, Matthew A.
2004-07-01
We present an overview of computational techniques for simulating the thermal cookoff of high explosives using a multi-physics hydrodynamics code, ALE3D. Recent improvements to the code have aided our computational capability in modeling the response of energetic materials systems exposed to extreme thermal environments, such as fires. We consider an idealized model process for a confined explosive involving the transition from slow heating to rapid deflagration in which the time scale changes from days to hundreds of microseconds. The heating stage involves thermal expansion and decomposition according to an Arrhenius kinetics model while a pressure-dependent burn model is employed during the explosive phase. We describe and demonstrate the numerical strategies employed to make the transition from slow to fast dynamics.
NASA Astrophysics Data System (ADS)
Vilain, J.
Approaches to major hazard assessment and prediction are reviewed. Source term: (phenomenology/modeling of release, influence on early stages of dispersion); dispersion (atmospheric advection, diffusion and deposition, emphasis on dense/cold gases); combustion (flammable clouds and mists covering flash fires, deflagration, transition to detonation; mostly unconfined/partly confined situations); blast formation, propagation, interaction with structures; catastrophic fires (pool fires, torches and fireballs; highly reactive substances) runaway reactions; features of more general interest; toxic substances, excluding toxicology; and dust explosions (phenomenology and protective measures) are discussed.
A plasma deflagration accelerator as a platform for laboratory astrophysics
NASA Astrophysics Data System (ADS)
Underwood, Thomas C.; Loebner, Keith T. K.; Cappelli, Mark A.
2017-06-01
The replication of astrophysical flows in the laboratory is critical for isolating particular phenomena and dynamics that appear in complex, highly-coupled natural systems. In particular, plasma jets are observed in astrophysical contexts at a variety of scales, typically at high magnetic Reynolds number and driven by internal currents. In this paper, we present detailed measurements of the plasma parameters within deflagration-produced plasma jets, the scaling of these parameters against both machine operating conditions and the corresponding astrophysical phenomena. Using optical and spectroscopic diagnostics, including Schlieren cinematography, we demonstrate the production of current-driven plasma jets of ∼100 km/s and magnetic Reynolds numbers of ∼100, and discuss the dynamics of their acceleration into vacuum. The results of this study will contribute to the reproduction of various types of astrophysical jets in the laboratory and indicate the ability to further probe active research areas such as jet collimation, stability, and interaction.
Measurements of observables during detonator function
NASA Astrophysics Data System (ADS)
Smilowitz, Laura; Henson, Bryan; Remelius, Dennis
Thermal explosion and detonation are two phenomena which can both occur as the response of explosives to thermal or mechanical insults. Thermal explosion is typically considered in the safety envelope and detonation is considered in the performance regime of explosive behavior. However, the two regimes are tied together by a phenomenon called deflagration to detonation transition (DDT). In this talk, I will discuss experiments on commercial detonators aimed at understanding the mechanism for energy release during detonator function. Diagnostic development towards measuring temperature, pressure, and density during the extreme conditions and time scales of detonation will be discussed. Our current ability to perform table-top dynamic radiography on functioning detonators will be described. Dynamic measurements of temperature, pressure, and density will be shown and discussion of the function of a detonator will be given in terms of our current understanding of deflagration, detonation, and the transition between the two.
In-situ measurement of temperature during rapid thermite deflagrations
NASA Astrophysics Data System (ADS)
Densmore, John; Sullivan, Kyle
Thermites are composite materials that consist of a fuel (metal) and oxidizer (metal oxide), that upon reaction can release a large amount of energy (20.8 kJ/cc for Al:CuO). The time scale for a thermite to release energy (ms) is much longer than a typical detonation (us). In-situ temperature and/or thermal flux measurements can provide fundamental insight into the reaction mechanisms. This information can inform the design and optimization of energy transport during a deflagration, to optimize the energy release rate. To measure the temperature we use a burn tube apparatus and various pyrometry techniques to measure the spatial temperature field as a reaction proceeds towards completion. We show that system properties can be adjusted to achieve custom thermal properties. Lawrence Livermore National Laboratory is operated by Lawrence Livermore National Security, LLC, for the U.S. Department of Energy, National Nuclear Security Administration under Contract DE-AC52-07NA27344.
NASA Astrophysics Data System (ADS)
Pal, Pinaki; Mansfield, Andrew B.; Arias, Paul G.; Wooldridge, Margaret S.; Im, Hong G.
2015-09-01
A computational study was conducted to investigate the characteristics of auto-ignition in a syngas mixture at high-pressure and low-temperature conditions in the presence of thermal inhomogeneities. Highly resolved one-dimensional numerical simulations incorporating detailed chemistry and transport were performed. The temperature inhomogeneities were represented by a global sinusoidal temperature profile and a local Gaussian temperature spike (hot spot). Reaction front speed and front Damköhler number analyses were employed to characterise the propagating ignition front. In the presence of a global temperature gradient, the ignition behaviour shifted from spontaneous propagation (strong) to deflagrative (weak), as the initial mean temperature of the reactant mixture was lowered. A predictive Zel'dovich-Sankaran criterion to determine the transition from strong to weak ignition was validated for different parametric sets. At sufficiently low temperatures, the strong ignition regime was recovered due to faster passive scalar dissipation of the imposed thermal fluctuations relative to the reaction timescale, which was quantified by the mixing Damköhler number. In the presence of local hot spots, only deflagrative fronts were observed. However, the fraction of the reactant mixture consumed by the propagating front was found to increase as the initial mean temperature was lowered, thereby leading to more enhanced compression-heating of the end-gas. Passive scalar mixing was not found to be important for the hot spot cases considered. The parametric study confirmed that the relative magnitude of the Sankaran number translates accurately to the quantitative strength of the deflagration front in the overall ignition advancement.
Compression testing of flammable liquids
NASA Technical Reports Server (NTRS)
Briles, O. M.; Hollenbaugh, R. P.
1979-01-01
Small cylindrical test chamber determines catalytic effect of given container material on fuel that might contribute to accidental deflagration or detonation below expected temperature under adiabatic compression. Device is useful to producers and users of flammable liquids and to safety specialists.
The Combustion of HMX. [burning rate at high pressures
NASA Technical Reports Server (NTRS)
Boggs, T. L.; Price, C. F.; Atwood, A. I.; Zurn, D. E.; Eisel, J. L.
1980-01-01
The burn rate of HMX was measured at high pressures (p more than 1000 psi). The self deflagration rate of HMX was determined from 1 atmosphere to 50,000 psi. The burning rate shows no significant slope breaks.
Narin, B; Ozyörük, Y; Ulas, A
2014-05-30
This paper describes a two-dimensional code developed for analyzing two-phase deflagration-to-detonation transition (DDT) phenomenon in granular, energetic, solid, explosive ingredients. The two-dimensional model is constructed in full two-phase, and based on a highly coupled system of partial differential equations involving basic flow conservation equations and some constitutive relations borrowed from some one-dimensional studies that appeared in open literature. The whole system is solved using an optimized high-order accurate, explicit, central-difference scheme with selective-filtering/shock capturing (SF-SC) technique, to augment central-diffencing and prevent excessive dispersion. The sources of the equations describing particle-gas interactions in terms of momentum and energy transfers make the equation system quite stiff, and hence its explicit integration difficult. To ease the difficulties, a time-split approach is used allowing higher time steps. In the paper, the physical model for the sources of the equation system is given for a typical explosive, and several numerical calculations are carried out to assess the developed code. Microscale intergranular and/or intragranular effects including pore collapse, sublimation, pyrolysis, etc. are not taken into account for ignition and growth, and a basic temperature switch is applied in calculations to control ignition in the explosive domain. Results for one-dimensional DDT phenomenon are in good agreement with experimental and computational results available in literature. A typical shaped-charge wave-shaper case study is also performed to test the two-dimensional features of the code and it is observed that results are in good agreement with those of commercial software. Copyright © 2014 Elsevier B.V. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Burstein, S.Z.; Lax, P.D.; Sod, G.A.
1978-09-01
Eleven lectures are presented on mathematical aspects of combustion: fluid dynamics, deflagrations and detonations, chemical kinetics, gas flows, combustion instability, flame spread above solids, spark ignition engines, burning rate of coal particles and hydrocarbon oxidation. Separate abstracts were prepared for three of the lectures. (DLC)
Nucleosynthesis by Type Ia Supernova for different Metallicity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ohkubo, Takuya; Umeda, Hideyuki; Nomoto, Ken'ichi
2006-07-12
We calculate nucleosynthesis by type Ia supernova for various metallicity. We adopt two typical hydrodynamical models, carbon deflagration and delayed detonation. The two main points of this research are to see that (1)how the ejected mass of 56Ni changes and (2)how abundance of each element (especially Fe-group elements) is influenced by varying metallicity. We find that (1)56Ni mass changes about 15% in the range of Z = 0.001 - 0.05 and insufficient to explain all of the observed variety of SNe Ia peak luminosity, and (2)[Mn/Fe] and [Ni/Fe] show fairy dependence on metallicity (especially for delayed detonation model) while [Cr/Fe]more » or [{alpha}/Fe] do not.« less
NASA Technical Reports Server (NTRS)
Osipov, Viatcheslav; Muratov, Cyrill; Hafiychuk, Halyna; Ponizovskya-Devine, Ekaterina; Smelyanskiy, Vadim; Mathias, Donovan; Lawrence, Scott; Werkheiser, Mary
2011-01-01
We analyze the data of purposeful rupture experiments with LOx and LH2 tanks, the Hydrogen-Oxygen Vertical Impact (HOVI) tests that were performed to clarify the ignition mechanisms, the explosive power of cryogenic H2/Ox mixtures under different conditions, and to elucidate the puzzling source of the initial formation of flames near the intertank section during the Challenger disaster. We carry out a physics-based analysis of general explosions scenarios for cryogenic gaseous H2/Ox mixtures and determine their realizability conditions, using the well-established simplified models from the detonation and deflagration theory. We study the features of aerosol H2/Ox mixture combustion and show, in particular, that aerosols intensify the deflagration flames and can induce detonation for any ignition mechanism. We propose a cavitation-induced mechanism of self-ignition of cryogenic H2/Ox mixtures that may be realized when gaseous H2 and Ox flows are mixed with a liquid Ox turbulent stream, as occurred in all HOVI tests. We present an overview of the HOVI tests to make conclusion on the risk of strong explosions in possible liquid rocket incidents and provide a semi-quantitative interpretation of the HOVI data based on aerosol combustion. We uncover the most dangerous situations and discuss the foreseeable risks which can arise in space missions and lead to tragic outcomes. Our analysis relates to only unconfined mixtures that are likely to arise as a result of liquid propellant space vehicle incidents.
NASA Astrophysics Data System (ADS)
Zhao, Tiejun; Li, Xiaojie; Lee, John H. S.; Yan, Honghao
2018-02-01
Using ferrocene, H2 and O2, Carbon nanomaterials were prepared with gaseous detonation (deflagration) method. The effects of H2 on the phase and morphology of carbon nanomaterials were studied by various proportions of H2 in the reaction. The prepared samples were characterized by x-ray diffractometer, transmission electron microscope and Raman spectrometer. The results show that hydrogen proportion has a great influence on the phase and morphology of carbon nanomaterials. The high hydrogen proportion leads to much unreacted hydrogen, which could protect the iron atom from oxidation of carbon and dilute the reactants contributing to uniform particle size. In addition, the graphitization degree of multi-walled carbon nanotubes, observed in samples with high H2 proportion, is high enough to see the lattice fringes, but the degree of graphitization of whole sample is lower than which fabricated with low H2 proportion, and it may result from the low energy generation.
Detonation engine fed by acetylene-oxygen mixture
NASA Astrophysics Data System (ADS)
Smirnov, N. N.; Betelin, V. B.; Nikitin, V. F.; Phylippov, Yu. G.; Koo, Jaye
2014-11-01
The advantages of a constant volume combustion cycle as compared to constant pressure combustion in terms of thermodynamic efficiency has focused the search for advanced propulsion on detonation engines. Detonation of acetylene mixed with oxygen in various proportions is studied using mathematical modeling. Simplified kinetics of acetylene burning includes 11 reactions with 9 components. Deflagration to detonation transition (DDT) is obtained in a cylindrical tube with a section of obstacles modeling a Shchelkin spiral; the DDT takes place in this section for a wide range of initial mixture compositions. A modified ka-omega turbulence model is used to simulate flame acceleration in the Shchelkin spiral section of the system. The results of numerical simulations were compared with experiments, which had been performed in the same size detonation chamber and turbulent spiral ring section, and with theoretical data on the Chapman-Jouguet detonation parameters.
Deflagration to Detonation Transition (DDT) Simulations of HMX Powder Using the HERMES Model
NASA Astrophysics Data System (ADS)
White, Bradley; Reaugh, John; Tringe, Joseph
2017-06-01
We performed computer simulations of DDT experiments with Class I HMX powder using the HERMES model (High Explosive Response to MEchanical Stimulus) in ALE3D. Parameters for the model were fitted to the limited available mechanical property data of the low-density powder, and to the Shock to Detonation Transition (SDT) test results. The DDT tests were carried out in steel-capped polycarbonate tubes. This arrangement permits direct observation of the event using both flash X-ray radiography and high speed camera imaging, and provides a stringent test of the model. We found the calculated detonation transition to be qualitatively similar to experiment. Through simulation we also explored the effects of confinement strength, the HMX particle size distribution and porosity on the computed detonation transition location. This work was performed under the auspices of the US DOE by LLNL under Contract DE-AC52-07NA27344.
Biocidal Energetic Materials for the Destruction of Spore Forming Bacteria
2015-07-01
Bacteria Spore Gas Antibacterial Thermal Unclassified Unclassified Unclassified SAR 47 Suhithi Peiris...naturally antibacterial and biocidal properties using combustion synthesis of mildly energetic reactants; and, (2) engineering an aerosolized spray...of biocidal gases using unique a deflagration synthesis approach. Accomplishments for all years: Major Activity 1: Creating highly porous
A numerical investigation of premixed combustion in wave rotors
NASA Technical Reports Server (NTRS)
Nalim, M. Razi; Paxson, Daniel E.
1996-01-01
Wave rotor cycles which utilize premixed combustion processes within the passages are examined numerically using a one-dimensional CFD-based simulation. Internal-combustion wave rotors are envisioned for use as pressure-gain combustors in gas turbine engines. The simulation methodology is described, including a presentation of the assumed governing equations for the flow and reaction in the channels, the numerical integration method used, and the modeling of external components such as recirculation ducts. A number of cycle simulations are then presented which illustrate both turbulent-deflagration and detonation modes of combustion. Estimates of performance and rotor wall temperatures for the various cycles are made, and the advantages and disadvantages of each are discussed.
Modeling the Risk of Fire/Explosion Due to Oxidizer/Fuel Leaks in the Ares I Interstage
NASA Technical Reports Server (NTRS)
Ring, Robert W.; Stott, James E.; Hales, Christy
2008-01-01
A significant flight hazard associated with liquid propellants, such as those used in the upper stage of NASA's new Ares I launch vehicle, is the possibility of leakage of hazardous fluids resulting in a catastrophic fire/explosion. The enclosed and vented interstage of the Ares I contains numerous oxidizer and fuel supply lines as well as ignition sources. The potential for fire/explosion due to leaks during ascent depends on the relative concentrations of hazardous and inert fluids within the interstage along with other variables such as pressure, temperature, leak rates, and fluid outgasing rates. This analysis improves on previous NASA Probabilistic Risk Assessment (PRA) estimates of the probability of deflagration, in which many of the variables pertinent to the problem were not explicitly modeled as a function of time. This paper presents the modeling methodology developed to analyze these risks.
Techniques for probing the effects of three-dimensional magneto-hydrodynamics in type Ia supernovae
NASA Astrophysics Data System (ADS)
Penney, Robert B.
Understanding Type Ia Supernovae provides an opportunity to study a wide range of modern physics as well as develop a key tool in cosmology. Here, we identify and investigate new observational signatures of underlying physical processes related to the thermonuclear runaway and the flame propagation and its environment, including three-dimensional effects. Understanding these signatures requires a consistent picture of the nuclear decay processes that power the Supernova and the energy transport of the high-energy particles created by these decays. Therefore we have developed and employed methods for simulating the transport of X-rays, gamma-rays, positrons and of low energy photons through a relativistically expanding envelope. With this, we identify signatures in the light curves, spectra and, in particular, line profiles that are produced in the context of a series of explosion models for Chandrasekhar Mass White Dwarfs. These models use the framework of a delayed detonation scenario in which a deflagration front transitions to a detonation phase. We use models that allow us to separate physical effects due to the flame physics and instabilities, the deflagration to detonation transitions and the initial conditions. Results from within this framework can be used directly for a variety of scenarios for Type Ias including merging White Dwarfs and explosions of sub-Chandrasekhar mass White Dwarfs. We found and developed X-rays as a tool to probe the outer layers of the Supernova envelope. Only models with radioactive material will show significant X-ray line fluxes during the first few months. We show that observations of these can put stringent limits on flame instabilities in delayed-detonation models, and can confirm or rule out mergers and sub-Chandrasekhar mass White Dwarfs as progenitors. Though these observations have not yet been done, the current generation of satellites are capable of observations of these spectra within about 10Mpc. As has been shown previously, optical lightcurves provide a possible diagnostic for magnetic fields. We confirm previous results that lightcurve changes due to these fields on the order of 0.1 mag are to be expected at 200--300 days, and changes may be up to 1 mag after 2--3 years. We show that the time evolution of late-time IR line profiles can probe the magentic fields and density and chemical distributions of the Supernova. To separate magnetic field from geometrical effects, we show the need for a time series of observations starting some 3--4 months after maximum light. Up until now, current observational programs commonly take snapshots at one late-time only though, technically, early time spectra are well within reach. Although rare, those observations that have been done show signatures of an off-center deflagration-detonation transition and, at least in some cases, our results strongly suggest the presence of magnetic fields well in excess of 106G. Fields of this size can be expected to alter the evolution of the nuclear runaway which determines the initial conditions which rule the flame propagation during the explosion. With the upcoming generation of large telescopes on ground and in space, observations can be obtained routinely at distances well beyond the local group of galaxies.
NASA Astrophysics Data System (ADS)
Boeck, Lorenz R.; Katzy, Peter; Hasslberger, Josef; Kink, Andreas; Sattelmayer, Thomas
2016-09-01
An open-access online platform containing data from experiments on deflagration-to-detonation transition conducted at the Institute of Thermodynamics, Technical University of Munich, has been developed and is accessible at http://www.td.mw.tum.de/ddt. The database provides researchers working on explosion dynamics with data for theoretical analyses and for the validation of numerical simulations.
In-Situ Imaging of Particles during Rapid Thermite Deflagrations
NASA Astrophysics Data System (ADS)
Grapes, Michael; Sullivan, Kyle; Reeves, Robert; Densmore, John; Willey, Trevor; van Buuren, Tony; Fezaa, Kamel
The dynamic behavior of rapidly deflagrating thermites is a highly complex process involving rapid decomposition, melting, and outgassing of intermediate and/or product gases. Few experimental techniques are capable of probing these phenomena in situ due to the small length and time scales associated with the reaction. Here we use a recently developed extended burn tube test, where we initiate a small pile of thermite on the closed end of a clear acrylic tube. The length of the tube is sufficient to fully contain the reaction as it proceeds and flows entrained particles down the tube. This experiment was brought to the Advanced Photon Source, and the particle formation was X-ray imaged at various positions down the tube. Several formulations, as well as formulation parameters were varied to investigate the size and morphology of the particles, as well as to look for dynamic behavior attributed to the reaction. In all cases, we see evidence of particle coalescence and condensed-phase interfacial reactions. The results improve our understanding of the procession of reactants to products in these systems. Funding provided by the LLNL LDRD program (PLS-16FS-028).
Deflagration-to-detonation transition in granular HMX
NASA Technical Reports Server (NTRS)
Campbell, A. W.
1980-01-01
Granular HMX of three degrees of fineness was packed into heavy-walled steel tubes closed at both ends. Ignition was obtained at one end using an intimate mixture of finely divided titanium and boron as an igniter that produced heat with little gas. The distance to detonation was determined by examination of the resulting tube fragments. By inserting tightly-fitted neoprene diaphragms periodically into the HMX column, it was shown that the role of convective combustion was limited to the initial stage of the deflagration to detonation (DDT) process. Experiments in which various combinations of two of the three types of HMX were loaded into the same tube showed that heating by adiabatic shear of explosive grains was an essential factor in the final buildup to detonation. A description of the DDT process is developed in which conductive burning is followed in turn by convective burning, bed collapse with plug formation, onset of accelerated burning at the front of the plug through heating by intercrystalline friction and adiabatic shear, and intense shock formation resulting in high-order detonation.
NASA Astrophysics Data System (ADS)
Mernier, F.; de Plaa, J.; Pinto, C.; Kaastra, J. S.; Kosec, P.; Zhang, Y.-Y.; Mao, J.; Werner, N.; Pols, O. R.; Vink, J.
2016-11-01
The hot intra-cluster medium (ICM) is rich in metals, which are synthesised by supernovae (SNe) and accumulate over time into the deep gravitational potential well of clusters of galaxies. Since most of the elements visible in X-rays are formed by type Ia (SNIa) and/or core-collapse (SNcc) supernovae, measuring their abundances gives us direct information on the nucleosynthesis products of billions of SNe since the epoch of the star formation peak (z 2-3). In this study, we compare the most accurate average X/Fe abundance ratios (compiled in a previous work from XMM-Newton EPIC and RGS observations of 44 galaxy clusters, groups, and ellipticals), representative of the chemical enrichment in the nearby ICM, to various SNIa and SNcc nucleosynthesis models found in the literature. The use of a SNcc model combined to any favoured standard SNIa model (deflagration or delayed-detonation) fails to reproduce our abundance pattern. In particular, the Ca/Fe and Ni/Fe ratios are significantly underestimated by the models. We show that the Ca/Fe ratio can be reproduced better, either by taking a SNIa delayed-detonation model that matches the observations of the Tycho supernova remnant, or by adding a contribution from the "Ca-rich gap transient" SNe, whose material should easily mix into the hot ICM. On the other hand, the Ni/Fe ratio can be reproduced better by assuming that both deflagration and delayed-detonation SNIa contribute in similar proportions to the ICM enrichment. In either case, the fraction of SNIa over the total number of SNe (SNIa+SNcc) contributing to the ICM enrichment ranges within 29-45%. This fraction is found to be systematically higher than the corresponding SNIa/(SNIa+SNcc) fraction contributing to the enrichment of the proto-solar environnement (15-25%). We also discuss and quantify two useful constraints on both SNIa (I.e. the initial metallicity on SNIa progenitors and the fraction of low-mass stars that result in SNIa) and SNcc (I.e. the effect of the IMF and the possible contribution of pair-instability SNe to the enrichment) that can be inferred from the ICM abundance ratios. Finally, we show that detonative sub-Chandrasekhar WD explosions (resulting, for example, from violent WD mergers) cannot be a dominant channel for SNIa progenitors in galaxy clusters.
Numerical approaches to combustion modeling. Progress in Astronautics and Aeronautics. Vol. 135
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oran, E.S.; Boris, J.P.
1991-01-01
Various papers on numerical approaches to combustion modeling are presented. The topics addressed include; ab initio quantum chemistry for combustion; rate coefficient calculations for combustion modeling; numerical modeling of combustion of complex hydrocarbons; combustion kinetics and sensitivity analysis computations; reduction of chemical reaction models; length scales in laminar and turbulent flames; numerical modeling of laminar diffusion flames; laminar flames in premixed gases; spectral simulations of turbulent reacting flows; vortex simulation of reacting shear flow; combustion modeling using PDF methods. Also considered are: supersonic reacting internal flow fields; studies of detonation initiation, propagation, and quenching; numerical modeling of heterogeneous detonations, deflagration-to-detonationmore » transition to reactive granular materials; toward a microscopic theory of detonations in energetic crystals; overview of spray modeling; liquid drop behavior in dense and dilute clusters; spray combustion in idealized configurations: parallel drop streams; comparisons of deterministic and stochastic computations of drop collisions in dense sprays; ignition and flame spread across solid fuels; numerical study of pulse combustor dynamics; mathematical modeling of enclosure fires; nuclear systems.« less
Design and optimization of a deflagration to detonation transition (ddt) section
NASA Astrophysics Data System (ADS)
Romo, Francisco X.
Throughout the previous century, hydrocarbon-fueled engines have used and optimized the `traditional' combustion process called deflagration (subsonic combustion). An alternative form of combustion, detonation (supersonic combustion), can increase the thermal efficiency of the process by anywhere from 20 - 50%. Even though several authors have studied detonation waves since the 1890's and a plethora of papers and books have been published, it was not until 2008 that the first detonation-powered flight took place. It lasted for 10 seconds at 100 ft. altitude. Achieving detonation presents its own challenges: some fuels are not prone to detonate, severe vibrations caused by the cyclic nature of the engine and its intense noise are some of the key areas that need further research. Also, to directly achieve detonation either a high-energy, bulky, ignition system is required, or the combustion chamber must be fairly long (5 ft. or more in some cases). In the latter method, a subsonic flame front accelerates within the combustion chamber until it reaches supersonic speeds, thus detonation is attained. This is called deflagration-todetonation transition (DDT). Previous papers and experiments have shown that obstacles, such as discs with an orifice, located inside the combustion chamber can shorten the distance required to achieve detonation. This paper describes a hands-on implementation of a DDT device. Different disc geometries inside the chamber alter the wave characteristics at the exit of the tube. Although detonation was reached only when using pure oxygen, testing identified an obstacle configuration for LPG and air mixtures that increased pressure and wave speed significantly when compared to baseline or other obstacle configurations. Mixtures of LPG and air were accelerated to Mach 0.96 in the downstream frame of reference, which would indicate a transition to detonation was close. Reasons for not achieving detonation may include poor fuel and oxidizer mixing, and/or the need for a longer DDT section.
NASA Technical Reports Server (NTRS)
Allgood, Daniel C.
2016-01-01
The objective of the presented work was to develop validated computational fluid dynamics (CFD) based methodologies for predicting propellant detonations and their associated blast environments. Applications of interest were scenarios relevant to rocket propulsion test and launch facilities. All model development was conducted within the framework of the Loci/CHEM CFD tool due to its reliability and robustness in predicting high-speed combusting flow-fields associated with rocket engines and plumes. During the course of the project, verification and validation studies were completed for hydrogen-fueled detonation phenomena such as shock-induced combustion, confined detonation waves, vapor cloud explosions, and deflagration-to-detonation transition (DDT) processes. The DDT validation cases included predicting flame acceleration mechanisms associated with turbulent flame-jets and flow-obstacles. Excellent comparison between test data and model predictions were observed. The proposed CFD methodology was then successfully applied to model a detonation event that occurred during liquid oxygen/gaseous hydrogen rocket diffuser testing at NASA Stennis Space Center.
Pulsating Instability of Turbulent Thermonuclear Flames in Type Ia Supernovae
NASA Astrophysics Data System (ADS)
Poludnenko, Alexei Y.
2014-01-01
Presently, one of the main explosion scenarios of type Ia supernovae (SNIa), aimed at explaining both "normal" and subluminous events, is the thermonuclear incineration of a white-dwarf in a single-degenerate system. The underlying engine of such explosions is the turbulent thermonuclear flame. Modern, large-scale, multidimensional simulations of SNIa cannot resolve the internal flame structure, and instead must include a subgrid-scale prescription for the turbulent-flame properties. As a result, development of robust, parameter-free, large-scale models of SNIa crucially relies on the detailed understanding of the turbulent flame properties during each stage of the flame evolution. Due to the complexity of the flame dynamics, such understanding must be validated by the first-principles direct numerical simulations (DNS). In our previous work, we showed that sufficiently fast turbulent flames are inherently susceptible to the development of detonations, which may provide the mechanism for the deflagration-to-detonation transition (DDT) in the delayed-detonation model of SNIa. Here we extend this study by performing detailed analysis of the turbulent flame properties at turbulent intensities below the critical threshold for DDT. We carried out a suite of 3D DNS of turbulent flames for a broad range of turbulent intensities and system sizes using a simplified, single-step, Arrhenius-type reaction kinetics. Our results show that at the later stages of the explosion, as the turbulence intensity increases prior to the possible onset of DDT, the flame front will become violently unstable. We find that the burning rate exhibits periodic pulsations with the energy release rate varying by almost an order of magnitude. Furthermore, such flame pulsations can produce pressure waves and shocks as the flame speed approaches the critical Chapman-Jouguet deflagration speed. Finally, in contrast with the current theoretical understanding, such fast turbulent flames can propagate at speeds, which are much higher than the characteristic speeds of turbulent fluctuations. These effects can qualitatively change the dynamics of the explosion and, therefore, must be properly accounted for in the turbulent-flame subgrid-scale models.
Numerical investigation of spontaneous flame propagation under RCCI conditions
Bhagatwala, Ankit V; Sankaran, Ramanan; Kokjohn, Sage; ...
2015-06-30
This paper presents results from one and two-dimensional direct numerical simulations under Reactivity Controlled Compression Ignition (RCCI) conditions of a primary reference fuel (PRF) mixture consisting of n-heptane and iso-octane. RCCI uses in-cylinder blending of two fuels with different autoignition characteristics to control combustion phasing and the rate of heat release. These simulations employ an improved model of compression heating through mass source/sink terms developed in a previous work by Bhagatwala et al. (2014), which incorporates feedback from the flow to follow a predetermined experimental pressure trace. Two-dimensional simulations explored parametric variations with respect to temperature stratification, pressure profiles andmore » n-heptane concentration. Furthermore, statistics derived from analysis of diffusion/reaction balances locally normal to the flame surface were used to elucidate combustion characteristics for the different cases. Both deflagration and spontaneous ignition fronts were observed to co-exist, however it was found that higher n-heptane concentration provided a greater degree of flame propagation, whereas lower n-heptane concentration (higher fraction of iso-octane) resulted in more spontaneous ignition fronts. A significant finding was that simulations initialized with a uniform initial temperature and a stratified n-heptane concentration field, resulted in a large fraction of combustion occurring through flame propagation. The proportion of spontaneous ignition fronts increased at higher pressures due to shorter ignition delay when other factors were held constant. For the same pressure and fuel concentration, the contribution of flame propagation to the overall combustion was found to depend on the level of thermal stratification, with higher initial temperature gradients resulting in more deflagration and lower gradients generating more ignition fronts. Statistics of ignition delay are computed to assess the Zel’dovich (1980) theory for the mode of combustion propagation based on ignition delay gradients.« less
Ultrashort-pulse laser generated nanoparticles of energetic materials
Welle, Eric J [Niceville, NM; Tappan, Alexander S [Albuquerque, NM; Palmer, Jeremy A [Albuquerque, NM
2010-08-03
A process for generating nanoscale particles of energetic materials, such as explosive materials, using ultrashort-pulse laser irradiation. The use of ultrashort laser pulses in embodiments of this invention enables one to generate particles by laser ablation that retain the chemical identity of the starting material while avoiding ignition, deflagration, and detonation of the explosive material.
Supernovae, supernebulae, and nucleosynthesis
NASA Astrophysics Data System (ADS)
Wheeler, J. Craig; Harkness, Robert P.; Barkat, Zalman; Swartz, Douglas
1986-10-01
Supernova atmosphere calculations continue to show that variants of previously calculated carbon-deflagration models provide a good representation of the maximum light spectra of classical type Ia supernovae including the ultraviolet deficit. Careful consideration of the conditions leading to central thermonuclear runaway of degenerate carbon shows that runaway can, however, lead to detonation and direct conflict with observations. As witnessed by the spectra of type Ib supernovae, massive stars are expected to be the primary source of oxygen. Estimates of the absolute production of oxygen in massive stars suggest that if all stars more massive than ≡12 M_sun; explode as supernovae, oxygen would be overproduced in the solar neighborhood, an effect exacerbated by the recent increase in the reaction rate for 12C(α, γ)16O.
Explosively Generated Plasmas: Measurement and Models of Shock Generation and Material Interactions
NASA Astrophysics Data System (ADS)
Emery, Samuel; Elert, Mark; Giannuzzi, Paul; Le, Ryan; McCarthy, Daniel; Schweigert, Igor
2017-06-01
Explosively generated plasmas (EGPs) are created by the focusing of a shock produced from an explosive driver via a conical waveguide. In the waveguide, the gases from the explosive along with the trapped air are accelerated and compressed (via Mach stemming) to such extent that plasma is produced. These EGPs have been measured in controlled experiments to achieve temperatures on the order of 1 eV and velocities as high as 25 km/s. We have conducted a combined modeling and measurement effort to increase the understanding for design purposes of the shock generation of EGPs and the interaction of EGP with explosive materials. Such efforts have led to improved measures of pressure and temperature, spatial structure of the plasma, and the decomposition/deflagration behavior of RDX upon exposure to an EGP. Funding provided by the Environmental Security Technology Certification Program (ESTCP) Munitions Response program area.
Numerical investigation of combustion phenomena in pulse detonation engine with different fuels
NASA Astrophysics Data System (ADS)
Alam, Noor; Sharma, K. K.; Pandey, K. M.
2018-05-01
The effects of different fuel-air mixture on the cyclic operation of pulse detonation engine (PDE) are numerically investigated. The present simulation is to be consider 1200 mm long straight tube combustor channel and 60 mm internal diameter, and filled with stoichiometric ethane-air and ethylene-air (C2H6-air & C2H4) fuel mixture at atmospheric pressure and temperature of 0.1 MPa and 300 K respectively. The obstacles of blockage ratio (BR) 0.5 and having 60 mm spacing among them are allocated inside the combustor tube. There are realizable k-ɛ turbulence model used to analyze characteristic of combustion flame. The objective of present simulation is to analyze the variation in combustion mechanism for two different fuels with one-step reduced chemical reaction model. The obstacles were creating perturbation inside the PDE tube. Therefore, flame surface area increases and reduces deflagration-to-detonation transition (DDT) run-up length.
Random Fields and Collective Effects in Molecular Magnets
2018-01-29
longitudinal fields the final state consists of only partially reversed spins. Further, we measured the front speed as a function of applied magnetic...field. The theory of magnetic deflagration, together with a modification that takes into account the partial spin reversal, fits the transverse field...Conference Paper or Presentation Conference Name: APS March Meeting 2016 Conference Location: Baltimore, Paper Title: Time-resolved Measurements
Spark Ignition of Combustible Vapor in a Plastic Bottle as a Demonstration of Rocket Propulsion
ERIC Educational Resources Information Center
Mattox, J. R.
2017-01-01
I report an innovation that provides a compelling demonstration of rocket propulsion, appropriate for students of physics and other physical sciences. An electrical spark is initiated from a distance to cause the deflagration of a combustible vapor mixed with air in a lightweight plastic bottle that is consequently propelled as a rocket by the…
Controlling material reactivity using architecture
NASA Astrophysics Data System (ADS)
Sullivan, Kyle
2017-06-01
The reactivity of thermites can be tailored through selection of several parameters, and can range from very slow burns to rapid deflagrations. 3D printing is a rapidly emerging field, and offers the potential to build architected parts. Here we sought to explore whether controlling such features could be a suitable path forward for gaining additional control of the reactivity. This talk discusses several new methods for preparing thermite samples with controlled architectures using 3D printing. Additionally, we demonstrate that the architecture can play a role in the reactivity of an object. Our results suggest that architecture can be used to tailor the convective and/or advective energy transport during a deflagration, thus enhancing or retarding the reaction. The results are promising in that they give researchers an additional way of controlling the energy release rate without defaulting to the conventional approach of changing the formulation. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under contract DE-AC52-07NA27344. LLNL-ABS-708525. In collaboration with: Cheng Zhu, Eric Duoss, Matt Durban, Alex Gash, Alexandra Golobic, Michael Grapes, David Kolesky, Joshua Kuntz, Jennifer Lewis, Christopher Spadaccini; LAWRENCE LIVERMORE NATIONAL LAB.
Controlling material reactivity using architecture
NASA Astrophysics Data System (ADS)
Sullivan, Kyle; Zhu, Cheng; Duoss, Eric; Durban, Matt; Gash, Alex; Golobic, Alexandra; Grapes, Michael; Kuntz, Joshua; Spadaccini, Christopher; Kolesky, David; Lewis, Jennifer; LLNL Team; Harvard University Team
Thermites are mixtures of a metal fuel with a metal oxide as the oxidizer. The reactivity of such materials can be tailored through careful selection of a variety of parameters, and can range from very slow burns to rapid deflagrations when using nanoparticles. However, in some cases diminishing returns have been observed as the particle size is reduced. 3D printing is a rapidly emerging field, which offers the capability of printing architected parts; for example parts with controlled internal feature sizes and geometries. In this work, we investigated whether such features could be utilized to gain additional control of the reactivity. This talk introduces several new methods for preparing thermite samples with controlled architectures using direct 3D printing, deposition, and/or casting. Additionally, we demonstrate that 3D printing can be used to tailor the convective and/or advective energy transport during a deflagration, thus enhancing or retarding the reaction. The results are promising in that they give researchers additional ways to control the energy release rate, without defaulting to the classic approach of changing the formulation. This work was performed under the auspices of the U. S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344. LLNL-ABS-708525.
Measurement of the flow properties within a copper tube containing a deflagrating explosive
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hill, Larry G; Morris, John S; Jackson, Scott I
2009-01-01
We report on the propagation of deflagration waves in the high explosive (HE) PBX 9501 (95 wt % HMX, 5 wt% binder). Our test configuration, which we call the def1agration cylinder test (DFCT), is fashioned after the detonation cylinder test (DTCT) that is used to calibrate the JWL detonation product equation of state (EOS). In the DFCT, the HE is heated to a uniform slightly subcritical temperature, and is ignited at one end by a hot wire. For some configurations and initial conditions, we observe a quasi-steady wave that flares the tube into a funnel shape, stretching it to themore » point of rupture. This behavior is qualitatively like the DTCT, such that, by invoking certain additional approximations that we discuss, its behavior can be analyzed by the same methods. We employ an analysis proposed by G.I. Taylor to infer the pressure-volume curve for the burning, expanding flow. By comparing this result to the EOS of HMX product gas alone. we infer that only {approx}20 wt% of the HMX has burned at tube rupture. This result confirms pre-existing observations about the role of convective burning in HMX cookoff explosions.« less
NASA Astrophysics Data System (ADS)
Mays, Owen; Tringe, Joe; Souers, Clark; Lauderbach, Lisa; Baluyot, Emer; Converse, Mark; Kane, Ron
2017-06-01
Microwave interferometry (MI) presents several advantages over more traditional existing shock and deflagration front diagnostics. Most importantly, it directly interrogates these fronts, instead of measuring the evolution of containment surfaces or explosive edges. Here we present the results of MI measurements on detonator-initiated cylinder tests, as well as on deflagration-to-detonation transition experiments, with emphasis on optimization of signal strength through coupling devices and through microwave-transparent windows. Full-wave electromagnetic field finite element simulations were employed to better understand microwave coupling into porous and near full theoretical maximum density (TMD) explosives. HMX and TATB-based explosives were investigated. Data was collected simultaneously at 26.5 GHz and 39 GHz, allowing for direct comparison of the front characteristics and providing insight into the dielectric properties of explosives at these high frequencies. MI measurements are compared against detonation velocity results from photonic Doppler velocimetry probes and high speed cameras, demonstrating the accuracy of the MI technique. Our results illustrate features of front propagation behavior that are difficult to observe with other techniques. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Reacting Chemistry Based Burn Model for Explosive Hydrocodes
NASA Astrophysics Data System (ADS)
Schwaab, Matthew; Greendyke, Robert; Steward, Bryan
2017-06-01
Currently, in hydrocodes designed to simulate explosive material undergoing shock-induced ignition, the state of the art is to use one of numerous reaction burn rate models. These burn models are designed to estimate the bulk chemical reaction rate. Unfortunately, these models are largely based on empirical data and must be recalibrated for every new material being simulated. We propose that the use of an equilibrium Arrhenius rate reacting chemistry model in place of these empirically derived burn models will improve the accuracy for these computational codes. Such models have been successfully used in codes simulating the flow physics around hypersonic vehicles. A reacting chemistry model of this form was developed for the cyclic nitramine RDX by the Naval Research Laboratory (NRL). Initial implementation of this chemistry based burn model has been conducted on the Air Force Research Laboratory's MPEXS multi-phase continuum hydrocode. In its present form, the burn rate is based on the destruction rate of RDX from NRL's chemistry model. Early results using the chemistry based burn model show promise in capturing deflagration to detonation features more accurately in continuum hydrocodes than previously achieved using empirically derived burn models.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Reaugh, J E; Lee, E L
2002-07-01
Mixture rules for partially reacted explosives differ amongst various models. For instance, JWL++ uses a partial pressure addition to compute an average zonal pressure, Ignition and Growth requires pressure equilibration and thermal equilibration of temperature dependent JWL EOSs, CHEETAH In Line RF also assumes temperature and pressure equilibration. It has been suggested in the past that a more realistic equilibration scheme should comprise isentropic pressure equilibration of the separate reacted and unreacted phases. This turns out not to be a proper path for equilibration. Rather, we find that the only internally consistent method is the evaluation of the equilibrium pressuremore » that satisfies the particular conditions of reactant and product resulting from deflagration in a fixed volume.« less
Fluidically Augmented Nozzles for Pulse Detonation Engine Applications
2011-12-01
25 captured the flow soon after the leading shock wave passed through the diverging section of the nozzle. As can be seen, the “pillow” has begun to...35 Figure 25. Initial Detonation Wave Enters the Diverging Section of the Nozzle...charging the combustor with an appropriate fuel/air mixture. This mixture is then ignited, producing a flame that is initially a deflagration wave . A
Degeneracy breakdown as a source of supernovae Ia
NASA Astrophysics Data System (ADS)
Clavelli, L.
2018-01-01
In a confined system of multiple Fermions, the particles are forced into high energy levels by the Pauli Exclusion Principle. We refer to this system as a Pauli tower. We pursue the investigation of a model for sub-Chandrasekhar supernovae Ia explosions (SNIa) in which the energy stored in the Pauli tower is released to trigger a nuclear deflagration. The simplest physical model for such a degeneracy breakdown and collapse of the Pauli tower is a phase transition to an exactly supersymmetric state in which the scalar partners of protons, neutrons, and leptons become degenerate with the familiar fermions of our world as in the supersymmetric standard model with susy breaking parameters relaxed to zero. We focus on the ability of the susy phase transition model to fit the total SNIa rate as well as the delay time distribution of SNIa after the birth of a progenitor white dwarf. We also study the ejected mass distribution and its correlation with delay time. Finally, we discuss the expected SNIa remnant in the form of a black hole of roughly Jupiter mass and the prospects for detecting such remnants.
2009-12-01
minimal pressure losses. 15. NUMBER OF PAGES 113 14. SUBJECT TERMS Pulse Detonation Combustors, PDC, Pulse Detonation Engines, PDE , PDE ...Postgraduate School PDC Pulse Detonation Combustor PDE Pulse Detonation Engine RAM Random Access Memory RDT Research, Design and Test RPL...inhibiting the implementation of this advanced propulsion system. The primary advantage offered by pulse detonation engines ( PDEs ) is the high efficiency
Experimental Investigation of Turbojet Thrust Augmentation Using an Ejector
2007-03-01
mechanisms in which a particle can exchange energy. Thrust augmenting devices can be divided into two categories: ones that exchange net work or heat and...two categories from the energy equation discussion above. Thrust augmentation is achieved through turbulent entrainment where work and/or heat is...front sustained by compression waves from a trailing reaction zone. A deflagration wave is a subsonic flame front sustained by heat transfer
Self-extinguishing behavior of kerosene spray fire in a completely enclosed compartment
NASA Astrophysics Data System (ADS)
Wang, Changjian; Guo, Jin; Yan, Weigang; Lu, Shouxiang
2013-10-01
The self-extinguishing behavior of kerosene spray fire was investigated in a completely enclosed compartment with the size of 3 m × 3 m × 3.4 m. The spray was generated by locating one BETE nozzle at the center of the bottom wall. A series of spray fire videos were obtained by changing BETE nozzle type and injecting pressure. The results show that spray fire undergoes four stages: the growth stage, the quasi-steady stage, the stretch stage and the self-extinguishing stage. Consumption of large quantities of oxygen causes spray fire to first be stretched and then quench. In this process, fire base migrates away from spray region and leads to the emergence of ghosting fire. Ghosting fire promotes the instability of spray fire and large fluctuation of its height, which provides help to its self-extinguishing. With increasing the injecting pressure or the nozzle diameter, the self-extinguishing time decreases. It is found that the self-extinguishing time is approximately in inverse relation with injecting flow rate. Additionally, we also observed the occurrence of two-phase deflagration just after ignition, and it accelerates the spray fire growth and induces a larger fire height than the following quasi-steady spray fire. The deflagration turns stronger with increasing the injecting pressure.
In-Situ Imaging of Particles during Rapid Thermite Deflagrations
NASA Astrophysics Data System (ADS)
Grapes, Michael; Reeves, Robert; Densmore, John; Fezzaa, Kamel; van Buuren, Tony; Willey, Trevor; Sullivan, Kyle
2017-06-01
The dynamic behavior of rapidly deflagrating thermites is a highly complex process involving rapid decomposition, melting, and outgassing of intermediate and/or product gases. Few experimental techniques are capable of probing these phenomena in situ due to the small length and time scales associated with the reaction. Here we use a recently developed extended burn tube test, where we initiate a small pile of thermite on the closed end of a clear acrylic tube. The length of the tube is sufficient to fully contain the reaction as it proceeds and flows entrained particles down the tube. This experiment was brought to the Advanced Photon Source, and the particle formation was X-ray imaged at various positions down the tube. Several formulations, as well as formulation parameters were varied to investigate the size and morphology of the particles, as well as to look for dynamic behavior attributed to the reaction. In all cases, we see evidence of particle coalescence and condensed-phase interfacial reactions. The results improve our understanding of the procession of reactants to products in these systems. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344. LLNL-ABS-691140.
Explosive Characteristics of Carbonaceous Nanoparticles
NASA Astrophysics Data System (ADS)
Turkevich, Leonid; Fernback, Joseph; Dastidar, Ashok
2013-03-01
Explosion testing has been performed on 20 codes of carbonaceous particles. These include SWCNTs (single-walled carbon nanotubes), MWCNTs (multi-walled carbon nanotubes), CNFs (carbon nanofibers), graphene, diamond, fullerene, carbon blacks and graphites. Explosion screening was performed in a 20 L explosion chamber (ASTM E1226-10 protocol), at a (dilute) concentration of 500 g/m3, using a 5 kJ ignition source. Time traces of overpressure were recorded. Samples exhibited overpressures of 5-7 bar, and deflagration index KSt = V1/3 (dp/pt)max ~ 10 - 80 bar-m/s, which places these materials in European Dust Explosion Class St-1 (similar to cotton and wood dust). There was minimal variation between these different materials. The explosive characteristics of these carbonaceous powders are uncorrelated with particle size (BET specific surface area). Additional tests were performed on selected materials to identify minimum explosive concentration [MEC]. These materials exhibit MEC ~ 101 -102 g/m3 (lower than the MEC for coals). The concentration scans confirm that the earlier screening was performed under fuel-rich conditions (i.e. the maximum over-pressure and deflagration index exceed the screening values); e.g. the true fullerene KSt ~ 200 bar-m/s, placing it borderline St-1/St-2. Work supported through the NIOSH Nanotechnology Research Center (NTRC)
Pressure Amplification Off High Impedance Barriers in DDT
DOE Office of Scientific and Technical Information (OSTI.GOV)
Heatwole, Eric Mann; Broilo, Robert M.; Kistle, Trevin Joseph
The Deflagration-to-Detonation Transition (DDT) in one-dimensional porous explosive, where combustion in an explosive transitions to detonation, can be described by the following model. This simplified model proceeds in five steps, as follows: 1) Ignition of the explosive, surface burning. 2) Convective burning, with the flame front penetrating through the porous network of the explosive. This proceeds until the pressure grows high enough to result in choked flow in the pores restricting the convective burn. 3) The choked flow results in the formation of a high-density compact of explosive. This compact is driven into undisturbed material by the pressure of themore » burning explosive. See Figure1. 4) The compression of the undisturbed porous explosive by the compact leads to the ignition of a compressive burn. This builds in pressure until a supported shock forms. 5) The shock builds in pressure until detonation occurs. See Figure 2 for an overview streak of the proceeding steps.« less
Critical velocities for deflagration and detonation triggered by voids in a REBO high explosive
DOE Office of Scientific and Technical Information (OSTI.GOV)
Herring, Stuart Davis; Germann, Timothy C; Jensen, Niels G
2010-01-01
The effects of circular voids on the shock sensitivity of a two-dimensional model high explosive crystal are considered. We simulate a piston impact using molecular dynamics simulations with a Reactive Empirical Bond Order (REBO) model potential for a sub-micron, sub-ns exothermic reaction in a diatomic molecular solid. The probability of initiating chemical reactions is found to rise more suddenly with increasing piston velocity for larger voids that collapse more deterministically. A void with radius as small as 10 nm reduces the minimum initiating velocity by a factor of 4. The transition at larger velocities to detonation is studied in amore » micron-long sample with a single void (and its periodic images). The reaction yield during the shock traversal increases rapidly with velocity, then becomes a prompt, reliable detonation. A void of radius 2.5 nm reduces the critical velocity by 10% from the perfect crystal. A Pop plot of the time-to-detonation at higher velocities shows a characteristic pressure dependence.« less
Prediction of Ignition of High Explosive When Submitted To Impact
NASA Astrophysics Data System (ADS)
Picart, Didier; Delmaire-Sizes, Franck; Gruau, Cyril; Trumel, Herve
2009-06-01
High explosive structures may unintentionally ignite and transit to deflagration or detonation, when subjected to mechanical loadings, such as low velocity impact. We focus our attention on ignition. The Browning and Scammon [1] criterion has been adapted. A concrete like constitutive law is derived, with an up-to-date experimental characterization. These models have been implemented in Abaqus/Explicit [2]. Numerical simulations are used to calibrate the ignition threshold. The presentation or the poster will detail the main assumptions, the models (Browning et al, mechanical behavior) and the calibration procedure. Comparisons between numerical results and experiments [3] will show the interest of this method but also its limitations (numerical artifacts, lack of mechanical data, misinterpretation of reactive tests). [1] R. Browning and R. Scammon, Shock compression of condensed matter, pp. 987-990, (2001). [2] C. Gruau, D. Picart et al., 17^th Dymat technical meeting, Cambridge, UK, (2007). [3] F. Delmaire-Sizes et al., 3^rd International symposium on energetic materials, Tokyo, Japan, (2008).
NASA Astrophysics Data System (ADS)
Nguyen, V. B.; Li, J.; Chang, P.-H.; Phan, Q. T.; Teo, C. J.; Khoo, B. C.
2018-01-01
In this paper, numerical simulations are performed to study the dynamics of the deflagration-to-detonation transition (DDT) in pulse detonation engines (PDE) using energetic aluminum particles. The DDT process and detonation wave propagation toward the unburnt hydrogen/air mixture containing solid aluminum particles is numerically studied using the Eulerian-Lagrangian approach. A hybrid numerical methodology combined with appropriate sub-models is used to capture the gas dynamic characteristics, particle behavior, combustion characteristics, and two-way solid-particle-gas flow interactions. In our approach, the gas mixture is expressed in the Eulerian frame of reference, while the solid aluminum particles are tracked in the Lagrangian frame of reference. The implemented computer code is validated using published benchmark problems. The obtained results show that the aluminum particles not only shorten the DDT length but also reduce the DDT time. The improvement of DDT is primarily attributed to the heat released from surface chemical reactions on the aluminum particles. The temperatures associated with the DDT process are greater than the case of non-reacting particles added, with an accompanying rise in the pressure. For an appropriate range of particle volume fraction, particularly in this study, the higher volume fraction of the micro-aluminum particles added in the detonation chamber can lead to more heat energy released and more local instabilities in the combustion process (caused by the local high temperature), thereby resulting in a faster DDT process. In essence, the aluminum particles contribute to the DDT process of successfully transitioning to detonation waves for (failure) cases in which the fuel gas mixture can be either too lean or too rich. With a better understanding of the influence of added aluminum particles on the dynamics of the DDT and detonation process, we can apply it to modify the geometry of the detonation chamber (e.g., the length of the detonation tube) accordingly to improve the operational performance of the PDE.
Towards a predictive thermal explosion model for energetic materials
NASA Astrophysics Data System (ADS)
Yoh, Jack J.; McClelland, Matthew A.; Maienschein, Jon L.; Wardell, Jeffrey F.
2005-01-01
We present an overview of models and computational strategies for simulating the thermal response of high explosives using a multi-physics hydrodynamics code, ALE3D. Recent improvements to the code have aided our computational capability in modeling the behavior of energetic materials systems exposed to strong thermal environments such as fires. We apply these models and computational techniques to a thermal explosion experiment involving the slow heating of a confined explosive. The model includes the transition from slow heating to rapid deflagration in which the time scale decreases from days to hundreds of microseconds. Thermal, mechanical, and chemical effects are modeled during all phases of this process. The heating stage involves thermal expansion and decomposition according to an Arrhenius kinetics model while a pressure-dependent burn model is employed during the explosive phase. We describe and demonstrate the numerical strategies employed to make the transition from slow to fast dynamics. In addition, we investigate the sensitivity of wall expansion rates to numerical strategies and parameters. Results from a one-dimensional model show that violence is influenced by the presence of a gap between the explosive and container. In addition, a comparison is made between 2D model and measured results for the explosion temperature and tube wall expansion profiles.
Direct Initiation Through Detonation Branching in a Pulsed Detonation Engine
2008-03-01
important features noted ................................. 33 Figure 20. GM Quad 4 engine head used as the PDE research engine with the detonation tube...Deflagration to Detonation Transition EF – Engine Frequency FF – Fill Fraction NPT – National Pipe Thread MPT – Male National Pipe Thread PDE – Pulsed... Detonation Engines ( PDE ) has increased greatly in recent years due in part to the potential for increased thermal efficiency derived from constant
Burning to Detonation Transition in Porous Beds of a High Energy Propellant
1979-11-01
deflagration to detonation transition (DDT) behavior of porous charges of a high-energy propellant (VL-U) has been stualed at both high and low...transition (DDT) behavior of porous charges of a high-eneryy propellant (VLU) has been studied at both high confinement (steel tube) and low confinement...plastic tube). The physical nature of thet propellant ised to construct the porous beds was varied: shredded, powdered CI ,-: and cuboid materials were
Reflected stochastic differential equation models for constrained animal movement
Hanks, Ephraim M.; Johnson, Devin S.; Hooten, Mevin B.
2017-01-01
Movement for many animal species is constrained in space by barriers such as rivers, shorelines, or impassable cliffs. We develop an approach for modeling animal movement constrained in space by considering a class of constrained stochastic processes, reflected stochastic differential equations. Our approach generalizes existing methods for modeling unconstrained animal movement. We present methods for simulation and inference based on augmenting the constrained movement path with a latent unconstrained path and illustrate this augmentation with a simulation example and an analysis of telemetry data from a Steller sea lion (Eumatopias jubatus) in southeast Alaska.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hicks, E. P.; Rosner, R., E-mail: eph2001@columbia.edu
In this paper, we provide support for the Rayleigh-Taylor-(RT)-based subgrid model used in full-star simulations of deflagrations in Type Ia supernovae explosions. We use the results of a parameter study of two-dimensional direct numerical simulations of an RT unstable model flame to distinguish between the two main types of subgrid models (RT or turbulence dominated) in the flamelet regime. First, we give scalings for the turbulent flame speed, the Reynolds number, the viscous scale, and the size of the burning region as the non-dimensional gravity (G) is varied. The flame speed is well predicted by an RT-based flame speed model.more » Next, the above scalings are used to calculate the Karlovitz number (Ka) and to discuss appropriate combustion regimes. No transition to thin reaction zones is seen at Ka = 1, although such a transition is expected by turbulence-dominated subgrid models. Finally, we confirm a basic physical premise of the RT subgrid model, namely, that the flame is fractal, and thus self-similar. By modeling the turbulent flame speed, we demonstrate that it is affected more by large-scale RT stretching than by small-scale turbulent wrinkling. In this way, the RT instability controls the flame directly from the large scales. Overall, these results support the RT subgrid model.« less
2008-09-01
projectiles containing small amounts of a reactive material. The mechanism is that limited deflagration of the ANFO creates sufficient pressure to...resulting pressurization of the container causes the container to rupture, thus producing a render-safe solution. Several free-field shots demonstrated...the ANFO creates sufficient pressure to rupture plastic or steel containers. 1 Introduction Vehicle-borne improvised explosive devices (VBIEDs) have
Deflagration to Shock to Detonation Transition of Energetic Propellants.
1984-07-01
Champaign Urbana, IL 61801 Prepared for: Air Force Office of Scientific Research Aerospace Sciences Directorate: Bldg. 410 Boiling Air Force Base, DC 20332...Program Manager: Dr. Henry R. Radoski For Research Performed under ()Grant AFOSR 81-0145 During Period I June 1983 through 30 May 1984 AIR ,,1 C~i...ADDRESS 12. REPORT DATE / Air Force Office of Scientific Research/Q July 1984 Building 410; ATTN: Henry R. Radoski 13. NUMBER OF AGES Bolling Air Force
KA-111, Phase C, M-1 Propellant Tests: Deflagration in Partial Confinement.
1991-07-01
DNA Test Director and Mr. R. !. Flory, Washington Research Center, was Program Coordinator. The DDESB, NDCS , and SSO Technical Monitors for Phase C...to simulate the chamber and access tunnel proportions of the Shallow Underground Tunnel /Chamber Explosion Test conducted at China Lake, CA, in 1988...The chamber and access tunnel at China Lake had the following dimensions (volume, cross-sectional area and length): Chamber: V. - 331.2 m 3 Tunnel : Vt
Design of a Simple Blast Pressure Gauge Based on a Heterodyne Velocimetry Measuring Technique
2016-08-01
deployed in an experiment during which the blast pressure was measured from detonation of 114 g of Primasheet 1000 high explosive. The gauge reported... detonation of high explosive where accelerated projectiles and debris may occur. Many times, overpressures generated by such events can be a nuisance to...as that generated by release of energy from a high-explosive detonation or deflagration, materials such as metals or ceramics may be needed. A
Unsteady specific work and isentropic efficiency of a radial turbine driven by pulsed detonations
NASA Astrophysics Data System (ADS)
Rouser, Kurt P.
There has been longstanding government and industry interest in pressure-gain combustion for use in Brayton cycle based engines. Theoretically, pressure-gain combustion allows heat addition with reduced entropy loss. The pulsed detonation combustor (PDC) is a device that can provide such pressure-gain combustion and possibly replace typical steady deflagration combustors. The PDC is inherently unsteady, however, and comparisons with conventional steady deflagration combustors must be based upon time-integrated performance variables. In this study, the radial turbine of a Garrett automotive turbocharger was coupled directly to and driven, full admission, by a PDC in experiments fueled by hydrogen or ethylene. Data included pulsed cycle time histories of turbine inlet and exit temperature, pressure, velocity, mass flow, and enthalpy. The unsteady inlet flowfield showed momentary reverse flow, and thus unsteady accumulation and expulsion of mass and enthalpy within the device. The coupled turbine-driven compressor provided a time-resolved measure of turbine power. Peak power increased with PDC fill fraction, and duty cycle increased with PDC frequency. Cycle-averaged unsteady specific work increased with fill fraction and frequency. An unsteady turbine efficiency formulation is proposed, including heat transfer effects, enthalpy flux-weighted total pressure ratio, and ensemble averaging over multiple cycles. Turbine efficiency increased with frequency but was lower than the manufacturer reported conventional steady turbine efficiency.
Ajrash, Mohammed J; Zanganeh, Jafar; Moghtaderi, Behdad
2017-10-05
There has been a surge of interest from the extractive industries in the application of mechanical means to the mitigation of flame deflagration. To verify the implementation and performance of passive and active mitigation protection, a comprehensive experimental investigation has been conducted on a large scale detonation tube, 30m long and 0.5m in diameter, with two mitigation valves (passive and active) and a burst panel venting system. The valves were used alternately to mitigate the flame deflagration of methane in concentrations ranging from 1.25% to 7.5%. The experimental work revealed that locating the passive mitigation valve at 22m distance from the ignition source mitigates the flame by fully isolating the tube. However, closing the valve structure in the axial direction generated another pressure wave upstream, which was approximately the same value as for the original pressure wave upstream. In the case of the active mitigation system, the system perfectly isolated upstream from downstream with no further pressure wave generation. When the vent was located at 6.5m from the ignition source, the total pressure was reduced by 0.48bar. Due to the counter flow of the reflected pressure wave the flame was extinguished at 12.5m from the ignition source. Copyright © 2017 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Maeda, S.; Minami, S.; Okamoto, D.; Obara, T.
2016-09-01
The deflagration-to-detonation transition in a 100 mm square cross-section channel was investigated for a highly reactive stoichiometric hydrogen oxygen mixture at 70 kPa. Obstacles of 5 mm width and 5, 10, and 15 mm heights were equally spaced 60 mm apart at the bottom of the channel. The phenomenon was investigated primarily by time-resolved schlieren visualization from two orthogonal directions using a high-speed video camera. The detonation transition occurred over a remarkably short distance within only three or four repeated obstacles. The global flame speed just before the detonation transition was well below the sound speed of the combustion products and did not reach the sound speed of the initial unreacted gas for tests with an obstacle height of 5 and 10 mm. These results indicate that a detonation transition does not always require global flame acceleration beyond the speed of sound for highly reactive combustible mixtures. A possible mechanism for this detonation initiation was the mixing of the unreacted and reacted gas in the vicinity of the flame front convoluted by the vortex present behind each obstacle, and the formation of a hot spot by the shock wave. The final onset of the detonation originated from the unreacted gas pocket, which was surrounded by the obstacle downstream face and the channel wall.
Duan, Qiangling; Xiao, Huahua; Gao, Wei; Gong, Liang; Sun, Jinhua
2016-12-15
An experimental investigation of spontaneous ignition and flame propagation at high-pressure hydrogen release via cylindrical tubes with varying cross-section is presented. Tubes with different transverse cross-sections are considered in the experiments: (1) local contraction, (2) local enlargement, (3) abrupt contraction, and (4) abrupt enlargement. The results show that the presence of the varying cross-section geometries can significantly promote the occurrence of spontaneous ignition. Compared to the tube with constant cross-section, the minimum pressure release needed for spontaneous ignition for the varying cross-sections tubes is considerably lower. Moreover, the initial ignition location is closer to the disk in the presence of varying cross-section geometries in comparison with straight channel. As the flame emerges from the outlet of the tube, the velocity of the flame front in the vicinity of the nozzle increases sharply. Then, a deflagration develops across the mixing zone of hydrogen/air mixture. The maximum deflagration overpressure increases linearly with the release pressure. Subsequently, a hydrogen jet flame is produced and evolves different shapes at different release stages. A fireball is formed after the jet flame spouts in the open air. Later, the fireball develops into a jet flame which shifts upward and continues to burn in the vertical direction. Copyright © 2016 Elsevier B.V. All rights reserved.
Premaximum observations of the type Ia SN 1990N
NASA Technical Reports Server (NTRS)
Leibundgut, Bruno; Kirshner, Robert P.; Filippenko, Alexei V.; Shields, Joseph C.; Foltz, Craig B.; Phillips, Mark M.; Sonneborn, George
1991-01-01
Spectroscopic and photometric observations of SN 1990N were obtained at ultraviolet and optical wavelengths, beginning 14 days before maximum light. The early observations reveal important differences from spectra of SN Ia's around maximum light. Photometry and spectroscopy obtained after maximum show that SN 1990N is a typical SN Ia and that most of the observed differences are due to the early epoch of the observations. The most significant characteristics are (1) the high velocities of Ca and Si up to 22,000 km/s; (2) the presence of Co and Fe 2 weeks before maximum; and (3) the more rapid increase in the UV flux compared to the optical. The most popular models for white dwarf deflagration that have provided the standard interpretation for SN Ia's at maximum light do not reproduce the high velocities of Ca II and Si II lines observed in SN 1990N.
'Tertiary' nuclear burning - Neutron star deflagration?
NASA Technical Reports Server (NTRS)
Michel, F. Curtis
1988-01-01
A motivation is presented for the idea that dense nuclear matter can burn to a new class of stable particles. One of several possibilities is an 'octet' particle which is the 16 baryon extension of alpha particle, but now composed of a pair of each of the two nucleons, (3Sigma, Delta, and 2Xi). Such 'tertiary' nuclear burning (here 'primary' is H-He and 'secondary' is He-Fe) may lead to neutron star explosions rather than collapse to a black hole, analogous to some Type I supernovae models wherein accreting white dwarfs are pushed over the Chandrasekhar mass limit but explode rather than collapse to form neutron stars. Such explosions could possibly give gamma-ray bursts and power quasars, with efficient particle acceleration in the resultant relativistic shocks. The new stable particles themselves could possibly be the sought-after weakly interacting, massive particles (WIMPs) or 'dark' matter.
Analysis of compaction shock interactions during DDT of low density HMX
NASA Astrophysics Data System (ADS)
Rao, Pratap T.; Gonthier, Keith A.
2017-01-01
Deflagration-to-Detonation Transition (DDT) in confined, low density granular HMX occurs by a complex mechanism that involves compaction shock interactions within the material. Piston driven DDT experiments indicate that detonation is abruptly triggered by the interaction of a strong combustion-supported secondary shock and a piston-supported primary (input) shock, where the nature of the interaction depends on initial packing density and primary shock strength. These interactions influence transition by affecting dissipative heating within the microstructure during pore collapse. Inert meso-scale simulations of successive shock loading of low density HMX are performed to examine how dissipation and hot-spot formation are affected by the initial density, and the primary and secondary shock strengths. This information is used to formulate an ignition and burn model for low density HMX that accounts for the effect of shock densensitization on burn. Preliminary DDT predictions are presented that illustrate how primary shock strength affects the transition mechanism.
Constrained optimization via simulation models for new product innovation
NASA Astrophysics Data System (ADS)
Pujowidianto, Nugroho A.
2017-11-01
We consider the problem of constrained optimization where the decision makers aim to optimize the primary performance measure while constraining the secondary performance measures. This paper provides a brief overview of stochastically constrained optimization via discrete event simulation. Most review papers tend to be methodology-based. This review attempts to be problem-based as decision makers may have already decided on the problem formulation. We consider constrained optimization models as there are usually constraints on secondary performance measures as trade-off in new product development. It starts by laying out different possible methods and the reasons using constrained optimization via simulation models. It is then followed by the review of different simulation optimization approach to address constrained optimization depending on the number of decision variables, the type of constraints, and the risk preferences of the decision makers in handling uncertainties.
Di Sarli, Valeria; Di Benedetto, Almerinda; Russo, Gennaro
2010-08-15
In this work, an assessment of different sub-grid scale (sgs) combustion models proposed for large eddy simulation (LES) of steady turbulent premixed combustion (Colin et al., Phys. Fluids 12 (2000) 1843-1863; Flohr and Pitsch, Proc. CTR Summer Program, 2000, pp. 61-82; Kim and Menon, Combust. Sci. Technol. 160 (2000) 119-150; Charlette et al., Combust. Flame 131 (2002) 159-180; Pitsch and Duchamp de Lageneste, Proc. Combust. Inst. 29 (2002) 2001-2008) was performed to identify the model that best predicts unsteady flame propagation in gas explosions. Numerical results were compared to the experimental data by Patel et al. (Proc. Combust. Inst. 29 (2002) 1849-1854) for premixed deflagrating flame in a vented chamber in the presence of three sequential obstacles. It is found that all sgs combustion models are able to reproduce qualitatively the experiment in terms of step of flame acceleration and deceleration around each obstacle, and shape of the propagating flame. Without adjusting any constants and parameters, the sgs model by Charlette et al. also provides satisfactory quantitative predictions for flame speed and pressure peak. Conversely, the sgs combustion models other than Charlette et al. give correct predictions only after an ad hoc tuning of constants and parameters. Copyright 2010 Elsevier B.V. All rights reserved.
Reduced Order Models for Reactions of Energetic Materials
NASA Astrophysics Data System (ADS)
Kober, Edward
The formulation of reduced order models for the reaction chemistry of energetic materials under high pressures is needed for the development of mesoscale models in the areas of initiation, deflagration and detonation. Phenomenologically, 4-8 step models have been formulated from the analysis of cook-off data by analyzing the temperature rise of heated samples. Reactive molecular dynamics simulations have been used to simulate many of these processes, but reducing the results of those simulations to simple models has not been achieved. Typically, these efforts have focused on identifying molecular species and detailing specific chemical reactions. An alternative approach is presented here that is based on identifying the coordination geometries of each atom in the simulation and tracking classes of reactions by correlated changes in these geometries. Here, every atom and type of reaction is documented for every time step; no information is lost from unsuccessful molecular identification. Principal Component Analysis methods can then be used to map out the effective chemical reaction steps. For HMX and TATB decompositions simulated with ReaxFF, 90% of the data can be explained by 4-6 steps, generating models similar to those from the cook-off analysis. By performing these simulations at a variety of temperatures and pressures, both the activation and reaction energies and volumes can then be extracted.
NASA Technical Reports Server (NTRS)
Margolis, Stephen B.
1997-01-01
The burning of liquid propellants is a fundamental combustion problem that is applicable to various types of propulsion and energetic systems. The deflagration process is often rather complex, with vaporization and pyrolysis occurring at the liquid/gas interface and distributed combustion occurring either in the gas phase or in a spray. Nonetheless, there are realistic limiting cases in which combustion may be approximated by an overall reaction at the liquid/gas interface. In one such limit, the gas flame occurs under near-breakaway conditions, exerting little thermal or hydrodynamic influence on the burning propellant. In another such limit, distributed combustion occurs in an intrusive regime, the reaction zone lying closer to the liquid/gas interface than the length scale of any disturbance of interest. Finally, the liquid propellant may simply undergo exothermic decomposition at the surface without any significant distributed combustion, such as appears to occur in some types of HydroxylAmmonium Nitrate (HAN)-based liquid propellants at low pressures. Such limiting models have recently been formulated,thereby significantly generalizing earlier classical models that were originally introduced to study the hydrodynamic stability of a reactive liquid/gas interface. In all of these investigations, gravity appears explicitly and plays a significant role, along with surface tension, viscosity, and, in the more recent models, certain reaction-rate parameters associated with the pressure and temperature sensitivities of the reaction itself. In particular, these parameters determine the stability of the deflagration with respect to not only classical hydrodynamic disturbances, but also with respect to reactive/diffusive influences as well. Indeed, the inverse Froude number, representing the ratio of buoyant to inertial forces, appears explicitly in all of these models, and consequently, in the dispersion relation that determines the neutral stability boundaries beyond which steady, planar burning is unstable to nonsteady, and/or nonplanar (cellular) modes of burning. These instabilities thus lead to a number of interesting phenomena, such as the sloshing type of waves that have been observed in mixtures of HAN and TriEthanolAmmonium Nitrate (TEAN) with water. Although the Froude number was treated as an O(1) quantity in these studies, the limit of small inverse Froude number corresponding to the microgravity regime is increasingly of interest and can be treated explicitly, leading to various limiting forms of the models, the neutral stability boundaries, and, ultimately, the evolution equations that govern the nonlinear dynamics of the propagating reaction front. In the present work, we formally exploit this limiting parameter regime to compare some of the features of hydrodynamic instability of liquid-propellant combustion at reduced gravity with the same phenomenon at normal gravity.
Path correction of free flight projectiles by cross firing of subsidiary projectiles
NASA Astrophysics Data System (ADS)
Stroem, L.
1982-10-01
Terminal guidance of gun-fired shells is described. The path is corrected by shooting out throw-bodies from the shell casing. The drawbacks of the method, e.g., casing deformation, were eliminated. Using deflagrating substances instead of explosives, higher impulses were obtained, and at lower pressure levels. At acceleration distances of only 10 to 15 mm throw-body speeds of 400 to 500 m/sec were noted, allowing this method to be applied to rotation-stabilized shells.
Visualization and Measurement of the Deflagration of JA2 Bonded to Various Metal Foils
2016-01-01
Disclaimers The findings in this report are not to be construed as an official Department of the Army position unless so designated by other... design tools. There are numerous design parameters, including 1) the wires’ thermophysical properties and diameter(s), 2) their quantity, spacing, and...gasified by combustion. This is suggested by Figs. 7 and 8, which present posttest examinations of 2- and 3-mil-thick foils. On the 2-mil-thick Cu foils, a
Microstructure of the combustion zone: Thin-binder AP-polymer sandwiches
NASA Technical Reports Server (NTRS)
Price, E. W.; Panyam, R. R.; Sigman, R. K.
1980-01-01
Experimental results are summarized for systematic quench-burning tests on ammonium perchlorate-HC binder sandwiches with binder thicknesses in the range 10 - 150 microns. Tests included three binders (polysulfide, polybutadiene-acrylonitrile, and hydroxy terminated polybutadiene), and pressures from 1.4 to 14 MPa. In addition, deflagration limits were determined in terms of binder thickness and pressure. Results are discussed in terms of a qualitative theory of sandwich burning consolidated from various sources. Some aspects of the observed results are explained only speculatively.
Analytical Solution of a Generalized Hirota-Satsuma Equation
NASA Astrophysics Data System (ADS)
Kassem, M.; Mabrouk, S.; Abd-el-Malek, M.
A modified version of generalized Hirota-Satsuma is here solved using a two parameter group transformation method. This problem in three dimensions was reduced by Estevez [1] to a two dimensional one through a Lie transformation method and left unsolved. In the present paper, through application of symmetry transformation the Lax pair has been reduced to a system of ordinary equations. Three transformations cases are investigated. The obtained analytical solutions are plotted and show a profile proper to deflagration processes, well described by Degasperis-Procesi equation.
Constrained reduced-order models based on proper orthogonal decomposition
Reddy, Sohail R.; Freno, Brian Andrew; Cizmas, Paul G. A.; ...
2017-04-09
A novel approach is presented to constrain reduced-order models (ROM) based on proper orthogonal decomposition (POD). The Karush–Kuhn–Tucker (KKT) conditions were applied to the traditional reduced-order model to constrain the solution to user-defined bounds. The constrained reduced-order model (C-ROM) was applied and validated against the analytical solution to the first-order wave equation. C-ROM was also applied to the analysis of fluidized beds. Lastly, it was shown that the ROM and C-ROM produced accurate results and that C-ROM was less sensitive to error propagation through time than the ROM.
1974-01-01
REGRESSION MODEL - THE UNCONSTRAINED, LINEAR EQUALITY AND INEQUALITY CONSTRAINED APPROACHES January 1974 Nelson Delfino d’Avila Mascarenha;? Image...Report 520 DIGITAL IMAGE RESTORATION UNDER A REGRESSION MODEL THE UNCONSTRAINED, LINEAR EQUALITY AND INEQUALITY CONSTRAINED APPROACHES January...a two- dimensional form adequately describes the linear model . A dis- cretization is performed by using quadrature methods. By trans
Role of nuclear reactions on stellar evolution of intermediate-mass stars
NASA Astrophysics Data System (ADS)
Möller, H.; Jones, S.; Fischer, T.; Martínez-Pinedo, G.
2018-01-01
The evolution of intermediate-mass stars (8 - 12 solar masses) represents one of the most challenging subjects in nuclear astrophysics. Their final fate is highly uncertain and strongly model dependent. They can become white dwarfs, they can undergo electron-capture or core-collapse supernovae or they might even proceed towards explosive oxygen burning and a subsequent thermonuclear explosion. We believe that an accurate description of nuclear reactions is crucial for the determination of the pre-supernova structure of these stars. We argue that due to the possible development of an oxygen-deflagration, a hydrodynamic description has to be used. We implement a nuclear reaction network with ∼200 nuclear species into the implicit hydrodynamic code AGILE. The reaction network considers all relevant nuclear electron captures and beta-decays. For selected relevant nuclear species, we include a set of updated reaction rates, for which we discuss the role for the evolution of the stellar core, at the example of selected stellar models. We find that the final fate of these intermediate-mass stars depends sensitively on the density threshold for weak processes that deleptonize the core.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Groth, Katrina M.; Zumwalt, Hannah Ruth; Clark, Andrew Jordan
2016-03-01
Hydrogen Risk Assessment Models (HyRAM) is a prototype software toolkit that integrates data and methods relevant to assessing the safety of hydrogen fueling and storage infrastructure. The HyRAM toolkit integrates deterministic and probabilistic models for quantifying accident scenarios, predicting physical effects, and characterizing the impact of hydrogen hazards, including thermal effects from jet fires and thermal pressure effects from deflagration. HyRAM version 1.0 incorporates generic probabilities for equipment failures for nine types of components, and probabilistic models for the impact of heat flux on humans and structures, with computationally and experimentally validated models of various aspects of gaseous hydrogen releasemore » and flame physics. This document provides an example of how to use HyRAM to conduct analysis of a fueling facility. This document will guide users through the software and how to enter and edit certain inputs that are specific to the user-defined facility. Description of the methodology and models contained in HyRAM is provided in [1]. This User’s Guide is intended to capture the main features of HyRAM version 1.0 (any HyRAM version numbered as 1.0.X.XXX). This user guide was created with HyRAM 1.0.1.798. Due to ongoing software development activities, newer versions of HyRAM may have differences from this guide.« less
Simulating Hadronic-to-Quark-Matter with Burn-UD: Recent work and astrophysical applications
NASA Astrophysics Data System (ADS)
Welbanks, Luis; Ouyed, Amir; Koning, Nico; Ouyed, Rachid
2017-06-01
We present the new developments in Burn-UD, our in-house hydrodynamic combustion code used to model the phase transition of hadronic-to-quark matter. Our two new modules add neutrino transport and the time evolution of a (u, d, s) quark star (QS). Preliminary simulations show that the inclusion of neutrino transport points towards new hydrodynamic instabilities that increase the burning speed. A higher burning speed could elicit the deflagration to detonation of a neutron star (NS) into a QS. We propose that a Quark-Nova (QN: the explosive transition of a NS to a QS) could help us explain the most energetic astronomical events to this day: superluminous supernovae (SLSNe). Our models consider a QN occurring in a massive binary, experiencing two common envelope stages and a QN occurring after the supernova explosion of a Wolf-Rayet (WO) star. Both models have been successful in explaining the double humped light curves of over half a dozen SLSNe. We also introduce SiRop our r-process simulation code and propose that a QN site has the hot temperatures and neutron densities required to make it an ideal site for the r-process.
Type Ia Supernova Explosions from Hybrid Carbon-Oxygen-Neon White Dwarf Progenitors
NASA Astrophysics Data System (ADS)
Willcox, Donald E.; Townsley, Dean M.; Calder, Alan C.; Denissenkov, Pavel A.; Herwig, Falk
2016-11-01
Motivated by recent results in stellar evolution that predict the existence of hybrid white dwarf (WD) stars with a C-O core inside an O-Ne shell, we simulate thermonuclear (Type Ia) supernovae from these hybrid progenitors. We use the FLASH code to perform multidimensional simulations in the deflagration-to-detonation transition (DDT) explosion paradigm. Our hybrid progenitor models were produced with the MESA stellar evolution code and include the effects of the Urca process, and we map the progenitor model to the FLASH grid. We performed a suite of DDT simulations over a range of ignition conditions consistent with the progenitor’s thermal and convective structure assuming multiple ignition points. To compare the results from these hybrid WD stars to previous results from C-O WDs, we construct a set of C-O WD models with similar properties and similarly simulate a suite of explosions. We find that despite significant variability within each suite, trends distinguishing the explosions are apparent in their {}56{Ni} yields and the kinetic properties of the ejecta. We compare our results with other recent work that studies explosions from these hybrid progenitors.
Modelling shock to detonation transition in PETN using HERMES and CREST
NASA Astrophysics Data System (ADS)
Maheswaran, Mary-Ann; Curtis, John; Reaugh, Jack
2013-06-01
The High Explosive Response to MEchanical Stimulus (HERMES) model has been developed to address High Explosive Violent Response (HEVR). It is a material model for use in the both the LS-DYNA finite element and ALE3D hydrocodes that enables the modelling of both shock to detonation (SDT) and deflagration to detonation (DDT) transition. As part of its ongoing development and application, model parameters for the explosive PETN were found by using experimental data for PETN at different densities. PETN was selected because of the availability of both SDT and DDT data. To model SDT and DDT, HERMES uses a subset of the CREST reactive burn model with the Mie-Gruneisen equation of state (EOS) for the unreacted explosive and a look-up table for the gas EOS as generated by Cheetah. The unreacted EOS parameters were found first by calculating the principal isentrope of unreacted PETN at TMD from PETN shock Hugoniot data. Then Pop-plot data for PETN was used to fit the CREST parameters at each density. The resulting new PETN HERMES material model provides a platform for further investigations of SDT and DDT in low density PETN powder. JER's activity was performed under the auspices of the US DOE by LLNL under Contract DE-AC52-07NA27344, and partially funded by the Joint US DoD/DOE Munitions Technology Development Program.
Vibration control of multiferroic fibrous composite plates using active constrained layer damping
NASA Astrophysics Data System (ADS)
Kattimani, S. C.; Ray, M. C.
2018-06-01
Geometrically nonlinear vibration control of fiber reinforced magneto-electro-elastic or multiferroic fibrous composite plates using active constrained layer damping treatment has been investigated. The piezoelectric (BaTiO3) fibers are embedded in the magnetostrictive (CoFe2O4) matrix forming magneto-electro-elastic or multiferroic smart composite. A three-dimensional finite element model of such fiber reinforced magneto-electro-elastic plates integrated with the active constrained layer damping patches is developed. Influence of electro-elastic, magneto-elastic and electromagnetic coupled fields on the vibration has been studied. The Golla-Hughes-McTavish method in time domain is employed for modeling a constrained viscoelastic layer of the active constrained layer damping treatment. The von Kármán type nonlinear strain-displacement relations are incorporated for developing a three-dimensional finite element model. Effect of fiber volume fraction, fiber orientation and boundary conditions on the control of geometrically nonlinear vibration of the fiber reinforced magneto-electro-elastic plates is investigated. The performance of the active constrained layer damping treatment due to the variation of piezoelectric fiber orientation angle in the 1-3 Piezoelectric constraining layer of the active constrained layer damping treatment has also been emphasized.
1980-10-27
Reference 13. The 94/6 RDX/ wax (X893) and 97/3 RDX/ wax (X758) were mechanical mixtures prepared from Class A RDX (X597) and carnauba wax (N134). The...UKLAS9*TE SE,’CRITY CLASSIFICATION OF THIS PAGE (When Data Entered) ionization probes in previous steel tube studies. In charges of 94/6 RDX/ wax ...explosives (picric acid, tetryl, and RDX/ wax ) were among those materials in previous steel tube studies at NSWC which achieved deflagration to
1984-06-25
the weight mean particle sizes were:ammonium picrate (285um), TNT (325pm), RDX (200um) and tetryl (470 and 160m). The carnauba wax (125pm) and the...WORDS (enti.e n reverse side if noceemdy and bnt, ock ,by br) PBXN103 ’"mber) Gas loading RDX Expl D PBXW-108 Sensitivity 91/9 ROX/ wax Cast H-6 PBXN...porous HE for their susceptibility to undergoing DDT. Granuiar explosives studied were RDX, waxed RDX, tetryl, TNT, and Explosive D; cast HE, TNT based
Dimensionality effects in void-induced explosive sensitivity
Herring, Stuart Davis; Germann, Timothy Clark; Gronbech-Jensen, Niels
2016-09-02
Here, the dimensionality of defects in high explosives controls their heat generation and the expansion of deflagrations from them. We compare the behaviour of spherical voids in three dimensions to that of circular voids in two dimensions. The behaviour is qualitatively similar, but the additional focusing along the extra transverse dimension significantly reduces the piston velocity needed to initiate reactions. However, the reactions do not grow as well in three dimensions, so detonations require larger piston velocities. Pressure exponents are seen to be similar to those for the two-dimensional system.
An asymptotic analysis of supersonic reacting mixing layers
NASA Technical Reports Server (NTRS)
Jackson, T. L.; Hussaini, M. Y.
1987-01-01
The purpose of this paper is to present an asymptotic analysis of the laminar mixing of the simultaneous chemical reaction between parallel supersonic streams of two reacting species. The study is based on a one-step irreversible Arrhenius reaction and on large activation energy asymptotics. Essentially it extends the work of Linan and Crespo to include the effect of free shear and Mach number on the ignition regime, the deflagration regime and the diffusion flame regime. It is found that the effective parameter is the product of the characteristic Mach number and a shear parameter.
An Efficient Composition for Bengal Lights
NASA Astrophysics Data System (ADS)
Comet, M.; Schreyeck, L.; Fuzellier, H.
2002-01-01
Fuel-oxidizer mixtures based on potassium chlorate or sodium chlorate are well known. These mixtures have interesting properties of deflagration and are often used in propellants. Drastic reactivity of alkaline chlorates with ammonium salts due to the formation of ammonium chlorate NH4ClO3, a very unstable salt, is famous. By analogy, we tested the reactivity of different molecules containing nitrogen atoms, and we found an efficient fuel-oxidizer composed of potassium chlorate and thiocarbamide. Impressive bengal lights of various colors can easily be achieved using this basic composition.
Deflagration-to-Detonation Transition in Heteorogeneous Solids: A Bibliography.
1980-11-01
and Rockets, Vol. 9. No. 6, 1972, pp. 415-419. 1.4 Francois, D., and L. Joly; " La Rupture des Metaux; Ecole d’ete de la Colle sur Loup ," Masson et Cie...Computer Program for Multifield Fluid Flows," Los Alamos Scientific Laboratory, LA -5680, 1974. 5.3, 9 Nnderssen, K. E. B.; "Pressure Drop in Ideal...5.6, 6, 9 Forest, C. A.: "Burning and Detonation," Los Alamos Scientific Laboratory, LA -7245, July 1978. 2, 3, 4 Fox, J.; "Flow Regimes in
NASA Astrophysics Data System (ADS)
Pohle, Ina; Niebisch, Michael; Müller, Hannes; Schümberg, Sabine; Zha, Tingting; Maurer, Thomas; Hinz, Christoph
2018-07-01
To simulate the impacts of within-storm rainfall variabilities on fast hydrological processes, long precipitation time series with high temporal resolution are required. Due to limited availability of observed data such time series are typically obtained from stochastic models. However, most existing rainfall models are limited in their ability to conserve rainfall event statistics which are relevant for hydrological processes. Poisson rectangular pulse models are widely applied to generate long time series of alternating precipitation events durations and mean intensities as well as interstorm period durations. Multiplicative microcanonical random cascade (MRC) models are used to disaggregate precipitation time series from coarse to fine temporal resolution. To overcome the inconsistencies between the temporal structure of the Poisson rectangular pulse model and the MRC model, we developed a new coupling approach by introducing two modifications to the MRC model. These modifications comprise (a) a modified cascade model ("constrained cascade") which preserves the event durations generated by the Poisson rectangular model by constraining the first and last interval of a precipitation event to contain precipitation and (b) continuous sigmoid functions of the multiplicative weights to consider the scale-dependency in the disaggregation of precipitation events of different durations. The constrained cascade model was evaluated in its ability to disaggregate observed precipitation events in comparison to existing MRC models. For that, we used a 20-year record of hourly precipitation at six stations across Germany. The constrained cascade model showed a pronounced better agreement with the observed data in terms of both the temporal pattern of the precipitation time series (e.g. the dry and wet spell durations and autocorrelations) and event characteristics (e.g. intra-event intermittency and intensity fluctuation within events). The constrained cascade model also slightly outperformed the other MRC models with respect to the intensity-frequency relationship. To assess the performance of the coupled Poisson rectangular pulse and constrained cascade model, precipitation events were stochastically generated by the Poisson rectangular pulse model and then disaggregated by the constrained cascade model. We found that the coupled model performs satisfactorily in terms of the temporal pattern of the precipitation time series, event characteristics and the intensity-frequency relationship.
ODE constrained mixture modelling: a method for unraveling subpopulation structures and dynamics.
Hasenauer, Jan; Hasenauer, Christine; Hucho, Tim; Theis, Fabian J
2014-07-01
Functional cell-to-cell variability is ubiquitous in multicellular organisms as well as bacterial populations. Even genetically identical cells of the same cell type can respond differently to identical stimuli. Methods have been developed to analyse heterogeneous populations, e.g., mixture models and stochastic population models. The available methods are, however, either incapable of simultaneously analysing different experimental conditions or are computationally demanding and difficult to apply. Furthermore, they do not account for biological information available in the literature. To overcome disadvantages of existing methods, we combine mixture models and ordinary differential equation (ODE) models. The ODE models provide a mechanistic description of the underlying processes while mixture models provide an easy way to capture variability. In a simulation study, we show that the class of ODE constrained mixture models can unravel the subpopulation structure and determine the sources of cell-to-cell variability. In addition, the method provides reliable estimates for kinetic rates and subpopulation characteristics. We use ODE constrained mixture modelling to study NGF-induced Erk1/2 phosphorylation in primary sensory neurones, a process relevant in inflammatory and neuropathic pain. We propose a mechanistic pathway model for this process and reconstructed static and dynamical subpopulation characteristics across experimental conditions. We validate the model predictions experimentally, which verifies the capabilities of ODE constrained mixture models. These results illustrate that ODE constrained mixture models can reveal novel mechanistic insights and possess a high sensitivity.
Nayhouse, Michael; Kwon, Joseph Sang-Il; Orkoulas, G
2012-05-28
In simulation studies of fluid-solid transitions, the solid phase is usually modeled as a constrained system in which each particle is confined to move in a single Wigner-Seitz cell. The constrained cell model has been used in the determination of fluid-solid coexistence via thermodynamic integration and other techniques. In the present work, the phase diagram of such a constrained system of Lennard-Jones particles is determined from constant-pressure simulations. The pressure-density isotherms exhibit inflection points which are interpreted as the mechanical stability limit of the solid phase. The phase diagram of the constrained system contains a critical and a triple point. The temperature and pressure at the critical and the triple point are both higher than those of the unconstrained system due to the reduction in the entropy caused by the single occupancy constraint.
NASA Astrophysics Data System (ADS)
Strickland, Shawn L.; Vandersall, Kevin S.; Dehaven, Martin R.
2017-06-01
The potential for deflagration-to-detonation transition (DDT) in LLM-105 and RX-55-DQ (94/6 LLM-105/Viton) has been investigated as a function of loading density using high confinement tubes. The high confinement arrangement uses a 76 mm outer diameter by 25 mm inner diameter mild steel tube 320 mm in length with 25 mm thick mild steel end caps ignited using a thermite igniter and was loaded with samples of varying densities. None of the experiments showed a transition to detonation over the entire length with non-violent burning or extinguishing of the burning observed. The hand packed RX-55-DQ molding powder or neat LLM-105 ( 1.1 g/cm3) burned nearly completely and vented non-violently by deforming or splitting the end caps. The RX-55-DQ was tested at higher densities with 1.35 g/cm3 resulting in a burning reaction on the 2nd attempt that fractured the end cap while the 1.85 g/cm3 resulted in the burning reaction extinguishing in the first 15 mm on the 2nd attempt. This work will outline the testing details, present the results, and compare them to the relatively high binder content HMX-based LX-04 (85% HMX and 15% Viton) and ultra-fine TATB results tested under similar confinement. This work performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
Astrophysical Model Selection in Gravitational Wave Astronomy
NASA Technical Reports Server (NTRS)
Adams, Matthew R.; Cornish, Neil J.; Littenberg, Tyson B.
2012-01-01
Theoretical studies in gravitational wave astronomy have mostly focused on the information that can be extracted from individual detections, such as the mass of a binary system and its location in space. Here we consider how the information from multiple detections can be used to constrain astrophysical population models. This seemingly simple problem is made challenging by the high dimensionality and high degree of correlation in the parameter spaces that describe the signals, and by the complexity of the astrophysical models, which can also depend on a large number of parameters, some of which might not be directly constrained by the observations. We present a method for constraining population models using a hierarchical Bayesian modeling approach which simultaneously infers the source parameters and population model and provides the joint probability distributions for both. We illustrate this approach by considering the constraints that can be placed on population models for galactic white dwarf binaries using a future space-based gravitational wave detector. We find that a mission that is able to resolve approximately 5000 of the shortest period binaries will be able to constrain the population model parameters, including the chirp mass distribution and a characteristic galaxy disk radius to within a few percent. This compares favorably to existing bounds, where electromagnetic observations of stars in the galaxy constrain disk radii to within 20%.
Describing litho-constrained layout by a high-resolution model filter
NASA Astrophysics Data System (ADS)
Tsai, Min-Chun
2008-05-01
A novel high-resolution model (HRM) filtering technique was proposed to describe litho-constrained layouts. Litho-constrained layouts are layouts that have difficulties to pattern or are highly sensitive to process-fluctuations under current lithography technologies. HRM applies a short-wavelength (or high NA) model simulation directly on the pre-OPC, original design layout to filter out low spatial-frequency regions, and retain high spatial-frequency components which are litho-constrained. Since no OPC neither mask-synthesis steps are involved, this new technique is highly efficient in run time and can be used in design stage to detect and fix litho-constrained patterns. This method has successfully captured all the hot-spots with less than 15% overshoots on a realistic 80 mm2 full-chip M1 layout in 65nm technology node. A step by step derivation of this HRM technique is presented in this paper.
Late-time Near-infrared Observations of SN 2005df
NASA Astrophysics Data System (ADS)
Diamond, Tiara R.; Hoeflich, Peter; Gerardy, Christopher L.
2015-06-01
We present late-time near-infrared (NIR) spectral evolution, at 200-400 days, for the Type Ia supernova SN 2005df. The spectra show numerous strong emission features of [Co ii], [Co iii], and [Fe ii] throughout the 0.8-1.8 μm region. As the spectrum ages, the cobalt features fade as would be expected from the decay of 56Co to 56Fe. We show that the strong and isolated [Fe ii] emission line at 1.644 μ {m} provides a unique tool to analyze NIR spectra of SNe Ia. Normalization of spectra to this line allows the separation of features produced by stable versus unstable isotopes of iron group elements. We develop a new method of determining the initial central density, {ρ }c, and the magnetic field, B, of the white dwarf (WD) using the width of the 1.644 μ {m} line. The line width (LW) is sensitive because of electron capture in the early stages of burning, which increases as a function of density. The sensitivity of the LW to B increases with time, and the effects of the magnetic field shift toward later times with decreasing {ρ }c. Through comparison with spherical models, the initial central density for SN 2005df is measured as {ρ }c=0.9(+/- 0.2)× {10}9 {g} {{cm}}-3, which corresponds to a WD close to the Chandrasekhar mass, with {M}{WD}=1.31(+/- 0.03) {M}⊙ and systematic error less than 0.04 {M}⊙. This error estimate is based on spherical models. We discuss the potential uncertainties due to multi-dimensional effects, mixing, and rotation. The latter two effects would increase the estimate of the WD mass. Within {M}{Ch} explosions, however, the central density found for SN 2005df is very low for a H-accretor, possibly suggesting a helium star companion or a tidally disrupted WD companion. As an alternative, we suggest mixing of the central region. We find some support for high initial magnetic fields of strength {10}6 {G} for SN 2005df, however, 0 {G} cannot be ruled out because of noise in the spectra combined with low {ρ }c. We discuss our findings in the context of mixing by Rayleigh-Taylor instabilities during deflagration burning and a wide variety of explosion scenarios. Observations strongly support a very limited amount of mixing during a deflagration phase and high central densities characteristic of a {M}{Ch} WD.
NASA Astrophysics Data System (ADS)
Shi, Z.; Crowell, S.; Luo, Y.; Rayner, P. J.; Moore, B., III
2015-12-01
Uncertainty in predicted carbon-climate feedback largely stems from poor parameterization of global land models. However, calibration of global land models with observations has been extremely challenging at least for two reasons. First we lack global data products from systematical measurements of land surface processes. Second, computational demand is insurmountable for estimation of model parameter due to complexity of global land models. In this project, we will use OCO-2 retrievals of dry air mole fraction XCO2 and solar induced fluorescence (SIF) to independently constrain estimation of net ecosystem exchange (NEE) and gross primary production (GPP). The constrained NEE and GPP will be combined with data products of global standing biomass, soil organic carbon and soil respiration to improve the community land model version 4.5 (CLM4.5). Specifically, we will first develop a high fidelity emulator of CLM4.5 according to the matrix representation of the terrestrial carbon cycle. It has been shown that the emulator fully represents the original model and can be effectively used for data assimilation to constrain parameter estimation. We will focus on calibrating those key model parameters (e.g., maximum carboxylation rate, turnover time and transfer coefficients of soil carbon pools, and temperature sensitivity of respiration) for carbon cycle. The Bayesian Markov chain Monte Carlo method (MCMC) will be used to assimilate the global databases into the high fidelity emulator to constrain the model parameters, which will be incorporated back to the original CLM4.5. The calibrated CLM4.5 will be used to make scenario-based projections. In addition, we will conduct observing system simulation experiments (OSSEs) to evaluate how the sampling frequency and length could affect the model constraining and prediction.
Constrained and Unconstrained Partial Adjacent Category Logit Models for Ordinal Response Variables
ERIC Educational Resources Information Center
Fullerton, Andrew S.; Xu, Jun
2018-01-01
Adjacent category logit models are ordered regression models that focus on comparisons of adjacent categories. These models are particularly useful for ordinal response variables with categories that are of substantive interest. In this article, we consider unconstrained and constrained versions of the partial adjacent category logit model, which…
Exploring Systematic Effects in Thermonuclear Supernovae
NASA Astrophysics Data System (ADS)
Jackson, Aaron Perry
Type Ia supernovae (SNe) are bright astrophysical explosions that form a remarkably homogeneous class of objects serving as the premier distance indicators for studying the expansion history of the Universe and the nature of dark energy. Despite the widespread acceptance of the surprising discovery of the acceleration of the expansion of the Universe and the existence of the mysterious dark energy driving it that followed from these studies, the progenitor systems of these explosions are unknown. Knowledge of the progenitor system is required to understand possible systematic effects due to properties of the parent stellar population or host galaxy. While several scenarios have been proposed, the most widely accepted one is the thermonuclear explosion of a near-Chandrasekharmass, carbon-oxygen white dwarf (WD). Under this scenario, the explosive burning begins near the center as a deflagration (subsonic burning) that transitions to a detonation (supersonic burning) some time later after the WD has expanded in response to the energy release. Turbulence, either pre-existing or generated by burning, serves to increase the surface area of the burning front, thus enhancing the fuel consumption rate. In addition, turbulence--flame interaction (TFI) may be responsible for deflagration--detonation transition (DDT). Simulations of this explosion scenario typically parameterize the DDT to occur when the flame reaches a particular density. I performed a suite of two-dimensional (2D) simulations with the compressible, hydrodynamics code FLASH to evaluate the influence of the DDT density on the average yield of radioactive 56Ni that powers the SN light curve. In addition, I considered the compositional dependence of the DDT density to explore one way in which metallicity may influence the explosion outcome. My results have confirmed a new pathway to explain observed trends in the average peak brightness of SNe Ia with host galaxy metallicity. In a separate study, I address the basic physics of modeling flames and turbulent combustion. The disparate length scales in the SN necessitate use of a flame model to capture the effect of burning on unresolved scales. I implemented a method to measure the strength of unresolved turbulence, which is used to estimate the amount of wrinkling of the unresolved flame surface. In addition, the measure of turbulent strength may be used to improve the criterion by which DDT is initiated. These improvements will allow three-dimensional (3D) simulations of the early flame evolution in the presence of strong pre-existing turbulence. The research conducted for this dissertation has led to important insights into the explosion mechanism of SNe Ia. In addition, improvements to the model have allowed and will continue to allow simulations of unprecedented realism of the complex process of exploding WDs in a thermonuclear SN.
ODE Constrained Mixture Modelling: A Method for Unraveling Subpopulation Structures and Dynamics
Hasenauer, Jan; Hasenauer, Christine; Hucho, Tim; Theis, Fabian J.
2014-01-01
Functional cell-to-cell variability is ubiquitous in multicellular organisms as well as bacterial populations. Even genetically identical cells of the same cell type can respond differently to identical stimuli. Methods have been developed to analyse heterogeneous populations, e.g., mixture models and stochastic population models. The available methods are, however, either incapable of simultaneously analysing different experimental conditions or are computationally demanding and difficult to apply. Furthermore, they do not account for biological information available in the literature. To overcome disadvantages of existing methods, we combine mixture models and ordinary differential equation (ODE) models. The ODE models provide a mechanistic description of the underlying processes while mixture models provide an easy way to capture variability. In a simulation study, we show that the class of ODE constrained mixture models can unravel the subpopulation structure and determine the sources of cell-to-cell variability. In addition, the method provides reliable estimates for kinetic rates and subpopulation characteristics. We use ODE constrained mixture modelling to study NGF-induced Erk1/2 phosphorylation in primary sensory neurones, a process relevant in inflammatory and neuropathic pain. We propose a mechanistic pathway model for this process and reconstructed static and dynamical subpopulation characteristics across experimental conditions. We validate the model predictions experimentally, which verifies the capabilities of ODE constrained mixture models. These results illustrate that ODE constrained mixture models can reveal novel mechanistic insights and possess a high sensitivity. PMID:24992156
Optimal vibration control of a rotating plate with self-sensing active constrained layer damping
NASA Astrophysics Data System (ADS)
Xie, Zhengchao; Wong, Pak Kin; Lo, Kin Heng
2012-04-01
This paper proposes a finite element model for optimally controlled constrained layer damped (CLD) rotating plate with self-sensing technique and frequency-dependent material property in both the time and frequency domain. Constrained layer damping with viscoelastic material can effectively reduce the vibration in rotating structures. However, most existing research models use complex modulus approach to model viscoelastic material, and an additional iterative approach which is only available in frequency domain has to be used to include the material's frequency dependency. It is meaningful to model the viscoelastic damping layer in rotating part by using the anelastic displacement fields (ADF) in order to include the frequency dependency in both the time and frequency domain. Also, unlike previous ones, this finite element model treats all three layers as having the both shear and extension strains, so all types of damping are taken into account. Thus, in this work, a single layer finite element is adopted to model a three-layer active constrained layer damped rotating plate in which the constraining layer is made of piezoelectric material to work as both the self-sensing sensor and actuator under an linear quadratic regulation (LQR) controller. After being compared with verified data, this newly proposed finite element model is validated and could be used for future research.
ERIC Educational Resources Information Center
Hoijtink, Herbert; Molenaar, Ivo W.
1997-01-01
This paper shows that a certain class of constrained latent class models may be interpreted as a special case of nonparametric multidimensional item response models. Parameters of this latent class model are estimated using an application of the Gibbs sampler, and model fit is investigated using posterior predictive checks. (SLD)
Shock Development and Transition to Detonation Initiated by Burning in Porous Propellant Beds
1981-02-01
CAT ALCIG NIJMIIFERvPFOSR-TR. 31 - 5 4I4-~ 1 AD 4 TITLE (Anil SubfrfieJ S TYPE OP REPORT & PERIOD COVERt 0 " S1 10(K DI)ELVIOIMI NT AND) [BAV,1TIION...SEZUR11 ’Y CLA~SIFICAT1st4 OF V P AGF"" l 0)1- ~ N .- -- ABSTRACT This report summarizes the analyses of deflagration to detonation transition (DUT...is reported , and a similar analysis for one-dimensional , two-phase reactive flow is discussed. Improvements made in the gas phase nonideal equation
Spark Ignition of Combustible Vapor in a Plastic Bottle as a Demonstration of Rocket Propulsion
NASA Astrophysics Data System (ADS)
Mattox, J. R.
2017-01-01
I report an innovation that provides a compelling demonstration of rocket propulsion, appropriate for students of physics and other physical sciences. An electrical spark is initiated from a distance to cause the deflagration of a combustible vapor mixed with air in a lightweight plastic bottle that is consequently propelled as a rocket by the release of combustion products, i.e., a "whoosh rocket." My recommendation is that the standard fuel for pedagogical whoosh demonstrations be isopropanol, and the recommended vessel is the 3.8-L high-density polyethylene (HDPE) bottle.
Empirical models of Jupiter's interior from Juno data. Moment of inertia and tidal Love number k2
NASA Astrophysics Data System (ADS)
Ni, Dongdong
2018-05-01
Context. The Juno spacecraft has significantly improved the accuracy of gravitational harmonic coefficients J4, J6 and J8 during its first two perijoves. However, there are still differences in the interior model predictions of core mass and envelope metallicity because of the uncertainties in the hydrogen-helium equations of state. New theoretical approaches or observational data are hence required in order to further constrain the interior models of Jupiter. A well constrained interior model of Jupiter is helpful for understanding not only the dynamic flows in the interior, but also the formation history of giant planets. Aims: We present the radial density profiles of Jupiter fitted to the Juno gravity field observations. Also, we aim to investigate our ability to constrain the core properties of Jupiter using its moment of inertia and tidal Love number k2 which could be accessible by the Juno spacecraft. Methods: In this work, the radial density profile was constrained by the Juno gravity field data within the empirical two-layer model in which the equations of state are not needed as an input model parameter. Different two-layer models are constructed in terms of core properties. The dependence of the calculated moment of inertia and tidal Love number k2 on the core properties was investigated in order to discern their abilities to further constrain the internal structure of Jupiter. Results: The calculated normalized moment of inertia (NMOI) ranges from 0.2749 to 0.2762, in reasonable agreement with the other predictions. There is a good correlation between the NMOI value and the core properties including masses and radii. Therefore, measurements of NMOI by Juno can be used to constrain both the core mass and size of Jupiter's two-layer interior models. For the tidal Love number k2, the degeneracy of k2 is found and analyzed within the two-layer interior model. In spite of this, measurements of k2 can still be used to further constrain the core mass and size of Jupiter's two-layer interior models.
A review of direct numerical simulations of astrophysical detonations and their implications
Parete-Koon, Suzanne T.; Smith, Christopher R.; Papatheodore, Thomas L.; ...
2013-04-11
Multi-dimensional direct numerical simulations (DNS) of astrophysical detonations in degenerate matter have revealed that the nuclear burning is typically characterized by cellular structure caused by transverse instabilities in the detonation front. Type Ia supernova modelers often use one- dimensional DNS of detonations as inputs or constraints for their whole star simulations. While these one-dimensional studies are useful tools, the true nature of the detonation is multi-dimensional. The multi-dimensional structure of the burning influences the speed, stability, and the composition of the detonation and its burning products, and therefore, could have an impact on the spectra of Type Ia supernovae. Considerablemore » effort has been expended modeling Type Ia supernovae at densities above 1x10 7 g∙cm -3 where the complexities of turbulent burning dominate the flame propagation. However, most full star models turn the nuclear burning schemes off when the density falls below 1x10 7 g∙cm -3 and distributed burning begins. The deflagration to detonation transition (DDT) is believed to occur at just these densities and consequently they are the densities important for studying the properties of the subsequent detonation. In conclusion, this work reviews the status of DNS studies of detonations and their possible implications for Type Ia supernova models. It will cover the development of Detonation theory from the first simple Chapman-Jouguet (CJ) detonation models to the current models based on the time-dependent, compressible, reactive flow Euler equations of fluid dynamics.« less
ERIC Educational Resources Information Center
Mare, Robert D.; Mason, William M.
An important class of applications of measurement error or constrained factor analytic models consists of comparing models for several populations. In such cases, it is appropriate to make explicit statistical tests of model similarity across groups and to constrain some parameters of the models to be equal across groups using a priori substantive…
Order-Constrained Bayes Inference for Dichotomous Models of Unidimensional Nonparametric IRT
ERIC Educational Resources Information Center
Karabatsos, George; Sheu, Ching-Fan
2004-01-01
This study introduces an order-constrained Bayes inference framework useful for analyzing data containing dichotomous scored item responses, under the assumptions of either the monotone homogeneity model or the double monotonicity model of nonparametric item response theory (NIRT). The framework involves the implementation of Gibbs sampling to…
Methodology for assessing the safety of Hydrogen Systems: HyRAM 1.1 technical reference manual
DOE Office of Scientific and Technical Information (OSTI.GOV)
Groth, Katrina; Hecht, Ethan; Reynolds, John Thomas
The HyRAM software toolkit provides a basis for conducting quantitative risk assessment and consequence modeling for hydrogen infrastructure and transportation systems. HyRAM is designed to facilitate the use of state-of-the-art science and engineering models to conduct robust, repeatable assessments of hydrogen safety, hazards, and risk. HyRAM is envisioned as a unifying platform combining validated, analytical models of hydrogen behavior, a stan- dardized, transparent QRA approach, and engineering models and generic data for hydrogen installations. HyRAM is being developed at Sandia National Laboratories for the U. S. De- partment of Energy to increase access to technical data about hydrogen safety andmore » to enable the use of that data to support development and revision of national and international codes and standards. This document provides a description of the methodology and models contained in the HyRAM version 1.1. HyRAM 1.1 includes generic probabilities for hydrogen equipment fail- ures, probabilistic models for the impact of heat flux on humans and structures, and computa- tionally and experimentally validated analytical and first order models of hydrogen release and flame physics. HyRAM 1.1 integrates deterministic and probabilistic models for quantifying accident scenarios, predicting physical effects, and characterizing hydrogen hazards (thermal effects from jet fires, overpressure effects from deflagrations), and assessing impact on people and structures. HyRAM is a prototype software in active development and thus the models and data may change. This report will be updated at appropriate developmental intervals.« less
NASA Astrophysics Data System (ADS)
Li, L.; Li, J.; Teo, C. J.; Chang, P. H.; Khoo, B. C.
2018-03-01
The study of deflagration-to-detonation transition (DDT) in bent tubes is important with many potential applications including fuel pipeline and mine tunnel designs for explosion prevention and detonation engines for propulsion. The aim of this study is to exploit low-speed incident shock waves for DDT using an S-shaped geometry and investigate its effectiveness as a DDT enhancement device. Experiments were conducted in a valveless detonation chamber using ethylene-air mixture at room temperature and pressure (303 K, 1 bar). High-speed Schlieren photography was employed to keep track of the wave dynamic evolution. Results showed that waves with velocity as low as 500 m/s can experience a successful DDT process through this S-shaped geometry. To better understand the mechanism, clear images of local explosion processes were captured in either the first curved section or the second curved section depending on the inlet wave velocity, thus proving that this S-shaped tube can act as a two-stage device for DDT. Owing to the curved wall structure, the passing wave was observed to undergo a continuous compression phase which could ignite the local unburnt mixture and finally lead to a local explosion and a detonation transition. Additionally, the phenomenon of shock-vortex interaction near the wave diffraction region was also found to play an important role in the whole process. It was recorded that this interaction could not only result in local head-on reflection of the reflected wave on the wall that could ignite the local mixture, and it could also contribute to the recoupling of the shock-flame complex when a detonation wave is successfully formed in the first curved section.
A Method to Constrain Mass and Spin of GRB Black Holes within the NDAF Model
NASA Astrophysics Data System (ADS)
Liu, Tong; Xue, Li; Zhao, Xiao-Hong; Zhang, Fu-Wen; Zhang, Bing
2016-04-01
Black holes (BHs) hide themselves behind various astronomical phenomena and their properties, I.e., mass and spin, are usually difficult to constrain. One leading candidate for the central engine model of gamma-ray bursts (GRBs) invokes a stellar mass BH and a neutrino-dominated accretion flow (NDAF), with the relativistic jet launched due to neutrino-anti-neutrino annihilations. Such a model gives rise to a matter-dominated fireball, and is suitable to interpret GRBs with a dominant thermal component with a photospheric origin. We propose a method to constrain BH mass and spin within the framework of this model and apply the method to the thermally dominant GRB 101219B, whose initial jet launching radius, r0, is constrained from the data. Using our numerical model of NDAF jets, we estimate the following constraints on the central BH: mass MBH ˜ 5-9 M⊙, spin parameter a* ≳ 0.6, and disk mass 3 M⊙ ≲ Mdisk ≲ 4 M⊙. Our results also suggest that the NDAF model is a competitive candidate for the central engine of GRBs with a strong thermal component.
NASA Astrophysics Data System (ADS)
Elkhateeb, Esraa
2018-01-01
We consider a cosmological model based on a generalization of the equation of state proposed by Nojiri and Odintsov (2004) and Štefančić (2005, 2006). We argue that this model works as a dark fluid model which can interpolate between dust equation of state and the dark energy equation of state. We show how the asymptotic behavior of the equation of state constrained the parameters of the model. The causality condition for the model is also studied to constrain the parameters and the fixed points are tested to determine different solution classes. Observations of Hubble diagram of SNe Ia supernovae are used to further constrain the model. We present an exact solution of the model and calculate the luminosity distance and the energy density evolution. We also calculate the deceleration parameter to test the state of the universe expansion.
The Three-Dimensional Expansion of the Ejecta from Tycho's Supernova Remnant
NASA Technical Reports Server (NTRS)
Williams, Brian J.; Coyle, Nina M.; Yamaguchi, Hiroya; Depasquale, Joseph; Seitenzahl, Ivo R.; Hewitt, John W.; Blondin, John M.; Borkowski, Kazimierz J.; Ghavamian, Parviz; Petre, Robert;
2017-01-01
We present the first 3D measurements of the velocity of various ejecta knots in Tycho's supernova remnant, known to result from a Type Ia explosion. Chandra X-ray observations over a 12 yr baseline from 2003 to 2015 allow us to measure the proper motion of nearly 60 tufts of Si-rich ejecta, giving us the velocity in the plane of the sky. For the line-of-sight velocity, we use two different methods: a nonequilibrium ionization model fit to the strong Si and S lines in the 1.22.8 keV regime, and a fit consisting of a series of Gaussian lines. These methods give consistent results, allowing us to determine the redshift or blueshift of each of the knots. Assuming a distance of 3.5 kpc, we find total velocities that range from 2400 to 6600 km/s, with a mean of 4430 km/s. We find several regions where the ejecta knots have overtaken the forward shock. These regions have proper motions in excess of 6000 km/s. Some SN Ia explosion models predict a velocity asymmetry in the ejecta. We find no such velocity asymmetries in Tycho, and we discuss our findings in light of various explosion models, favoring those delayed-detonation models with relatively vigorous and symmetrical deflagrations. Finally, we compare measurements with models of the remnant's evolution that include both smooth and clumpy ejecta profiles, finding that both ejecta profiles can be accommodated by the observations.
On meeting capital requirements with a chance-constrained optimization model.
Atta Mills, Ebenezer Fiifi Emire; Yu, Bo; Gu, Lanlan
2016-01-01
This paper deals with a capital to risk asset ratio chance-constrained optimization model in the presence of loans, treasury bill, fixed assets and non-interest earning assets. To model the dynamics of loans, we introduce a modified CreditMetrics approach. This leads to development of a deterministic convex counterpart of capital to risk asset ratio chance constraint. We pursue the scope of analyzing our model under the worst-case scenario i.e. loan default. The theoretical model is analyzed by applying numerical procedures, in order to administer valuable insights from a financial outlook. Our results suggest that, our capital to risk asset ratio chance-constrained optimization model guarantees banks of meeting capital requirements of Basel III with a likelihood of 95 % irrespective of changes in future market value of assets.
Constraining new physics models with isotope shift spectroscopy
NASA Astrophysics Data System (ADS)
Frugiuele, Claudia; Fuchs, Elina; Perez, Gilad; Schlaffer, Matthias
2017-07-01
Isotope shifts of transition frequencies in atoms constrain generic long- and intermediate-range interactions. We focus on new physics scenarios that can be most strongly constrained by King linearity violation such as models with B -L vector bosons, the Higgs portal, and chameleon models. With the anticipated precision, King linearity violation has the potential to set the strongest laboratory bounds on these models in some regions of parameter space. Furthermore, we show that this method can probe the couplings relevant for the protophobic interpretation of the recently reported Be anomaly. We extend the formalism to include an arbitrary number of transitions and isotope pairs and fit the new physics coupling to the currently available isotope shift measurements.
Transient/structural analysis of a combustor under explosive loads
NASA Technical Reports Server (NTRS)
Gregory, Peyton B.; Holland, Anne D.
1992-01-01
The 8-Foot High Temperature Tunnel (HTT) at NASA Langley Research Center is a combustion-driven blow-down wind tunnel. A major potential failure mode that was considered during the combustor redesign was the possibility of a deflagration and/or detonation in the combustor. If a main burner flame-out were to occur, then unburned fuel gases could accumulate and, if reignited, an explosion could occur. An analysis has been performed to determine the safe operating limits of the combustor under transient explosive loads. The failure criteria was defined and the failure mechanisms were determined for both peak pressures and differential pressure loadings. An overview of the gas dynamics analysis was given. A finite element model was constructed to evaluate 13 transient load cases. The sensitivity of the structure to the frequency content of the transient loading was assessed. In addition, two closed form dynamic analyses were conducted to verify the finite element analysis. It was determined that the differential pressure load or thrust load was the critical load mechanism and that the nozzle is the weak link in the combustor system.
NASA Astrophysics Data System (ADS)
Glaze, L. S.; Baloga, S. M.; Garvin, J. B.; Quick, L. C.
2014-05-01
Lava flows and flow fields on Venus lack sufficient topographic data for any type of quantitative modeling to estimate eruption rates and durations. Such modeling can constrain rates of resurfacing and provide insights into magma plumbing systems.
NASA Astrophysics Data System (ADS)
Frolov, S. M.; Dubrovskii, A. V.; Ivanov, V. S.
2016-07-01
The possibility of integrating the Continuous Detonation Chamber (CDC) in a gas turbine engine (GTE) is demonstrated by means of three-dimensional (3D) numerical simulations, i. e., the feasibility of the operation process in the annular combustion chamber with a wide gap and with separate feeding of fuel (hydrogen) and oxidizer (air) is proved computationally. The CDC with an upstream isolator damping pressure disturbances propagating towards the compressor is shown to exhibit a gain in the total pressure of 15% as compared with the same combustion chamber operating in the deflagration mode.
NASA Astrophysics Data System (ADS)
Volk, Brent L.; Lagoudas, Dimitris C.; Maitland, Duncan J.
2011-09-01
In this work, tensile tests and one-dimensional constitutive modeling were performed on a high recovery force polyurethane shape memory polymer that is being considered for biomedical applications. The tensile tests investigated the free recovery (zero load) response as well as the constrained displacement recovery (stress recovery) response at extension values up to 25%, and two consecutive cycles were performed during each test. The material was observed to recover 100% of the applied deformation when heated at zero load in the second thermomechanical cycle, and a stress recovery of 1.5-4.2 MPa was observed for the constrained displacement recovery experiments. After the experiments were performed, the Chen and Lagoudas model was used to simulate and predict the experimental results. The material properties used in the constitutive model—namely the coefficients of thermal expansion, shear moduli, and frozen volume fraction—were calibrated from a single 10% extension free recovery experiment. The model was then used to predict the material response for the remaining free recovery and constrained displacement recovery experiments. The model predictions match well with the experimental data.
Method and system to estimate variables in an integrated gasification combined cycle (IGCC) plant
Kumar, Aditya; Shi, Ruijie; Dokucu, Mustafa
2013-09-17
System and method to estimate variables in an integrated gasification combined cycle (IGCC) plant are provided. The system includes a sensor suite to measure respective plant input and output variables. An extended Kalman filter (EKF) receives sensed plant input variables and includes a dynamic model to generate a plurality of plant state estimates and a covariance matrix for the state estimates. A preemptive-constraining processor is configured to preemptively constrain the state estimates and covariance matrix to be free of constraint violations. A measurement-correction processor may be configured to correct constrained state estimates and a constrained covariance matrix based on processing of sensed plant output variables. The measurement-correction processor is coupled to update the dynamic model with corrected state estimates and a corrected covariance matrix. The updated dynamic model may be configured to estimate values for at least one plant variable not originally sensed by the sensor suite.
Building fast well-balanced two-stage numerical schemes for a model of two-phase flows
NASA Astrophysics Data System (ADS)
Thanh, Mai Duc
2014-06-01
We present a set of well-balanced two-stage schemes for an isentropic model of two-phase flows arisen from the modeling of deflagration-to-detonation transition in granular materials. The first stage is to absorb the source term in nonconservative form into equilibria. Then in the second stage, these equilibria will be composed into a numerical flux formed by using a convex combination of the numerical flux of a stable Lax-Friedrichs-type scheme and the one of a higher-order Richtmyer-type scheme. Numerical schemes constructed in such a way are expected to get the interesting property: they are fast and stable. Tests show that the method works out until the parameter takes on the value CFL, and so any value of the parameter between zero and this value is expected to work as well. All the schemes in this family are shown to capture stationary waves and preserves the positivity of the volume fractions. The special values of the parameter 0,1/2,1/(1+CFL), and CFL in this family define the Lax-Friedrichs-type, FAST1, FAST2, and FAST3 schemes, respectively. These schemes are shown to give a desirable accuracy. The errors and the CPU time of these schemes and the Roe-type scheme are calculated and compared. The constructed schemes are shown to be well-balanced and faster than the Roe-type scheme.
The added value of remote sensing products in constraining hydrological models
NASA Astrophysics Data System (ADS)
Nijzink, Remko C.; Almeida, Susana; Pechlivanidis, Ilias; Capell, René; Gustafsson, David; Arheimer, Berit; Freer, Jim; Han, Dawei; Wagener, Thorsten; Sleziak, Patrik; Parajka, Juraj; Savenije, Hubert; Hrachowitz, Markus
2017-04-01
The calibration of a hydrological model still depends on the availability of streamflow data, even though more additional sources of information (i.e. remote sensed data products) have become more widely available. In this research, the model parameters of four different conceptual hydrological models (HYPE, HYMOD, TUW, FLEX) were constrained with remotely sensed products. The models were applied over 27 catchments across Europe to cover a wide range of climates, vegetation and landscapes. The fluxes and states of the models were correlated with the relevant products (e.g. MOD10A snow with modelled snow states), after which new a-posteriori parameter distributions were determined based on a weighting procedure using conditional probabilities. Briefly, each parameter was weighted with the coefficient of determination of the relevant regression between modelled states/fluxes and products. In this way, final feasible parameter sets were derived without the use of discharge time series. Initial results show that improvements in model performance, with regard to streamflow simulations, are obtained when the models are constrained with a set of remotely sensed products simultaneously. In addition, we present a more extensive analysis to assess a model's ability to reproduce a set of hydrological signatures, such as rising limb density or peak distribution. Eventually, this research will enhance our understanding and recommendations in the use of remotely sensed products for constraining conceptual hydrological modelling and improving predictive capability, especially for data sparse regions.
Yeast 5 – an expanded reconstruction of the Saccharomyces cerevisiae metabolic network
2012-01-01
Background Efforts to improve the computational reconstruction of the Saccharomyces cerevisiae biochemical reaction network and to refine the stoichiometrically constrained metabolic models that can be derived from such a reconstruction have continued since the first stoichiometrically constrained yeast genome scale metabolic model was published in 2003. Continuing this ongoing process, we have constructed an update to the Yeast Consensus Reconstruction, Yeast 5. The Yeast Consensus Reconstruction is a product of efforts to forge a community-based reconstruction emphasizing standards compliance and biochemical accuracy via evidence-based selection of reactions. It draws upon models published by a variety of independent research groups as well as information obtained from biochemical databases and primary literature. Results Yeast 5 refines the biochemical reactions included in the reconstruction, particularly reactions involved in sphingolipid metabolism; updates gene-reaction annotations; and emphasizes the distinction between reconstruction and stoichiometrically constrained model. Although it was not a primary goal, this update also improves the accuracy of model prediction of viability and auxotrophy phenotypes and increases the number of epistatic interactions. This update maintains an emphasis on standards compliance, unambiguous metabolite naming, and computer-readable annotations available through a structured document format. Additionally, we have developed MATLAB scripts to evaluate the model’s predictive accuracy and to demonstrate basic model applications such as simulating aerobic and anaerobic growth. These scripts, which provide an independent tool for evaluating the performance of various stoichiometrically constrained yeast metabolic models using flux balance analysis, are included as Additional files 1, 2 and 3. Conclusions Yeast 5 expands and refines the computational reconstruction of yeast metabolism and improves the predictive accuracy of a stoichiometrically constrained yeast metabolic model. It differs from previous reconstructions and models by emphasizing the distinction between the yeast metabolic reconstruction and the stoichiometrically constrained model, and makes both available as Additional file 4 and Additional file 5 and at http://yeast.sf.net/ as separate systems biology markup language (SBML) files. Through this separation, we intend to make the modeling process more accessible, explicit, transparent, and reproducible. PMID:22663945
Y. He; Q. Zhuang; A.D. McGuire; Y. Liu; M. Chen
2013-01-01
Model-data fusion is a process in which field observations are used to constrain model parameters. How observations are used to constrain parameters has a direct impact on the carbon cycle dynamics simulated by ecosystem models. In this study, we present an evaluation of several options for the use of observations inmodeling regional carbon dynamics and explore the...
NASA Astrophysics Data System (ADS)
Suparman, Yusep; Folmer, Henk; Oud, Johan H. L.
2014-01-01
Omitted variables and measurement errors in explanatory variables frequently occur in hedonic price models. Ignoring these problems leads to biased estimators. In this paper, we develop a constrained autoregression-structural equation model (ASEM) to handle both types of problems. Standard panel data models to handle omitted variables bias are based on the assumption that the omitted variables are time-invariant. ASEM allows handling of both time-varying and time-invariant omitted variables by constrained autoregression. In the case of measurement error, standard approaches require additional external information which is usually difficult to obtain. ASEM exploits the fact that panel data are repeatedly measured which allows decomposing the variance of a variable into the true variance and the variance due to measurement error. We apply ASEM to estimate a hedonic housing model for urban Indonesia. To get insight into the consequences of measurement error and omitted variables, we compare the ASEM estimates with the outcomes of (1) a standard SEM, which does not account for omitted variables, (2) a constrained autoregression model, which does not account for measurement error, and (3) a fixed effects hedonic model, which ignores measurement error and time-varying omitted variables. The differences between the ASEM estimates and the outcomes of the three alternative approaches are substantial.
Yu Wei; Michael Bevers; Erin Belval; Benjamin Bird
2015-01-01
This research developed a chance-constrained two-stage stochastic programming model to support wildfire initial attack resource acquisition and location on a planning unit for a fire season. Fire growth constraints account for the interaction between fire perimeter growth and construction to prevent overestimation of resource requirements. We used this model to examine...
Volk, Brent L; Lagoudas, Dimitris C; Maitland, Duncan J
2011-01-01
In this work, tensile tests and one-dimensional constitutive modeling are performed on a high recovery force polyurethane shape memory polymer that is being considered for biomedical applications. The tensile tests investigate the free recovery (zero load) response as well as the constrained displacement recovery (stress recovery) response at extension values up to 25%, and two consecutive cycles are performed during each test. The material is observed to recover 100% of the applied deformation when heated at zero load in the second thermomechanical cycle, and a stress recovery of 1.5 MPa to 4.2 MPa is observed for the constrained displacement recovery experiments. After performing the experiments, the Chen and Lagoudas model is used to simulate and predict the experimental results. The material properties used in the constitutive model – namely the coefficients of thermal expansion, shear moduli, and frozen volume fraction – are calibrated from a single 10% extension free recovery experiment. The model is then used to predict the material response for the remaining free recovery and constrained displacement recovery experiments. The model predictions match well with the experimental data. PMID:22003272
A chance-constrained stochastic approach to intermodal container routing problems.
Zhao, Yi; Liu, Ronghui; Zhang, Xi; Whiteing, Anthony
2018-01-01
We consider a container routing problem with stochastic time variables in a sea-rail intermodal transportation system. The problem is formulated as a binary integer chance-constrained programming model including stochastic travel times and stochastic transfer time, with the objective of minimising the expected total cost. Two chance constraints are proposed to ensure that the container service satisfies ship fulfilment and cargo on-time delivery with pre-specified probabilities. A hybrid heuristic algorithm is employed to solve the binary integer chance-constrained programming model. Two case studies are conducted to demonstrate the feasibility of the proposed model and to analyse the impact of stochastic variables and chance-constraints on the optimal solution and total cost.
A chance-constrained stochastic approach to intermodal container routing problems
Zhao, Yi; Zhang, Xi; Whiteing, Anthony
2018-01-01
We consider a container routing problem with stochastic time variables in a sea-rail intermodal transportation system. The problem is formulated as a binary integer chance-constrained programming model including stochastic travel times and stochastic transfer time, with the objective of minimising the expected total cost. Two chance constraints are proposed to ensure that the container service satisfies ship fulfilment and cargo on-time delivery with pre-specified probabilities. A hybrid heuristic algorithm is employed to solve the binary integer chance-constrained programming model. Two case studies are conducted to demonstrate the feasibility of the proposed model and to analyse the impact of stochastic variables and chance-constraints on the optimal solution and total cost. PMID:29438389
Hydrologic and hydraulic flood forecasting constrained by remote sensing data
NASA Astrophysics Data System (ADS)
Li, Y.; Grimaldi, S.; Pauwels, V. R. N.; Walker, J. P.; Wright, A. J.
2017-12-01
Flooding is one of the most destructive natural disasters, resulting in many deaths and billions of dollars of damages each year. An indispensable tool to mitigate the effect of floods is to provide accurate and timely forecasts. An operational flood forecasting system typically consists of a hydrologic model, converting rainfall data into flood volumes entering the river system, and a hydraulic model, converting these flood volumes into water levels and flood extents. Such a system is prone to various sources of uncertainties from the initial conditions, meteorological forcing, topographic data, model parameters and model structure. To reduce those uncertainties, current forecasting systems are typically calibrated and/or updated using ground-based streamflow measurements, and such applications are limited to well-gauged areas. The recent increasing availability of spatially distributed remote sensing (RS) data offers new opportunities to improve flood forecasting skill. Based on an Australian case study, this presentation will discuss the use of 1) RS soil moisture to constrain a hydrologic model, and 2) RS flood extent and level to constrain a hydraulic model.The GRKAL hydrological model is calibrated through a joint calibration scheme using both ground-based streamflow and RS soil moisture observations. A lag-aware data assimilation approach is tested through a set of synthetic experiments to integrate RS soil moisture to constrain the streamflow forecasting in real-time.The hydraulic model is LISFLOOD-FP which solves the 2-dimensional inertial approximation of the Shallow Water Equations. Gauged water level time series and RS-derived flood extent and levels are used to apply a multi-objective calibration protocol. The effectiveness with which each data source or combination of data sources constrained the parameter space will be discussed.
NASA Astrophysics Data System (ADS)
Burrage, Clare; Sakstein, Jeremy
2018-03-01
Theories of modified gravity, where light scalars with non-trivial self-interactions and non-minimal couplings to matter—chameleon and symmetron theories—dynamically suppress deviations from general relativity in the solar system. On other scales, the environmental nature of the screening means that such scalars may be relevant. The highly-nonlinear nature of screening mechanisms means that they evade classical fifth-force searches, and there has been an intense effort towards designing new and novel tests to probe them, both in the laboratory and using astrophysical objects, and by reinterpreting existing datasets. The results of these searches are often presented using different parametrizations, which can make it difficult to compare constraints coming from different probes. The purpose of this review is to summarize the present state-of-the-art searches for screened scalars coupled to matter, and to translate the current bounds into a single parametrization to survey the state of the models. Presently, commonly studied chameleon models are well-constrained but less commonly studied models have large regions of parameter space that are still viable. Symmetron models are constrained well by astrophysical and laboratory tests, but there is a desert separating the two scales where the model is unconstrained. The coupling of chameleons to photons is tightly constrained but the symmetron coupling has yet to be explored. We also summarize the current bounds on f( R) models that exhibit the chameleon mechanism (Hu and Sawicki models). The simplest of these are well constrained by astrophysical probes, but there are currently few reported bounds for theories with higher powers of R. The review ends by discussing the future prospects for constraining screened modified gravity models further using upcoming and planned experiments.
Dark matter, constrained minimal supersymmetric standard model, and lattice QCD.
Giedt, Joel; Thomas, Anthony W; Young, Ross D
2009-11-13
Recent lattice measurements have given accurate estimates of the quark condensates in the proton. We use these results to significantly improve the dark matter predictions in benchmark models within the constrained minimal supersymmetric standard model. The predicted spin-independent cross sections are at least an order of magnitude smaller than previously suggested and our results have significant consequences for dark matter searches.
NASA Technical Reports Server (NTRS)
Abercromby, Kira J.; Rapp, Jason; Bedard, Donald; Seitzer, Patrick; Cardona, Tommaso; Cowardin, Heather; Barker, Ed; Lederer, Susan
2013-01-01
Constrained Linear Least Squares model is generally more accurate than the "human-in-the-loop". However, "human-in-the-loop" can remove materials that make no sense. The speed of the model in determining a "first cut" at the material ID makes it a viable option for spectral unmixing of debris objects.
NASA Astrophysics Data System (ADS)
Quesada-Montano, Beatriz; Westerberg, Ida K.; Fuentes-Andino, Diana; Hidalgo-Leon, Hugo; Halldin, Sven
2017-04-01
Long-term hydrological data are key to understanding catchment behaviour and for decision making within water management and planning. Given the lack of observed data in many regions worldwide, hydrological models are an alternative for reproducing historical streamflow series. Additional types of information - to locally observed discharge - can be used to constrain model parameter uncertainty for ungauged catchments. Climate variability exerts a strong influence on streamflow variability on long and short time scales, in particular in the Central-American region. We therefore explored the use of climate variability knowledge to constrain the simulated discharge uncertainty of a conceptual hydrological model applied to a Costa Rican catchment, assumed to be ungauged. To reduce model uncertainty we first rejected parameter relationships that disagreed with our understanding of the system. We then assessed how well climate-based constraints applied at long-term, inter-annual and intra-annual time scales could constrain model uncertainty. Finally, we compared the climate-based constraints to a constraint on low-flow statistics based on information obtained from global maps. We evaluated our method in terms of the ability of the model to reproduce the observed hydrograph and the active catchment processes in terms of two efficiency measures, a statistical consistency measure, a spread measure and 17 hydrological signatures. We found that climate variability knowledge was useful for reducing model uncertainty, in particular, unrealistic representation of deep groundwater processes. The constraints based on global maps of low-flow statistics provided more constraining information than those based on climate variability, but the latter rejected slow rainfall-runoff representations that the low flow statistics did not reject. The use of such knowledge, together with information on low-flow statistics and constraints on parameter relationships showed to be useful to constrain model uncertainty for an - assumed to be - ungauged basin. This shows that our method is promising for reconstructing long-term flow data for ungauged catchments on the Pacific side of Central America, and that similar methods can be developed for ungauged basins in other regions where climate variability exerts a strong control on streamflow variability.
Wu, Sheng; Jin, Qibing; Zhang, Ridong; Zhang, Junfeng; Gao, Furong
2017-07-01
In this paper, an improved constrained tracking control design is proposed for batch processes under uncertainties. A new process model that facilitates process state and tracking error augmentation with further additional tuning is first proposed. Then a subsequent controller design is formulated using robust stable constrained MPC optimization. Unlike conventional robust model predictive control (MPC), the proposed method enables the controller design to bear more degrees of tuning so that improved tracking control can be acquired, which is very important since uncertainties exist inevitably in practice and cause model/plant mismatches. An injection molding process is introduced to illustrate the effectiveness of the proposed MPC approach in comparison with conventional robust MPC. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Moorkamp, M.; Fishwick, S.; Jones, A. G.
2015-12-01
Typical surface wave tomography can recover well the velocity structure of the upper mantle in the depth range between 70-200km. For a successful inversion, we have to constrain the crustal structure and assess the impact on the resulting models. In addition,we often observe potentially interesting features in the uppermost lithosphere which are poorly resolved and thus their interpretationhas to be approached with great care.We are currently developing a seismically constrained magnetotelluric (MT) inversion approach with the aim of better recovering the lithospheric properties (and thus seismic velocities) in these problematic areas. We perform a 3D MT inversion constrained by a fixed seismic velocity model from surface wave tomography. In order to avoid strong bias, we only utilize information on structural boundaries to combine these two methods. Within the region that is well resolved by both methods, we can then extract a velocity-conductivity relationship. By translating the conductivitiesretrieved from MT into velocities in areas where the velocity model is poorly resolved, we can generate an updated velocity model and test what impactthe updated velocities have on the predicted data.We test this new approach using a MT dataset acquired in central Botswana over the Okwa terrane and the adjacent Kaapvaal and Zimbabwe Cratons togetherwith a tomographic models for the region. Here, both datasets have previously been used to constrain lithospheric structure and show some similarities.We carefully asses the validity of our results by comparing with observations and petrophysical predictions for the conductivity-velocity relationship.
Periodic Forced Response of Structures Having Three-Dimensional Frictional Constraints
NASA Astrophysics Data System (ADS)
CHEN, J. J.; YANG, B. D.; MENQ, C. H.
2000-01-01
Many mechanical systems have moving components that are mutually constrained through frictional contacts. When subjected to cyclic excitations, a contact interface may undergo constant changes among sticks, slips and separations, which leads to very complex contact kinematics. In this paper, a 3-D friction contact model is employed to predict the periodic forced response of structures having 3-D frictional constraints. Analytical criteria based on this friction contact model are used to determine the transitions among sticks, slips and separations of the friction contact, and subsequently the constrained force which consists of the induced stick-slip friction force on the contact plane and the contact normal load. The resulting constrained force is often a periodic function and can be considered as a feedback force that influences the response of the constrained structures. By using the Multi-Harmonic Balance Method along with Fast Fourier Transform, the constrained force can be integrated with the receptance of the structures so as to calculate the forced response of the constrained structures. It results in a set of non-linear algebraic equations that can be solved iteratively to yield the relative motion as well as the constrained force at the friction contact. This method is used to predict the periodic response of a frictionally constrained 3-d.o.f. oscillator. The predicted results are compared with those of the direct time integration method so as to validate the proposed method. In addition, the effect of super-harmonic components on the resonant response and jump phenomenon is examined.
NASA Astrophysics Data System (ADS)
Pan, M.; Wood, E. F.
2004-05-01
This study explores a method to estimate various components of the water cycle (ET, runoff, land storage, etc.) based on a number of different info sources, including both observations and observation-enhanced model simulations. Different from existing data assimilations, this constrained Kalman filtering approach keeps the water budget perfectly closed while updating the states of the underlying model (VIC model) optimally using observations. Assimilating different data sources in this way has several advantages: (1) physical model is included to make estimation time series smooth, missing-free, and more physically consistent; (2) uncertainties in the model and observations are properly addressed; (3) model is constrained by observation thus to reduce model biases; (4) balance of water is always preserved along the assimilation. Experiments are carried out in Southern Great Plain region where necessary observations have been collected. This method may also be implemented in other applications with physical constraints (e.g. energy cycles) and at different scales.
Guenole, Nigel
2018-01-01
The test for item level cluster bias examines the improvement in model fit that results from freeing an item's between level residual variance from a baseline model with equal within and between level factor loadings and between level residual variances fixed at zero. A potential problem is that this approach may include a misspecified unrestricted model if any non-invariance is present, but the log-likelihood difference test requires that the unrestricted model is correctly specified. A free baseline approach where the unrestricted model includes only the restrictions needed for model identification should lead to better decision accuracy, but no studies have examined this yet. We ran a Monte Carlo study to investigate this issue. When the referent item is unbiased, compared to the free baseline approach, the constrained baseline approach led to similar true positive (power) rates but much higher false positive (Type I error) rates. The free baseline approach should be preferred when the referent indicator is unbiased. When the referent assumption is violated, the false positive rate was unacceptably high for both free and constrained baseline approaches, and the true positive rate was poor regardless of whether the free or constrained baseline approach was used. Neither the free or constrained baseline approach can be recommended when the referent indicator is biased. We recommend paying close attention to ensuring the referent indicator is unbiased in tests of cluster bias. All Mplus input and output files, R, and short Python scripts used to execute this simulation study are uploaded to an open access repository.
Guenole, Nigel
2018-01-01
The test for item level cluster bias examines the improvement in model fit that results from freeing an item's between level residual variance from a baseline model with equal within and between level factor loadings and between level residual variances fixed at zero. A potential problem is that this approach may include a misspecified unrestricted model if any non-invariance is present, but the log-likelihood difference test requires that the unrestricted model is correctly specified. A free baseline approach where the unrestricted model includes only the restrictions needed for model identification should lead to better decision accuracy, but no studies have examined this yet. We ran a Monte Carlo study to investigate this issue. When the referent item is unbiased, compared to the free baseline approach, the constrained baseline approach led to similar true positive (power) rates but much higher false positive (Type I error) rates. The free baseline approach should be preferred when the referent indicator is unbiased. When the referent assumption is violated, the false positive rate was unacceptably high for both free and constrained baseline approaches, and the true positive rate was poor regardless of whether the free or constrained baseline approach was used. Neither the free or constrained baseline approach can be recommended when the referent indicator is biased. We recommend paying close attention to ensuring the referent indicator is unbiased in tests of cluster bias. All Mplus input and output files, R, and short Python scripts used to execute this simulation study are uploaded to an open access repository. PMID:29551985
Late-time spectra and type Ia supernova models: New clues from the Hubble Space Telescope
NASA Technical Reports Server (NTRS)
Ruiz-Lapuente, P.; Kirshner, R. P.; Phillips, M. M.; Challis, P. M.; Schmidt, B. P.; Filippenko, A. V.; Wheeler, J. C.
1995-01-01
Calculated late-time spectra of two classical hydrodynamical models for Type Ia supernovae (deflagration model W7 of Nomoto, Thielemann, & Yokoi, and delayed detonation model DD4 of Woosley & Weaver) are compared with observations of SN 1992A and other spectroscopically normal SNe Ia. An important new piece of information is provided by observations done with the Hubble Space Telescope (HST) which cover the ultraviolet range at the nebular phase of a SN Ia: SN 1992A in NGC 1380. For the first time a picture of SN Ia emission from the ultraviolet through the optical is obtained at these phases. Predictions of the classical model (W7 and DD4) are compared with the observed spectrum of SN 1992A and with the optical spectra of SN 1989M in NGC 4579 and SN 1990N in NGC 4639 at similar epochs. The absolute B and V magnitudes of the models are also estimated at these late phases. Taken at face value the nebular spectra of these 'classical' models are more consistent with the long extragalactic distance scale, pointing to distances to NGC 4579 around 21 +/- 3 Mpc and a slightly larger distance, 22 +/- 3 Mpc, to NGC 4639, on the back side of the Virgo Cluster. However, the calculated Fe(+3) luminosity as predicted from the models exceeds the observed limit from the HST data of SN 1992A. Other differences in the ratios of the line intensities between calculated and observed spectra, show some disagreement with the observed spectra at the nebular phases. They may not be the best choice for spectroscopically normal SNe Ia, and their use as an independent calibration of the extragalactic distance scale should be viewed with caution.
Trajectory optimization and guidance law development for national aerospace plane applications
NASA Technical Reports Server (NTRS)
Calise, A. J.; Flandro, G. A.; Corban, J. E.
1988-01-01
The work completed to date is comprised of the following: a simple vehicle model representative of the aerospace plane concept in the hypersonic flight regime, fuel-optimal climb profiles for the unconstrained and dynamic pressure constrained cases generated using a reduced order dynamic model, an analytic switching condition for transition to rocket powered flight as orbital velocity is approached, simple feedback guidance laws for both the unconstrained and dynamic pressure constrained cases derived via singular perturbation theory and a nonlinear transformation technique, and numerical simulation results for ascent to orbit in the dynamic pressure constrained case.
Constraints on Dark Energy from Baryon Acoustic Peak and Galaxy Cluster Gas Mass Measurements
NASA Astrophysics Data System (ADS)
Samushia, Lado; Ratra, Bharat
2009-10-01
We use baryon acoustic peak measurements by Eisenstein et al. and Percival et al., together with the Wilkinson Microwave Anisotropy Probe (WMAP) measurement of the apparent acoustic horizon angle, and galaxy cluster gas mass fraction measurements of Allen et al., to constrain a slowly rolling scalar field dark energy model, phiCDM, in which dark energy's energy density changes in time. We also compare our phiCDM results with those derived for two more common dark energy models: the time-independent cosmological constant model, ΛCDM, and the XCDM parameterization of dark energy's equation of state. For time-independent dark energy, the Percival et al. measurements effectively constrain spatial curvature and favor a close to the spatially flat model, mostly due to the WMAP cosmic microwave background prior used in the analysis. In a spatially flat model the Percival et al. data less effectively constrain time-varying dark energy. The joint baryon acoustic peak and galaxy cluster gas mass constraints on the phiCDM model are consistent with but tighter than those derived from other data. A time-independent cosmological constant in a spatially flat model provides a good fit to the joint data, while the α parameter in the inverse power-law potential phiCDM model is constrained to be less than about 4 at 3σ confidence level.
Critical Robotic Lunar Missions
NASA Astrophysics Data System (ADS)
Plescia, J. B.
2018-04-01
Perhaps the most critical missions to understanding lunar history are in situ dating and network missions. These would constrain the volcanic and thermal history and interior structure. These data would better constrain lunar evolution models.
The Three-dimensional Expansion of the Ejecta from Tycho's Supernova Remnant
DOE Office of Scientific and Technical Information (OSTI.GOV)
Williams, Brian J.; Depasquale, Joseph; Coyle, Nina M.
2017-06-10
We present the first 3D measurements of the velocity of various ejecta knots in Tycho’s supernova remnant, known to result from a Type Ia explosion. Chandra X-ray observations over a 12 yr baseline from 2003 to 2015 allow us to measure the proper motion of nearly 60 “tufts” of Si-rich ejecta, giving us the velocity in the plane of the sky. For the line-of-sight velocity, we use two different methods: a nonequilibrium ionization model fit to the strong Si and S lines in the 1.2–2.8 keV regime, and a fit consisting of a series of Gaussian lines. These methods givemore » consistent results, allowing us to determine the redshift or blueshift of each of the knots. Assuming a distance of 3.5 kpc, we find total velocities that range from 2400 to 6600 km s{sup −1}, with a mean of 4430 km s{sup −1}. We find several regions where the ejecta knots have overtaken the forward shock. These regions have proper motions in excess of 6000 km s{sup −1}. Some SN Ia explosion models predict a velocity asymmetry in the ejecta. We find no such velocity asymmetries in Tycho, and we discuss our findings in light of various explosion models, favoring those delayed-detonation models with relatively vigorous and symmetrical deflagrations. Finally, we compare measurements with models of the remnant’s evolution that include both smooth and clumpy ejecta profiles, finding that both ejecta profiles can be accommodated by the observations.« less
Universally Sloppy Parameter Sensitivities in Systems Biology Models
Gutenkunst, Ryan N; Waterfall, Joshua J; Casey, Fergal P; Brown, Kevin S; Myers, Christopher R; Sethna, James P
2007-01-01
Quantitative computational models play an increasingly important role in modern biology. Such models typically involve many free parameters, and assigning their values is often a substantial obstacle to model development. Directly measuring in vivo biochemical parameters is difficult, and collectively fitting them to other experimental data often yields large parameter uncertainties. Nevertheless, in earlier work we showed in a growth-factor-signaling model that collective fitting could yield well-constrained predictions, even when it left individual parameters very poorly constrained. We also showed that the model had a “sloppy” spectrum of parameter sensitivities, with eigenvalues roughly evenly distributed over many decades. Here we use a collection of models from the literature to test whether such sloppy spectra are common in systems biology. Strikingly, we find that every model we examine has a sloppy spectrum of sensitivities. We also test several consequences of this sloppiness for building predictive models. In particular, sloppiness suggests that collective fits to even large amounts of ideal time-series data will often leave many parameters poorly constrained. Tests over our model collection are consistent with this suggestion. This difficulty with collective fits may seem to argue for direct parameter measurements, but sloppiness also implies that such measurements must be formidably precise and complete to usefully constrain many model predictions. We confirm this implication in our growth-factor-signaling model. Our results suggest that sloppy sensitivity spectra are universal in systems biology models. The prevalence of sloppiness highlights the power of collective fits and suggests that modelers should focus on predictions rather than on parameters. PMID:17922568
Universally sloppy parameter sensitivities in systems biology models.
Gutenkunst, Ryan N; Waterfall, Joshua J; Casey, Fergal P; Brown, Kevin S; Myers, Christopher R; Sethna, James P
2007-10-01
Quantitative computational models play an increasingly important role in modern biology. Such models typically involve many free parameters, and assigning their values is often a substantial obstacle to model development. Directly measuring in vivo biochemical parameters is difficult, and collectively fitting them to other experimental data often yields large parameter uncertainties. Nevertheless, in earlier work we showed in a growth-factor-signaling model that collective fitting could yield well-constrained predictions, even when it left individual parameters very poorly constrained. We also showed that the model had a "sloppy" spectrum of parameter sensitivities, with eigenvalues roughly evenly distributed over many decades. Here we use a collection of models from the literature to test whether such sloppy spectra are common in systems biology. Strikingly, we find that every model we examine has a sloppy spectrum of sensitivities. We also test several consequences of this sloppiness for building predictive models. In particular, sloppiness suggests that collective fits to even large amounts of ideal time-series data will often leave many parameters poorly constrained. Tests over our model collection are consistent with this suggestion. This difficulty with collective fits may seem to argue for direct parameter measurements, but sloppiness also implies that such measurements must be formidably precise and complete to usefully constrain many model predictions. We confirm this implication in our growth-factor-signaling model. Our results suggest that sloppy sensitivity spectra are universal in systems biology models. The prevalence of sloppiness highlights the power of collective fits and suggests that modelers should focus on predictions rather than on parameters.
JPRS report: Science and technology. Central Eurasia: Engineering and equipment
NASA Astrophysics Data System (ADS)
1993-10-01
Translated articles cover the following topics: transient gas dynamic processes in ramjet engines; aerodynamic characteristics of delta wings with detached leading edge shock wave at hypersonic flight velocities; effect of atmospheric density gradient on aerodynamic stabilization; measurement of target radar scattering characteristics using frequency synthesized signals; assessing survivability and ensuring safety of large axial-flow compressor blades; procedure for experimentally determining transient aerodynamic forces caused by flat vane cascade; analysis of aerodynamic interaction of profile and vortex; laser machine for balancing dynamically adjusted gyros; use of heat pumps in solar heat supply systems; numerical simulation of deflagration transition to detonation in homogeneous combustible fuel mixture; and investigation of chemically nonequilibrium flow about bodies allowing for vibrational relaxation.
Testing of a work bench for handling of explosives in the laboratory
NASA Astrophysics Data System (ADS)
Hank, R.; Johansson, K.; Lagman, L.
1981-01-01
A prototype work station was developed at which jobs can be carried out with explosives up to 10 gr and deflagrating products up to 50 gr. Tests were made to investigate the consequences of a spontaneous accident during work. Conclusions are: the workbench offers good protection against splinters provided the inside walls are coated with a shock absorber; the carbonate glass should be a minimum of eight mm thick; the risk of burns, except on arms and hands, is very low; the bench withstands the explosion with the given weight of explosives (10 gr); the risk of lesions on the lung are very low, for the operator as well as for somebody nearby.
Characterization of diacetone diperoxide (DADP)
NASA Astrophysics Data System (ADS)
Bowden, Patrick R.; Tappan, Bryce C.; Manner, Virginia W.; Preston, Daniel N.; Scott, Brian L.
2017-01-01
To date, diacetone diperoxide (DADP) has been significantly less studied than its well-known counterpart, triacetone triperoxide (TATP). Much of this disparity in the literature is due to the harsher conditions/multi-step syntheses required to obtain DADP leading to much lower evidence of frequency of use. Because of this, DADP is often misrepresented as being more dangerous (i.e. more sensitive and less stable) than TATP. This paper discusses the synthesis and characterization (sensitivity, thermal stability, etc.) of DADP with comparisons to other energetic organic peroxides (TATP, HMTD and MEKP) and differences in polymorphism, crystal habit and effects of aging and processing differences are discussed. Additionally, the deflagration-to-detonation transition (DDT) behavior of DADP is discussed with comparison to TATP.
Fluid dynamics of the unsteady two phase processes leading to DDT in granular solid propellants
NASA Technical Reports Server (NTRS)
Krier, H.; Butler, P. B.; Lembeck, M. F.
1980-01-01
Deflagration to Detonation (DDT) was predicted to occur in porous beds of high-energy solid propellants by solving the unsteady fluid mechanical convective heat transfer from hot gas products, obtained from the rapid burning at high pressures, provides the impetus to develop a narrow combustion zone and a resulting strong shock. A parametric study clearly indicates that DDT occurs only when a combination of the solids loading fraction, the burning rate constants, the propellant chemical energy, and the particle size provide for critical energy and gas release to support a detonation wave. Predictions for the run-up length to detonation as a function of these parameters are presented.
Ou, Guoliang; Tan, Shukui; Zhou, Min; Lu, Shasha; Tao, Yinghui; Zhang, Zuo; Zhang, Lu; Yan, Danping; Guan, Xingliang; Wu, Gang
2017-12-15
An interval chance-constrained fuzzy land-use allocation (ICCF-LUA) model is proposed in this study to support solving land resource management problem associated with various environmental and ecological constraints at a watershed level. The ICCF-LUA model is based on the ICCF (interval chance-constrained fuzzy) model which is coupled with interval mathematical model, chance-constrained programming model and fuzzy linear programming model and can be used to deal with uncertainties expressed as intervals, probabilities and fuzzy sets. Therefore, the ICCF-LUA model can reflect the tradeoff between decision makers and land stakeholders, the tradeoff between the economical benefits and eco-environmental demands. The ICCF-LUA model has been applied to the land-use allocation of Wujiang watershed, Guizhou Province, China. The results indicate that under highly land suitable conditions, optimized area of cultivated land, forest land, grass land, construction land, water land, unused land and landfill in Wujiang watershed will be [5015, 5648] hm 2 , [7841, 7965] hm 2 , [1980, 2056] hm 2 , [914, 1423] hm 2 , [70, 90] hm 2 , [50, 70] hm 2 and [3.2, 4.3] hm 2 , the corresponding system economic benefit will be between 6831 and 7219 billion yuan. Consequently, the ICCF-LUA model can effectively support optimized land-use allocation problem in various complicated conditions which include uncertainties, risks, economic objective and eco-environmental constraints. Copyright © 2017 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Wong, T. E.; Noone, D. C.; Kleiber, W.
2014-12-01
The single largest uncertainty in climate model energy balance is the surface latent heating over tropical land. Furthermore, the partitioning of the total latent heat flux into contributions from surface evaporation and plant transpiration is of great importance, but notoriously poorly constrained. Resolving these issues will require better exploiting information which lies at the interface between observations and advanced modeling tools, both of which are imperfect. There are remarkably few observations which can constrain these fluxes, placing strict requirements on developing statistical methods to maximize the use of limited information to best improve models. Previous work has demonstrated the power of incorporating stable water isotopes into land surface models for further constraining ecosystem processes. We present results from a stable water isotopically-enabled land surface model (iCLM4), including model experiments partitioning the latent heat flux into contributions from plant transpiration and surface evaporation. It is shown that the partitioning results are sensitive to the parameterization of kinetic fractionation used. We discuss and demonstrate an approach to calibrating select model parameters to observational data in a Bayesian estimation framework, requiring Markov Chain Monte Carlo sampling of the posterior distribution, which is shown to constrain uncertain parameters as well as inform relevant values for operational use. Finally, we discuss the application of the estimation scheme to iCLM4, including entropy as a measure of information content and specific challenges which arise in calibration models with a large number of parameters.
Minimal models from W-constrained hierarchies via the Kontsevich-Miwa transform
NASA Astrophysics Data System (ADS)
Gato-Rivera, B.; Semikhatov, A. M.
1992-08-01
A direct relation between the conformal formalism for 2D quantum gravity and the W-constrained KP hierarchy is found, without the need to invoke intermediate matrix model technology. The Kontsevich-Miwa transform of the KP hierarchy is used to establish an identification between W constraints on the KP tau function and decoupling equations corresponding to Virasoro null vectors. The Kontsevich-Miwa transform maps the W ( l) -constrained KP hierarchy to the ( p‧, p‧) minimal model, with the tau function being given by the correlator of a product of (dressed) ( l, 1) [or (1, l)] operators, provided the Miwa parameter ni and the free parameter (an abstract bc spin) present in the constraint are expressed through the ratio p‧/ p and the level l.
NASA Astrophysics Data System (ADS)
Reading, A. M.; Staal, T.; Halpin, J.; Whittaker, J. M.; Morse, P. E.
2017-12-01
The lithosphere of East Antarctica is one of the least explored regions of the planet, yet it is gaining in importance in global scientific research. Continental heat flux density and 3D glacial isostatic adjustment studies, for example, rely on a good knowledge of the deep structure in constraining model inputs.In this contribution, we use a multidisciplinary approach to constrain lithospheric domains. To seismic tomography models, we add constraints from magnetic studies and also new geological constraints. Geological knowledge exists around the periphery of East Antarctica and is reinforced in the knowledge of plate tectonic reconstructions. The subglacial geology of the Antarctic hinterland is largely unknown but the plate reconstructions allow the well-posed extrapolation of major terranes into the interior of the continent, guided by the seismic tomography and magnetic images. We find that the northern boundary of the lithospheric domain centred on the Gamburtsev Subglacial Mountains has a possible trend that runs south of the Lambert Glacier region, turning coastward through Wilkes Land. Other periphery-to-interior connections are less well constrained and the possibility of lithospheric domains that are entirely sub-glacial is high. We develop this framework to include a probabilistic method of handling alternate models and quantifiable uncertainties. We also show first results in using a Bayesian approach to predicting lithospheric boundaries from multivariate data.Within the newly constrained domains, we constrain heat flux (density) as the sum of basal heat flux and upper crustal heat flux. The basal heat flux is constrained by geophysical methods while the upper crustal heat flux is constrained by geology or predicted geology. In addition to heat flux constraints, we also consider the variations in friction experienced by moving ice sheets due to varying geology.
NASA Astrophysics Data System (ADS)
Quan, Lulin; Yang, Zhixin
2010-05-01
To address the issues in the area of design customization, this paper expressed the specification and application of the constrained surface deformation, and reported the experimental performance comparison of three prevail effective similarity assessment algorithms on constrained surface deformation domain. Constrained surface deformation becomes a promising method that supports for various downstream applications of customized design. Similarity assessment is regarded as the key technology for inspecting the success of new design via measuring the difference level between the deformed new design and the initial sample model, and indicating whether the difference level is within the limitation. According to our theoretical analysis and pre-experiments, three similarity assessment algorithms are suitable for this domain, including shape histogram based method, skeleton based method, and U system moment based method. We analyze their basic functions and implementation methodologies in detail, and do a series of experiments on various situations to test their accuracy and efficiency using precision-recall diagram. Shoe model is chosen as an industrial example for the experiments. It shows that shape histogram based method gained an optimal performance in comparison. Based on the result, we proposed a novel approach that integrating surface constrains and shape histogram description with adaptive weighting method, which emphasize the role of constrains during the assessment. The limited initial experimental result demonstrated that our algorithm outperforms other three algorithms. A clear direction for future development is also drawn at the end of the paper.
Small-kernel, constrained least-squares restoration of sampled image data
NASA Technical Reports Server (NTRS)
Hazra, Rajeeb; Park, Stephen K.
1992-01-01
Following the work of Park (1989), who extended a derivation of the Wiener filter based on the incomplete discrete/discrete model to a more comprehensive end-to-end continuous/discrete/continuous model, it is shown that a derivation of the constrained least-squares (CLS) filter based on the discrete/discrete model can also be extended to this more comprehensive continuous/discrete/continuous model. This results in an improved CLS restoration filter, which can be efficiently implemented as a small-kernel convolution in the spatial domain.
Identification of different geologic units using fuzzy constrained resistivity tomography
NASA Astrophysics Data System (ADS)
Singh, Anand; Sharma, S. P.
2018-01-01
Different geophysical inversion strategies are utilized as a component of an interpretation process that tries to separate geologic units based on the resistivity distribution. In the present study, we present the results of separating different geologic units using fuzzy constrained resistivity tomography. This was accomplished using fuzzy c means, a clustering procedure to improve the 2D resistivity image and geologic separation within the iterative minimization through inversion. First, we developed a Matlab-based inversion technique to obtain a reliable resistivity image using different geophysical data sets (electrical resistivity and electromagnetic data). Following this, the recovered resistivity model was converted into a fuzzy constrained resistivity model by assigning the highest probability value of each model cell to the cluster utilizing fuzzy c means clustering procedure during the iterative process. The efficacy of the algorithm is demonstrated using three synthetic plane wave electromagnetic data sets and one electrical resistivity field dataset. The presented approach shows improvement on the conventional inversion approach to differentiate between different geologic units if the correct number of geologic units will be identified. Further, fuzzy constrained resistivity tomography was performed to examine the augmentation of uranium mineralization in the Beldih open cast mine as a case study. We also compared geologic units identified by fuzzy constrained resistivity tomography with geologic units interpreted from the borehole information.
How well can future CMB missions constrain cosmic inflation?
DOE Office of Scientific and Technical Information (OSTI.GOV)
Martin, Jérôme; Vennin, Vincent; Ringeval, Christophe, E-mail: jmartin@iap.fr, E-mail: christophe.ringeval@uclouvain.be, E-mail: vennin@iap.fr
2014-10-01
We study how the next generation of Cosmic Microwave Background (CMB) measurement missions (such as EPIC, LiteBIRD, PRISM and COrE) will be able to constrain the inflationary landscape in the hardest to disambiguate situation in which inflation is simply described by single-field slow-roll scenarios. Considering the proposed PRISM and LiteBIRD satellite designs, we simulate mock data corresponding to five different fiducial models having values of the tensor-to-scalar ratio ranging from 10{sup -1} down to 10{sup -7}. We then compute the Bayesian evidences and complexities of all Encyclopædia Inflationaris models in order to assess the constraining power of PRISM alone andmore » LiteBIRD complemented with the Planck 2013 data. Within slow-roll inflation, both designs have comparable constraining power and can rule out about three quarters of the inflationary scenarios, compared to one third for Planck 2013 data alone. However, we also show that PRISM can constrain the scalar running and has the capability to detect a violation of slow roll at second order. Finally, our results suggest that describing an inflationary model by its potential shape only, without specifying a reheating temperature, will no longer be possible given the accuracy level reached by the future CMB missions.« less
NASA Astrophysics Data System (ADS)
Li, Duo; Liu, Yajing
2017-04-01
Along-strike segmentation of slow-slip events (SSEs) and nonvolcanic tremors in Cascadia may reflect heterogeneities of the subducting slab or overlying continental lithosphere. However, the nature behind this segmentation is not fully understood. We develop a 3-D model for episodic SSEs in northern and central Cascadia, incorporating both seismological and gravitational observations to constrain the heterogeneities in the megathrust fault properties. The 6 year automatically detected tremors are used to constrain the rate-state friction parameters. The effective normal stress at SSE depths is constrained by along-margin free-air and Bouguer gravity anomalies. The along-strike variation in the long-term plate convergence rate is also taken into consideration. Simulation results show five segments of ˜Mw6.0 SSEs spontaneously appear along the strike, correlated to the distribution of tremor epicenters. Modeled SSE recurrence intervals are equally comparable to GPS observations using both types of gravity anomaly constraints. However, the model constrained by free-air anomaly does a better job in reproducing the cumulative slip as well as more consistent surface displacements with GPS observations. The modeled along-strike segmentation represents the averaged slip release over many SSE cycles, rather than permanent barriers. Individual slow-slip events can still propagate across the boundaries, which may cause interactions between adjacent SSEs, as observed in time-dependent GPS inversions. In addition, the moment-duration scaling is sensitive to the selection of velocity criteria for determining when SSEs occur. Hence, the detection ability of the current GPS network should be considered in the interpretation of slow earthquake source parameter scaling relations.
NASA Astrophysics Data System (ADS)
SUN, D.; TONG, L.
2002-05-01
A detailed model for the beams with partially debonded active constraining damping (ACLD) treatment is presented. In this model, the transverse displacement of the constraining layer is considered to be non-identical to that of the host structure. In the perfect bonding region, the viscoelastic core is modelled to carry both peel and shear stresses, while in the debonding area, it is assumed that no peel and shear stresses be transferred between the host beam and the constraining layer. The adhesive layer between the piezoelectric sensor and the host beam is also considered in this model. In active control, the positive position feedback control is employed to control the first mode of the beam. Based on this model, the incompatibility of the transverse displacements of the active constraining layer and the host beam is investigated. The passive and active damping behaviors of the ACLD patch with different thicknesses, locations and lengths are examined. Moreover, the effects of debonding of the damping layer on both passive and active control are examined via a simulation example. The results show that the incompatibility of the transverse displacements is remarkable in the regions near the ends of the ACLD patch especially for the high order vibration modes. It is found that a thinner damping layer may lead to larger shear strain and consequently results in a larger passive and active damping. In addition to the thickness of the damping layer, its length and location are also key factors to the hybrid control. The numerical results unveil that edge debonding can lead to a reduction of both passive and active damping, and the hybrid damping may be more sensitive to the debonding of the damping layer than the passive damping.
NASA Astrophysics Data System (ADS)
Wang, Yu; Fan, Jie; Xu, Ye; Sun, Wei; Chen, Dong
2018-05-01
In this study, an inexact log-normal-based stochastic chance-constrained programming model was developed for solving the non-point source pollution issues caused by agricultural activities. Compared to the general stochastic chance-constrained programming model, the main advantage of the proposed model is that it allows random variables to be expressed as a log-normal distribution, rather than a general normal distribution. Possible deviations in solutions caused by irrational parameter assumptions were avoided. The agricultural system management in the Erhai Lake watershed was used as a case study, where critical system factors, including rainfall and runoff amounts, show characteristics of a log-normal distribution. Several interval solutions were obtained under different constraint-satisfaction levels, which were useful in evaluating the trade-off between system economy and reliability. The applied results show that the proposed model could help decision makers to design optimal production patterns under complex uncertainties. The successful application of this model is expected to provide a good example for agricultural management in many other watersheds.
Maximum entropy production: Can it be used to constrain conceptual hydrological models?
M.C. Westhoff; E. Zehe
2013-01-01
In recent years, optimality principles have been proposed to constrain hydrological models. The principle of maximum entropy production (MEP) is one of the proposed principles and is subject of this study. It states that a steady state system is organized in such a way that entropy production is maximized. Although successful applications have been reported in...
The in situ transverse lamina strength of composite laminates
NASA Technical Reports Server (NTRS)
Flaggs, D. L.
1983-01-01
The objective of the work reported in this presentation is to determine the in situ transverse strength of a lamina within a composite laminate. From a fracture mechanics standpoint, in situ strength may be viewed as constrained cracking that has been shown to be a function of both lamina thickness and the stiffness of adjacent plies that serve to constrain the cracking process. From an engineering point of view, however, constrained cracking can be perceived as an apparent increase in lamina strength. With the growing need to design more highly loaded composite structures, the concept of in situ strength may prove to be a viable means of increasing the design allowables of current and future composite material systems. A simplified one dimensional analytical model is presented that is used to predict the strain at onset of transverse cracking. While it is accurate only for the most constrained cases, the model is important in that the predicted failure strain is seen to be a function of a lamina's thickness d and of the extensional stiffness bE theta of the adjacent laminae that constrain crack propagation in the 90 deg laminae.
Wait for It: Post-supernova Winds Driven by Delayed Radioactive Decays
NASA Astrophysics Data System (ADS)
Shen, Ken J.; Schwab, Josiah
2017-01-01
In most astrophysical situations, the radioactive decay of {}56{Ni} to {}56{Co} occurs via electron capture with a fixed half-life of 6.1 days. However, this decay rate is significantly slowed when the nuclei are fully ionized because K-shell electrons are unavailable for capture. In this paper, we explore the effect of these delayed decays on white dwarfs (WDs) that may survive Type Ia and Type Iax supernovae (SNe Ia and SNe Iax). The energy released by the delayed radioactive decays of {}56{Ni} and {}56{Co} drives a persistent wind from the surviving WD’s surface that contributes to the late-time appearance of these SNe after emission from the bulk of the SN ejecta has faded. We use the stellar evolution code MESA to calculate the hydrodynamic evolution and resulting light curves of these winds. Our post-SN Ia models conflict with late-time observations of SN 2011fe, but uncertainties in our initial conditions prevent us from ruling out the existence of surviving WD donors. Much better agreement with observations is achieved with our models of post-SN Iax bound remnants, providing evidence that these explosions are due to deflagrations in accreting WDs that fail to completely unbind the WDs. Future radiative transfer calculations and wind models utilizing simulations of explosions for more accurate initial conditions will extend our study of radioactively powered winds from post-SN surviving WDs and enable their use as powerful discriminants among the various SN Ia and SN Iax progenitor scenarios.
Freezing Transition Studies Through Constrained Cell Model Simulation
NASA Astrophysics Data System (ADS)
Nayhouse, Michael; Kwon, Joseph Sang-Il; Heng, Vincent R.; Amlani, Ankur M.; Orkoulas, G.
2014-10-01
In the present work, a simulation method based on cell models is used to deduce the fluid-solid transition of a system of particles that interact via a pair potential, , which is of the form with . The simulations are implemented under constant-pressure conditions on a generalized version of the constrained cell model. The constrained cell model is constructed by dividing the volume into Wigner-Seitz cells and confining each particle in a single cell. This model is a special case of a more general cell model which is formed by introducing an additional field variable that controls the number of particles per cell and, thus, the relative stability of the solid against the fluid phase. High field values force configurations with one particle per cell and thus favor the solid phase. Fluid-solid coexistence on the isotherm that corresponds to a reduced temperature of 2 is determined from constant-pressure simulations of the generalized cell model using tempering and histogram reweighting techniques. The entire fluid-solid phase boundary is determined through a thermodynamic integration technique based on histogram reweighting, using the previous coexistence point as a reference point. The vapor-liquid phase diagram is obtained from constant-pressure simulations of the unconstrained system using tempering and histogram reweighting. The phase diagram of the system is found to contain a stable critical point and a triple point. The phase diagram of the corresponding constrained cell model is also found to contain both a stable critical point and a triple point.
Bennington, Ninfa; Thurber, Clifford; Feigl, Kurt; ,
2011-01-01
Several studies of the 2004 Parkfield earthquake have linked the spatial distribution of the event’s aftershocks to the mainshock slip distribution on the fault. Using geodetic data, we find a model of coseismic slip for the 2004 Parkfield earthquake with the constraint that the edges of coseismic slip patches align with aftershocks. The constraint is applied by encouraging the curvature of coseismic slip in each model cell to be equal to the negative of the curvature of seismicity density. The large patch of peak slip about 15 km northwest of the 2004 hypocenter found in the curvature-constrained model is in good agreement in location and amplitude with previous geodetic studies and the majority of strong motion studies. The curvature-constrained solution shows slip primarily between aftershock “streaks” with the continuation of moderate levels of slip to the southeast. These observations are in good agreement with strong motion studies, but inconsistent with the majority of published geodetic slip models. Southeast of the 2004 hypocenter, a patch of peak slip observed in strong motion studies is absent from our curvature-constrained model, but the available GPS data do not resolve slip in this region. We conclude that the geodetic slip model constrained by the aftershock distribution fits the geodetic data quite well and that inconsistencies between models derived from seismic and geodetic data can be attributed largely to resolution issues.
Hee, S.; Vázquez, J. A.; Handley, W. J.; ...
2016-12-01
Data-driven model-independent reconstructions of the dark energy equation of state w(z) are presented using Planck 2015 era CMB, BAO, SNIa and Lyman-α data. These reconstructions identify the w(z) behaviour supported by the data and show a bifurcation of the equation of state posterior in the range 1.5 < z < 3. Although the concordance ΛCDM model is consistent with the data at all redshifts in one of the bifurcated spaces, in the other a supernegative equation of state (also known as ‘phantom dark energy’) is identified within the 1.5σ confidence intervals of the posterior distribution. In order to identify themore » power of different datasets in constraining the dark energy equation of state, we use a novel formulation of the Kullback–Leibler divergence. Moreover, this formalism quantifies the information the data add when moving from priors to posteriors for each possible dataset combination. The SNIa and BAO datasets are shown to provide much more constraining power in comparison to the Lyman-α datasets. Furthermore, SNIa and BAO constrain most strongly around redshift range 0.1 - 0.5, whilst the Lyman-α data constrains weakly over a broader range. We do not attribute the supernegative favouring to any particular dataset, and note that the ΛCDM model was favoured at more than 2 log-units in Bayes factors over all the models tested despite the weakly preferred w(z) structure in the data.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hee, S.; Vázquez, J. A.; Handley, W. J.
Data-driven model-independent reconstructions of the dark energy equation of state w(z) are presented using Planck 2015 era CMB, BAO, SNIa and Lyman-α data. These reconstructions identify the w(z) behaviour supported by the data and show a bifurcation of the equation of state posterior in the range 1.5 < z < 3. Although the concordance ΛCDM model is consistent with the data at all redshifts in one of the bifurcated spaces, in the other a supernegative equation of state (also known as ‘phantom dark energy’) is identified within the 1.5σ confidence intervals of the posterior distribution. In order to identify themore » power of different datasets in constraining the dark energy equation of state, we use a novel formulation of the Kullback–Leibler divergence. Moreover, this formalism quantifies the information the data add when moving from priors to posteriors for each possible dataset combination. The SNIa and BAO datasets are shown to provide much more constraining power in comparison to the Lyman-α datasets. Furthermore, SNIa and BAO constrain most strongly around redshift range 0.1 - 0.5, whilst the Lyman-α data constrains weakly over a broader range. We do not attribute the supernegative favouring to any particular dataset, and note that the ΛCDM model was favoured at more than 2 log-units in Bayes factors over all the models tested despite the weakly preferred w(z) structure in the data.« less
Constrained minimization of smooth functions using a genetic algorithm
NASA Technical Reports Server (NTRS)
Moerder, Daniel D.; Pamadi, Bandu N.
1994-01-01
The use of genetic algorithms for minimization of differentiable functions that are subject to differentiable constraints is considered. A technique is demonstrated for converting the solution of the necessary conditions for a constrained minimum into an unconstrained function minimization. This technique is extended as a global constrained optimization algorithm. The theory is applied to calculating minimum-fuel ascent control settings for an energy state model of an aerospace plane.
Time-of-flight mass spectrometry of laser exploding foil initiated PETN samples
NASA Astrophysics Data System (ADS)
Fajardo, Mario E.; Molek, Christopher D.; Fossum, Emily C.
2017-01-01
We report the results of time-of-flight mass spectrometry (TOFMS) measurements of the gaseous products of thin-film pentaerythritol tetranitrate [PETN, C(CH2NO3)4] samples reacting in vacuo. The PETN sample spots are produced by masked physical vapor deposition [A.S. Tappan, et al., AIP Conf. Proc. 1426, 677 (2012)] onto a first-surface aluminum mirror. A pulsed laser beam imaged through the soda lime glass mirror substrate converts the aluminum layer into a high-temperature high-pressure plasma which initiates chemical reactions in the overlying PETN sample. We had previously proposed [E.C. Fossum, et al., AIP Conf. Proc. 1426, 235 (2012)] to exploit differences in gaseous product chemical identities and molecular velocities to provide a chemically-based diagnostic for distinguishing between "detonation-like" and deflagration responses. Briefly: we expect in-vacuum detonations to produce hyperthermal (v˜10 km/s) thermodynamically-stable products such as N2, CO2, and H2O, and for deflagrations to produce mostly reaction intermediates, such as NO and NO2, with much slower molecular velocities - consistent with the expansion-quenched thermal decomposition of PETN. We observe primarily slow reaction intermediates (NO2, CH2NO3) at low laser pulse energies, the appearance of NO at intermediate laser pulse energies, and the appearance of hyperthemal CO/N2 at mass 28 amu at the highest laser pulse energies. However, these results are somewhat ambiguous, as the NO, NO2, and CH2NO3 intermediates persist and all species become hyperthermal at the higher laser pulse energies. Also, the purported CO/N2 signal at 28 amu may be contaminated by silicon ablated from the glass mirror substrate. We plan to mitigate these problems in future experiments by adopting the "Buelow" sample configuration which employs an intermediate foil barrier to shield the energetic material from the laser and the laser driven plasma [S.J. Buelow, et al., AIP Conf. Proc. 706, 1377 (2003)].
Discontinuous Galerkin method for multicomponent chemically reacting flows and combustion
NASA Astrophysics Data System (ADS)
Lv, Yu; Ihme, Matthias
2014-08-01
This paper presents the development of a discontinuous Galerkin (DG) method for application to chemically reacting flows in subsonic and supersonic regimes under the consideration of variable thermo-viscous-diffusive transport properties, detailed and stiff reaction chemistry, and shock capturing. A hybrid-flux formulation is developed for treatment of the convective fluxes, combining a conservative Riemann-solver and an extended double-flux scheme. A computationally efficient splitting scheme is proposed, in which advection and diffusion operators are solved in the weak form, and the chemically stiff substep is advanced in the strong form using a time-implicit scheme. The discretization of the viscous-diffusive transport terms follows the second form of Bassi and Rebay, and the WENO-based limiter due to Zhong and Shu is extended to multicomponent systems. Boundary conditions are developed for subsonic and supersonic flow conditions, and the algorithm is coupled to thermochemical libraries to account for detailed reaction chemistry and complex transport. The resulting DG method is applied to a series of test cases of increasing physico-chemical complexity. Beginning with one- and two-dimensional multispecies advection and shock-fluid interaction problems, computational efficiency, convergence, and conservation properties are demonstrated. This study is followed by considering a series of detonation and supersonic combustion problems to investigate the convergence-rate and the shock-capturing capability in the presence of one- and multistep reaction chemistry. The DG algorithm is then applied to diffusion-controlled deflagration problems. By examining convergence properties for polynomial order and spatial resolution, and comparing these with second-order finite-volume solutions, it is shown that optimal convergence is achieved and that polynomial refinement provides advantages in better resolving the localized flame structure and complex flow-field features associated with multidimensional and hydrodynamic/thermo-diffusive instabilities in deflagration and detonation systems. Comparisons with standard third- and fifth-order WENO schemes are presented to illustrate the benefit of the DG scheme for application to detonation and multispecies flow/shock-interaction problems.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sullivan, Kyle T.
Reactive composites utilizing nanoparticles have been the topic of extensive research in the past two decades. The driver for this is that, as the particle size is decreased, the mixing scale between constituents is greatly reduced, which has long thought to increase the rate of chemical reaction. While a general trend of increased reactivity has been seen for metal / metal oxide, or thermite, reactive materials, some results have demonstrated diminishing returns as the particle size is further decreased. Recent results have shown that nanoparticles, which are typically aggregates of several primary particles, can undergo very rapid coalescence to formmore » micron particles once a critical temperature is reached. Experiments on this topic to date have been performed on very small sample masses, and sometimes under vacuum; conditions which are not representative of the environment during a deflagration. In this feasibility study, a custom burn tube was used to ignite and react 100 mg powdered thermite samples in long acrylic tubes. X-ray imaging at APS Sector 32 was performed to image the particle field as a function of distance and time as the rarefied particle cloud expanded and flowed down the tube. Five different thermite formulations were investigated, Al / CuO, Al / Fe 2O 3, Al / SnO 2, Al / WO 3, and Al / Fe 2O 3, along with Al / CuO formulations with different sizes of Al particles ranging from 80 nm to approximate 10 μm. The results clearly show that the sample powder reacts and unloads into a distribution of larger micron-scale particles (~5-500 μm), which continue to react and propagate as the particle-laden stream flows down the tube. This was the first direct imaging of the particle field during a thermite deflagration, and gives significant insight into the evolution of reactants to products. Analysis of phase is currently being pursued to determine whether this method can be used to extract reaction kinetics.« less
NASA Astrophysics Data System (ADS)
Fetherolf, B. L.; Litzinger, T. A.; Lu, Y.-C.; Kuo, Kenneth K.
1993-11-01
The RDX-based composite propellants XM39 and M43 are similar in composition but exhibit significant differences in burning behavior. Experimental studies of the physical and chemical processes governing the CO2 laser-induced pyrolysis and deflagration of these two materials were conducted to characterize these differences in behavior and to gain some insight into the mechanisms responsible for the observed differences. Tests were conducted at one, three, and five atmospheres and laser heat fluxes of 100 - 1000 W/sq cm. Quantitative gaseous species profiles were measured with a microprobe/mass spectrometer system and both gas-phase temperature profiles and surface temperatures were measured with fine-wire thermocouples. Both materials exhibited similar gas-phase reaction chemistry to that of RDX with a primary nonluminous flame zone due to the reaction of CH2O and NO2 and a final luminous flame zone where HCN, NO, and a smaller amount of N2O were consumed to form the final products. However, the gas-phase zonal structure was significantly stretched out in comparison to the structure for pure RDX. The luminous flame was only observed above three atmospheres for M43 and above five atmospheres for XM39. Species and temperature measurements at the surfaces of the pyrolyzing propellants appeared to indicate more reaction in the condensed phase (i.e., melt layer) for M43 than for XM39. Subsurface gas species were measured by placing a probe within a hole drilled partway through a sample of XM39. The results indicated substantially less H2O, CH2O, HCN, and NO2 than were measured directly above the surface. This result and the observation of a temperature rise of about 100 degrees within the first 150 microns above the surface for both XM39 and M43 support the possible existence of a thin gas-phase reaction zone directly above the propellant surface.
NASA Technical Reports Server (NTRS)
Pawson, Steven; Ott, Lesley E.; Zhu, Zhengxin; Bowman, Kevin; Brix, Holger; Collatz, G. James; Dutkiewicz, Stephanie; Fisher, Joshua B.; Gregg, Watson W.; Hill, Chris;
2011-01-01
Forward GEOS-5 AGCM simulations of CO2, with transport constrained by analyzed meteorology for 2009-2010, are examined. The CO2 distributions are evaluated using AIRS upper tropospheric CO2 and ACOS-GOSAT total column CO2 observations. Different combinations of surface C02 fluxes are used to generate ensembles of runs that span some uncertainty in surface emissions and uptake. The fluxes are specified in GEOS-5 from different inventories (fossil and biofuel), different data-constrained estimates of land biological emissions, and different data-constrained ocean-biology estimates. One set of fluxes is based on the established "Transcom" database and others are constructed using contemporary satellite observations to constrain land and ocean process models. Likewise, different approximations to sub-grid transport are employed, to construct an ensemble of CO2 distributions related to transport variability. This work is part of NASA's "Carbon Monitoring System Flux Pilot Project,"
Polarization and studies of evolved star mass loss
NASA Astrophysics Data System (ADS)
Sargent, Benjamin; Srinivasan, Sundar; Riebel, David; Meixner, Margaret
2012-05-01
Polarization studies of astronomical dust have proven very useful in constraining its properties. Such studies are used to constrain the spatial arrangement, shape, composition, and optical properties of astronomical dust grains. Here we explore possible connections between astronomical polarization observations to our studies of mass loss from evolved stars. We are studying evolved star mass loss in the Large Magellanic Cloud (LMC) by using photometry from the Surveying the Agents of a Galaxy's Evolution (SAGE; PI: M. Meixner) Spitzer Space Telescope Legacy program. We use the radiative transfer program 2Dust to create our Grid of Red supergiant and Asymptotic giant branch ModelS (GRAMS), in order to model this mass loss. To model emission of polarized light from evolved stars, however, we appeal to other radiative transfer codes. We probe how polarization observations might be used to constrain the dust shell and dust grain properties of the samples of evolved stars we are studying.
A Bayesian ensemble data assimilation to constrain model parameters and land-use carbon emissions
NASA Astrophysics Data System (ADS)
Lienert, Sebastian; Joos, Fortunat
2018-05-01
A dynamic global vegetation model (DGVM) is applied in a probabilistic framework and benchmarking system to constrain uncertain model parameters by observations and to quantify carbon emissions from land-use and land-cover change (LULCC). Processes featured in DGVMs include parameters which are prone to substantial uncertainty. To cope with these uncertainties Latin hypercube sampling (LHS) is used to create a 1000-member perturbed parameter ensemble, which is then evaluated with a diverse set of global and spatiotemporally resolved observational constraints. We discuss the performance of the constrained ensemble and use it to formulate a new best-guess version of the model (LPX-Bern v1.4). The observationally constrained ensemble is used to investigate historical emissions due to LULCC (ELUC) and their sensitivity to model parametrization. We find a global ELUC estimate of 158 (108, 211) PgC (median and 90 % confidence interval) between 1800 and 2016. We compare ELUC to other estimates both globally and regionally. Spatial patterns are investigated and estimates of ELUC of the 10 countries with the largest contribution to the flux over the historical period are reported. We consider model versions with and without additional land-use processes (shifting cultivation and wood harvest) and find that the difference in global ELUC is on the same order of magnitude as parameter-induced uncertainty and in some cases could potentially even be offset with appropriate parameter choice.
Interstage Flammability Analysis Approach
NASA Technical Reports Server (NTRS)
Little, Jeffrey K.; Eppard, William M.
2011-01-01
The Interstage of the Ares I launch platform houses several key components which are on standby during First Stage operation: the Reaction Control System (ReCS), the Upper Stage (US) Thrust Vector Control (TVC) and the J-2X with the Main Propulsion System (MPS) propellant feed system. Therefore potentially dangerous leaks of propellants could develop. The Interstage leaks analysis addresses the concerns of localized mixing of hydrogen and oxygen gases to produce deflagration zones in the Interstage of the Ares I launch vehicle during First Stage operation. This report details the approach taken to accomplish the analysis. Specified leakage profiles and actual flammability results are not presented due to proprietary and security restrictions. The interior volume formed by the Interstage walls, bounding interfaces with the Upper and First Stages, and surrounding the J2-X engine was modeled using Loci-CHEM to assess the potential for flammable gas mixtures to develop during First Stage operations. The transient analysis included a derived flammability indicator based on mixture ratios to maintain achievable simulation times. Validation of results was based on a comparison to Interstage pressure profiles outlined in prior NASA studies. The approach proved useful in the bounding of flammability risk in supporting program hazard reviews.
A full scale hydrodynamic simulation of pyrotechnic combustion
NASA Astrophysics Data System (ADS)
Kim, Bohoon; Jang, Seung-Gyo; Yoh, Jack
2017-06-01
A full scale hydrodynamic simulation that requires an accurate reproduction of shock-induced detonation was conducted for design of an energetic component system. A series of small scale gap tests and detailed hydrodynamic simulations were used to validate the reactive flow model for predicting the shock propagation in a train configuration and to quantify the shock sensitivity of the energetic materials. The energetic component system is composed of four main components, namely a donor unit (HNS + HMX), a bulkhead (STS), an acceptor explosive (RDX), and a propellant (BKNO3) for gas generation. The pressurized gases generated from the burning propellant were purged into a 10 cc release chamber for study of the inherent oscillatory flow induced by the interferences between shock and rarefaction waves. The pressure fluctuations measured from experiment and calculation were investigated to further validate the peculiar peak at specific characteristic frequency (ωc = 8.3 kHz). In this paper, a step-by-step numerical description of detonation of high explosive components, deflagration of propellant component, and deformation of metal component is given in order to facilitate the proper implementation of the outlined formulation into a shock physics code for a full scale hydrodynamic simulation of the energetic component system.
Implementation of remote sensing data for flood forecasting
NASA Astrophysics Data System (ADS)
Grimaldi, S.; Li, Y.; Pauwels, V. R. N.; Walker, J. P.; Wright, A. J.
2016-12-01
Flooding is one of the most frequent and destructive natural disasters. A timely, accurate and reliable flood forecast can provide vital information for flood preparedness, warning delivery, and emergency response. An operational flood forecasting system typically consists of a hydrologic model, which simulates runoff generation and concentration, and a hydraulic model, which models riverine flood wave routing and floodplain inundation. However, these two types of models suffer from various sources of uncertainties, e.g., forcing data initial conditions, model structure and parameters. To reduce those uncertainties, current forecasting systems are typically calibrated and/or updated using streamflow measurements, and such applications are limited in well-gauged areas. The recent increasing availability of spatially distributed Remote Sensing (RS) data offers new opportunities for flood events investigation and forecast. Based on an Australian case study, this presentation will discuss the use 1) of RS soil moisture data to constrain a hydrologic model, and 2) of RS-derived flood extent and level to constrain a hydraulic model. The hydrological model is based on a semi-distributed system coupled with a two-soil-layer rainfall-runoff model GRKAL and a linear Muskingum routing model. Model calibration was performed using either 1) streamflow data only or 2) both streamflow and RS soil moisture data. The model was then further constrained through the integration of real-time soil moisture data. The hydraulic model is based on LISFLOOD-FP which solves the 2D inertial approximation of the Shallow Water Equations. Streamflow data and RS-derived flood extent and levels were used to apply a multi-objective calibration protocol. The effectiveness with which each data source or combination of data sources constrained the parameter space was quantified and discussed.
Compromise Approach-Based Genetic Algorithm for Constrained Multiobjective Portfolio Selection Model
NASA Astrophysics Data System (ADS)
Li, Jun
In this paper, fuzzy set theory is incorporated into a multiobjective portfolio selection model for investors’ taking into three criteria: return, risk and liquidity. The cardinality constraint, the buy-in threshold constraint and the round-lots constraints are considered in the proposed model. To overcome the difficulty of evaluation a large set of efficient solutions and selection of the best one on non-dominated surface, a compromise approach-based genetic algorithm is presented to obtain a compromised solution for the proposed constrained multiobjective portfolio selection model.
Medvigy, David; Moorcroft, Paul R
2012-01-19
Terrestrial biosphere models are important tools for diagnosing both the current state of the terrestrial carbon cycle and forecasting terrestrial ecosystem responses to global change. While there are a number of ongoing assessments of the short-term predictive capabilities of terrestrial biosphere models using flux-tower measurements, to date there have been relatively few assessments of their ability to predict longer term, decadal-scale biomass dynamics. Here, we present the results of a regional-scale evaluation of the Ecosystem Demography version 2 (ED2)-structured terrestrial biosphere model, evaluating the model's predictions against forest inventory measurements for the northeast USA and Quebec from 1985 to 1995. Simulations were conducted using a default parametrization, which used parameter values from the literature, and a constrained model parametrization, which had been developed by constraining the model's predictions against 2 years of measurements from a single site, Harvard Forest (42.5° N, 72.1° W). The analysis shows that the constrained model parametrization offered marked improvements over the default model formulation, capturing large-scale variation in patterns of biomass dynamics despite marked differences in climate forcing, land-use history and species-composition across the region. These results imply that data-constrained parametrizations of structured biosphere models such as ED2 can be successfully used for regional-scale ecosystem prediction and forecasting. We also assess the model's ability to capture sub-grid scale heterogeneity in the dynamics of biomass growth and mortality of different sizes and types of trees, and then discuss the implications of these analyses for further reducing the remaining biases in the model's predictions.
Tropospheric transport differences between models using the same large-scale meteorological fields
NASA Astrophysics Data System (ADS)
Orbe, Clara; Waugh, Darryn W.; Yang, Huang; Lamarque, Jean-Francois; Tilmes, Simone; Kinnison, Douglas E.
2017-01-01
The transport of chemicals is a major uncertainty in the modeling of tropospheric composition. A common approach is to transport gases using the winds from meteorological analyses, either using them directly in a chemical transport model or by constraining the flow in a general circulation model. Here we compare the transport of idealized tracers in several different models that use the same meteorological fields taken from Modern-Era Retrospective analysis for Research and Applications (MERRA). We show that, even though the models use the same meteorological fields, there are substantial differences in their global-scale tropospheric transport related to large differences in parameterized convection between the simulations. Furthermore, we find that the transport differences between simulations constrained with the same-large scale flow are larger than differences between free-running simulations, which have differing large-scale flow but much more similar convective mass fluxes. Our results indicate that more attention needs to be paid to convective parameterizations in order to understand large-scale tropospheric transport in models, particularly in simulations constrained with analyzed winds.
NASA Astrophysics Data System (ADS)
Jungman, Gerard
1992-11-01
Yukawa-coupling-constant unification together with the known fermion masses is used to constrain SO(10) models. We consider the case of one (heavy) generation, with the tree-level relation mb=mτ, calculating the limits on the intermediate scales due to the known limits on fermion masses. This analysis extends previous analyses which addressed only the simplest symmetry-breaking schemes. In the case where the low-energy model is the standard model with one Higgs doublet, there are very strong constraints due to the known limits on the top-quark mass and the τ-neutrino mass. The two-Higgs-doublet case is less constrained, though we can make progress in constraining this model also. We identify those parameters to which the viability of the model is most sensitive. We also discuss the ``triviality'' bounds on mt obtained from the analysis of the Yukawa renormalization-group equations. Finally we address the role of a speculative constraint on the τ-neutrino mass, arising from the cosmological implications of anomalous B+L violation in the early Universe.
Terrestrial Sagnac delay constraining modified gravity models
NASA Astrophysics Data System (ADS)
Karimov, R. Kh.; Izmailov, R. N.; Potapov, A. A.; Nandi, K. K.
2018-04-01
Modified gravity theories include f(R)-gravity models that are usually constrained by the cosmological evolutionary scenario. However, it has been recently shown that they can also be constrained by the signatures of accretion disk around constant Ricci curvature Kerr-f(R0) stellar sized black holes. Our aim here is to use another experimental fact, viz., the terrestrial Sagnac delay to constrain the parameters of specific f(R)-gravity prescriptions. We shall assume that a Kerr-f(R0) solution asymptotically describes Earth's weak gravity near its surface. In this spacetime, we shall study oppositely directed light beams from source/observer moving on non-geodesic and geodesic circular trajectories and calculate the time gap, when the beams re-unite. We obtain the exact time gap called Sagnac delay in both cases and expand it to show how the flat space value is corrected by the Ricci curvature, the mass and the spin of the gravitating source. Under the assumption that the magnitude of corrections are of the order of residual uncertainties in the delay measurement, we derive the allowed intervals for Ricci curvature. We conclude that the terrestrial Sagnac delay can be used to constrain the parameters of specific f(R) prescriptions. Despite using the weak field gravity near Earth's surface, it turns out that the model parameter ranges still remain the same as those obtained from the strong field accretion disk phenomenon.
Modeling and simulating networks of interdependent protein interactions.
Stöcker, Bianca K; Köster, Johannes; Zamir, Eli; Rahmann, Sven
2018-05-21
Protein interactions are fundamental building blocks of biochemical reaction systems underlying cellular functions. The complexity and functionality of these systems emerge not only from the protein interactions themselves but also from the dependencies between these interactions, as generated by allosteric effects or mutual exclusion due to steric hindrance. Therefore, formal models for integrating and utilizing information about interaction dependencies are of high interest. Here, we describe an approach for endowing protein networks with interaction dependencies using propositional logic, thereby obtaining constrained protein interaction networks ("constrained networks"). The construction of these networks is based on public interaction databases as well as text-mined information about interaction dependencies. We present an efficient data structure and algorithm to simulate protein complex formation in constrained networks. The efficiency of the model allows fast simulation and facilitates the analysis of many proteins in large networks. In addition, this approach enables the simulation of perturbation effects, such as knockout of single or multiple proteins and changes of protein concentrations. We illustrate how our model can be used to analyze a constrained human adhesome protein network, which is responsible for the formation of diverse and dynamic cell-matrix adhesion sites. By comparing protein complex formation under known interaction dependencies versus without dependencies, we investigate how these dependencies shape the resulting repertoire of protein complexes. Furthermore, our model enables investigating how the interplay of network topology with interaction dependencies influences the propagation of perturbation effects across a large biochemical system. Our simulation software CPINSim (for Constrained Protein Interaction Network Simulator) is available under the MIT license at http://github.com/BiancaStoecker/cpinsim and as a Bioconda package (https://bioconda.github.io).
Sensitivity Analysis Tailored to Constrain 21st Century Terrestrial Carbon-Uptake
NASA Astrophysics Data System (ADS)
Muller, S. J.; Gerber, S.
2013-12-01
The long-term fate of terrestrial carbon (C) in response to climate change remains a dominant source of uncertainty in Earth-system model projections. Increasing atmospheric CO2 could be mitigated by long-term net uptake of C, through processes such as increased plant productivity due to "CO2-fertilization". Conversely, atmospheric conditions could be exacerbated by long-term net release of C, through processes such as increased decomposition due to higher temperatures. This balance is an important area of study, and a major source of uncertainty in long-term (>year 2050) projections of planetary response to climate change. We present results from an innovative application of sensitivity analysis to LM3V, a dynamic global vegetation model (DGVM), intended to identify observed/observable variables that are useful for constraining long-term projections of C-uptake. We analyzed the sensitivity of cumulative C-uptake by 2100, as modeled by LM3V in response to IPCC AR4 scenario climate data (1860-2100), to perturbations in over 50 model parameters. We concurrently analyzed the sensitivity of over 100 observable model variables, during the extant record period (1970-2010), to the same parameter changes. By correlating the sensitivities of observable variables with the sensitivity of long-term C-uptake we identified model calibration variables that would also constrain long-term C-uptake projections. LM3V employs a coupled carbon-nitrogen cycle to account for N-limitation, and we find that N-related variables have an important role to play in constraining long-term C-uptake. This work has implications for prioritizing field campaigns to collect global data that can help reduce uncertainties in the long-term land-atmosphere C-balance. Though results of this study are specific to LM3V, the processes that characterize this model are not completely divorced from other DGVMs (or reality), and our approach provides valuable insights into how data can be leveraged to be better constrain projections for the land carbon sink.
Current distribution measurements inside an electromagnetic plasma gun operated in a gas-puff mode.
Poehlmann, Flavio R; Cappelli, Mark A; Rieker, Gregory B
2010-12-01
Measurements are presented of the time-dependent current distribution inside a coaxial electromagnetic plasma gun. The measurements are carried out using an array of six axially distributed dual-Rogowski coils in a balanced circuit configuration. The radial current distributions indicate that operation in the gas-puff mode, i.e., the mode in which the electrode voltage is applied before injection of the gas, results in a stationary ionization front consistent with the presence of a plasma deflagration. The effects of varying the bank capacitance, transmission line inductance, and applied electrode voltage were studied over the range from 14 to 112 μF, 50 to 200 nH, and 1 to 3 kV, respectively.
Current distribution measurements inside an electromagnetic plasma gun operated in a gas-puff mode
Poehlmann, Flavio R.; Cappelli, Mark A.; Rieker, Gregory B.
2010-01-01
Measurements are presented of the time-dependent current distribution inside a coaxial electromagnetic plasma gun. The measurements are carried out using an array of six axially distributed dual-Rogowski coils in a balanced circuit configuration. The radial current distributions indicate that operation in the gas-puff mode, i.e., the mode in which the electrode voltage is applied before injection of the gas, results in a stationary ionization front consistent with the presence of a plasma deflagration. The effects of varying the bank capacitance, transmission line inductance, and applied electrode voltage were studied over the range from 14 to 112 μF, 50 to 200 nH, and 1 to 3 kV, respectively. PMID:21267082
Binary progenitors of supernovae
NASA Astrophysics Data System (ADS)
Trimble, V.
1984-12-01
Among the massive stars that are expected to produce Type II, hydrogen-rich supernovae, the presence of a close companion can increase the main sequence mass needed to yield a collapsing core. In addition, due to mass transfer from the primary to the secondary, the companion enhances the stripping of the stellar hydrogen envelope produced by single star winds and thereby makes it harder for the star to give rise to a typical SN II light curve. Among the less massive stars that may be the basis for Type I, hydrogen-free supernovae, a close companion could be an innocent bystander to carbon detonation/deflagration in the primary. It may alternatively be a vital participant which transfers material to a white dwarf primary and drives it to explosive conditions.
Programs for calculating the explosion resistance of buildings and structures
NASA Astrophysics Data System (ADS)
Shevchenko, Nikita; Manucharyan, Rachik; Gravit, Marina; Geraskin, Yuriy
2017-10-01
Trends in the development of the oil and gas processing industry and facilities using liquefied hydrocarbon gases in the technological processes lead to an increase in the number of emergency situations and as a result to the increase of the damage and environmental degradation. A gasified housing stock is a special group of explosive objects. The accident in such disasters lead to hundreds of human victims. The main danger of explosive effects on objects can lead to a large-scale destruction connected with an “internal explosion”. It results in the release of combustible substances (gas, oil, etc.) into enclosed or semi-enclosed rooms (modules) with the ignition (in explosion mode of deflagration or detonation) of air gas mixtures (AGM).
Report on JANNAF panel on shotgun/relative quickness testing
NASA Technical Reports Server (NTRS)
Gould, R. A.
1980-01-01
As the need for more energetic solid propellants continues, a number of problems arises. One of these is the tendency of high energy propellants to transition from burning (deflagration) to detonation in regions where the propellant is present in small particle sizes; e.g., in case bonding areas of a motor after a rapid depressurization causes a shear zone at the bond interface as the stressed propellant and motor case relax at different rates. In an effort to determine the susceptibility of propellants to high strain rate break up (friability), and subsequent DDT, the propulsion community uses the shotgun/relative quickness test as one of a number of screening tests for new propellant formulations. Efforts to standardize test techniques and equipment are described.
Study of thermite mixture consolidated by the cold gas dynamic spray process
NASA Astrophysics Data System (ADS)
Bacciochini, A.; Maines, G.; Poupart, C.; Akbarnejad, H.; Radulescu, M.; Jodoin, B.; Zhang, F.; Lee, J. J.
2014-05-01
The present study focused on the cold gas dynamic spray process for manufacturing porosity free, finely structured energetic materials with high reactivity and structural integrity. The experiments have focused the reaction between the aluminium and metal oxide, such as Al-CuO system. The consolidation of the materials used the cold gas dynamic spray technique, where the particles are accelerated to high speeds and consolidated via plastic deformation upon impact. Reactive composites are formed in arbitrary shapes with close to zero porosity and without any reactions during the consolidation phase. Reactivity of mixtures has been investigated through flame propagation analysis on cold sprayed samples and compacted powder mixture. Deflagration tests showed the influence of porosity on the reactivity.
Ignition Study on a Rotary-valved Air-breathing Pulse Detonation Engine
NASA Astrophysics Data System (ADS)
Wu, Yuwen; Han, Qixiang; Shen, Yujia; Zhao, Wei
2017-05-01
In the present study, the ignition effect on detonation initiation was investigated in the air-breathing pulse detonation engine. Two kinds of fuel injection and ignition methods were applied. For one method, fuel and air was pre-mixed outside the PDE and then injected into the detonation tube. The droplet sizes of mixtures were measured. An annular cavity was used as the ignition section. For the other method, fuel-air mixtures were mixed inside the PDE, and a pre-combustor was utilized as the ignition source. At firing frequency of 20 Hz, transition to detonation was obtained. Experimental results indicated that the ignition position and initial flame acceleration had important effects on the deflagration-to-detonation transition.
Hacker, David E; Hoinka, Jan; Iqbal, Emil S; Przytycka, Teresa M; Hartman, Matthew C T
2017-03-17
Highly constrained peptides such as the knotted peptide natural products are promising medicinal agents because of their impressive biostability and potent activity. Yet, libraries of highly constrained peptides are challenging to prepare. Here, we present a method which utilizes two robust, orthogonal chemical steps to create highly constrained bicyclic peptide libraries. This technology was optimized to be compatible with in vitro selections by mRNA display. We performed side-by-side monocyclic and bicyclic selections against a model protein (streptavidin). Both selections resulted in peptides with mid-nanomolar affinity, and the bicyclic selection yielded a peptide with remarkable protease resistance.
Fuzzy multi-objective chance-constrained programming model for hazardous materials transportation
NASA Astrophysics Data System (ADS)
Du, Jiaoman; Yu, Lean; Li, Xiang
2016-04-01
Hazardous materials transportation is an important and hot issue of public safety. Based on the shortest path model, this paper presents a fuzzy multi-objective programming model that minimizes the transportation risk to life, travel time and fuel consumption. First, we present the risk model, travel time model and fuel consumption model. Furthermore, we formulate a chance-constrained programming model within the framework of credibility theory, in which the lengths of arcs in the transportation network are assumed to be fuzzy variables. A hybrid intelligent algorithm integrating fuzzy simulation and genetic algorithm is designed for finding a satisfactory solution. Finally, some numerical examples are given to demonstrate the efficiency of the proposed model and algorithm.
Constraining the Mechanism of D" Anisotropy: Diversity of Observation Types Required
NASA Astrophysics Data System (ADS)
Creasy, N.; Pisconti, A.; Long, M. D.; Thomas, C.
2017-12-01
A variety of different mechanisms have been proposed as explanations for seismic anisotropy at the base of the mantle, including crystallographic preferred orientation of various minerals (bridgmanite, post-perovskite, and ferropericlase) and shape preferred orientation of elastically distinct materials such as partial melt. Investigations of the mechanism for D" anisotropy are usually ambiguous, as seismic observations rarely (if ever) uniquely constrain a mechanism. Observations of shear wave splitting and polarities of SdS and PdP reflections off the D" discontinuity are among our best tools for probing D" anisotropy; however, typical data sets cannot constrain a unique scenario suggested by the mineral physics literature. In this work, we determine what types of body wave observations are required to uniquely constrain a mechanism for D" anisotropy. We test multiple possible models based on both single-crystal and poly-phase elastic tensors provided by mineral physics studies. We predict shear wave splitting parameters for SKS, SKKS, and ScS phases and reflection polarities off the D" interface for a range of possible propagation directions. We run a series of tests that create synthetic data sets by random selection over multiple iterations, controlling the total number of measurements, the azimuthal distribution, and the type of phases. We treat each randomly drawn synthetic dataset with the same methodology as in Ford et al. (2015) to determine the possible mechanism(s), carrying out a grid search over all possible elastic tensors and orientations to determine which are consistent with the synthetic data. We find is it difficult to uniquely constrain the starting model with a realistic number of seismic anisotropy measurements with only one measurement technique or phase type. However, having a mix of SKS, SKKS, and ScS measurements, or a mix of shear wave splitting and reflection polarity measurements, dramatically increases the probability of uniquely constraining the starting model. We also explore what types of datasets are needed to uniquely constrain the orientation(s) of anisotropic symmetry if the mechanism is assumed.
NASA Astrophysics Data System (ADS)
Hee, S.; Vázquez, J. A.; Handley, W. J.; Hobson, M. P.; Lasenby, A. N.
2017-04-01
Data-driven model-independent reconstructions of the dark energy equation of state w(z) are presented using Planck 2015 era cosmic microwave background, baryonic acoustic oscillations (BAO), Type Ia supernova (SNIa) and Lyman α (Lyα) data. These reconstructions identify the w(z) behaviour supported by the data and show a bifurcation of the equation of state posterior in the range 1.5 < z < 3. Although the concordance Λ cold dark matter (ΛCDM) model is consistent with the data at all redshifts in one of the bifurcated spaces, in the other, a supernegative equation of state (also known as 'phantom dark energy') is identified within the 1.5σ confidence intervals of the posterior distribution. To identify the power of different data sets in constraining the dark energy equation of state, we use a novel formulation of the Kullback-Leibler divergence. This formalism quantifies the information the data add when moving from priors to posteriors for each possible data set combination. The SNIa and BAO data sets are shown to provide much more constraining power in comparison to the Lyα data sets. Further, SNIa and BAO constrain most strongly around redshift range 0.1-0.5, whilst the Lyα data constrain weakly over a broader range. We do not attribute the supernegative favouring to any particular data set, and note that the ΛCDM model was favoured at more than 2 log-units in Bayes factors over all the models tested despite the weakly preferred w(z) structure in the data.
Constraining ecosystem processes from tower fluxes and atmospheric profiles.
Hill, T C; Williams, M; Woodward, F I; Moncrieff, J B
2011-07-01
The planetary boundary layer (PBL) provides an important link between the scales and processes resolved by global atmospheric sampling/modeling and site-based flux measurements. The PBL is in direct contact with the land surface, both driving and responding to ecosystem processes. Measurements within the PBL (e.g., by radiosondes, aircraft profiles, and flask measurements) have a footprint, and thus an integrating scale, on the order of 1-100 km. We use the coupled atmosphere-biosphere model (CAB) and a Bayesian data assimilation framework to investigate the amount of biosphere process information that can be inferred from PBL measurements. We investigate the information content of PBL measurements in a two-stage study. First, we demonstrate consistency between the coupled model (CAB) and measurements, by comparing the model to eddy covariance flux tower measurements (i.e., water and carbon fluxes) and also PBL scalar profile measurements (i.e., water, carbon dioxide, and temperature) from Canadian boreal forest. Second, we use the CAB model in a set of Bayesian inversions experiments using synthetic data for a single day. In the synthetic experiment, leaf area and respiration were relatively well constrained, whereas surface albedo and plant hydraulic conductance were only moderately constrained. Finally, the abilities of the PBL profiles and the eddy covariance data to constrain the parameters were largely similar and only slightly lower than the combination of both observations.
Weighting climate model projections using observational constraints.
Gillett, Nathan P
2015-11-13
Projected climate change integrates the net response to multiple climate feedbacks. Whereas existing long-term climate change projections are typically based on unweighted individual climate model simulations, as observed climate change intensifies it is increasingly becoming possible to constrain the net response to feedbacks and hence projected warming directly from observed climate change. One approach scales simulated future warming based on a fit to observations over the historical period, but this approach is only accurate for near-term projections and for scenarios of continuously increasing radiative forcing. For this reason, the recent Fifth Assessment Report of the Intergovernmental Panel on Climate Change (IPCC AR5) included such observationally constrained projections in its assessment of warming to 2035, but used raw model projections of longer term warming to 2100. Here a simple approach to weighting model projections based on an observational constraint is proposed which does not assume a linear relationship between past and future changes. This approach is used to weight model projections of warming in 2081-2100 relative to 1986-2005 under the Representative Concentration Pathway 4.5 forcing scenario, based on an observationally constrained estimate of the Transient Climate Response derived from a detection and attribution analysis. The resulting observationally constrained 5-95% warming range of 0.8-2.5 K is somewhat lower than the unweighted range of 1.1-2.6 K reported in the IPCC AR5. © 2015 The Authors.
NASA Astrophysics Data System (ADS)
Özer, Ahmet Özkan
2016-04-01
An infinite dimensional model for a three-layer active constrained layer (ACL) beam model, consisting of a piezoelectric elastic layer at the top and an elastic host layer at the bottom constraining a viscoelastic layer in the middle, is obtained for clamped-free boundary conditions by using a thorough variational approach. The Rao-Nakra thin compliant layer approximation is adopted to model the sandwich structure, and the electrostatic approach (magnetic effects are ignored) is assumed for the piezoelectric layer. Instead of the voltage actuation of the piezoelectric layer, the piezoelectric layer is proposed to be activated by a charge (or current) source. We show that, the closed-loop system with all mechanical feedback is shown to be uniformly exponentially stable. Our result is the outcome of the compact perturbation argument and a unique continuation result for the spectral problem which relies on the multipliers method. Finally, the modeling methodology of the paper is generalized to the multilayer ACL beams, and the uniform exponential stabilizability result is established analogously.
A method for fitting regression splines with varying polynomial order in the linear mixed model.
Edwards, Lloyd J; Stewart, Paul W; MacDougall, James E; Helms, Ronald W
2006-02-15
The linear mixed model has become a widely used tool for longitudinal analysis of continuous variables. The use of regression splines in these models offers the analyst additional flexibility in the formulation of descriptive analyses, exploratory analyses and hypothesis-driven confirmatory analyses. We propose a method for fitting piecewise polynomial regression splines with varying polynomial order in the fixed effects and/or random effects of the linear mixed model. The polynomial segments are explicitly constrained by side conditions for continuity and some smoothness at the points where they join. By using a reparameterization of this explicitly constrained linear mixed model, an implicitly constrained linear mixed model is constructed that simplifies implementation of fixed-knot regression splines. The proposed approach is relatively simple, handles splines in one variable or multiple variables, and can be easily programmed using existing commercial software such as SAS or S-plus. The method is illustrated using two examples: an analysis of longitudinal viral load data from a study of subjects with acute HIV-1 infection and an analysis of 24-hour ambulatory blood pressure profiles.
NASA Astrophysics Data System (ADS)
He, W.; Ju, W.; Chen, H.; Peters, W.; van der Velde, I.; Baker, I. T.; Andrews, A. E.; Zhang, Y.; Launois, T.; Campbell, J. E.; Suntharalingam, P.; Montzka, S. A.
2016-12-01
Carbonyl sulfide (OCS) is a promising novel atmospheric tracer for studying carbon cycle processes. OCS shares a similar pathway as CO2 during photosynthesis but not released through a respiration-like process, thus could be used to partition Gross Primary Production (GPP) from Net Ecosystem-atmosphere CO2 Exchange (NEE). This study uses joint atmospheric observations of OCS and CO2 to constrain GPP and ecosystem respiration (Re). Flask data from tower and aircraft sites over North America are collected. We employ our recently developed CarbonTracker (CT)-Lagrange carbon assimilation system, which is based on the CT framework and the Weather Research and Forecasting - Stochastic Time-Inverted Lagrangian Transport (WRF-STILT) model, and the Simple Biosphere model with simulated OCS (SiB3-OCS) that provides prior GPP, Re and plant uptake fluxes of OCS. Derived plant OCS fluxes from both process model and GPP-scaled model are tested in our inversion. To investigate the ability of OCS to constrain GPP and understand the uncertainty propagated from OCS modeling errors to constrained fluxes in a dual-tracer system including OCS and CO2, two inversion schemes are implemented and compared: (1) a two-step scheme, which firstly optimizes GPP using OCS observations, and then simultaneously optimizes GPP and Re using CO2 observations with OCS-constrained GPP in the first step as prior; (2) a joint scheme, which simultaneously optimizes GPP and Re using OCS and CO2 observations. We will evaluate the result using an estimated GPP from space-borne solar-induced fluorescence observations and a data-driven GPP upscaled from FLUXNET data with a statistical model (Jung et al., 2011). Preliminary result for the year 2010 shows the joint inversion makes simulated mole fractions more consistent with observations for both OCS and CO2. However, the uncertainty of OCS simulation is larger than that of CO2. The two-step and joint schemes perform similarly in improving the consistence with observations for OCS, implicating that OCS could provide independent constraint in joint inversion. Optimization makes less total GPP and Re but more NEE, when testing with prior CO2 fluxes from two biosphere models. This study gives an in-depth insight into the role of joint atmospheric OCS and CO2 observations in constraining CO2 fluxes.
NASA Astrophysics Data System (ADS)
Zhang, Chenglong; Zhang, Fan; Guo, Shanshan; Liu, Xiao; Guo, Ping
2018-01-01
An inexact nonlinear mλ-measure fuzzy chance-constrained programming (INMFCCP) model is developed for irrigation water allocation under uncertainty. Techniques of inexact quadratic programming (IQP), mλ-measure, and fuzzy chance-constrained programming (FCCP) are integrated into a general optimization framework. The INMFCCP model can deal with not only nonlinearities in the objective function, but also uncertainties presented as discrete intervals in the objective function, variables and left-hand side constraints and fuzziness in the right-hand side constraints. Moreover, this model improves upon the conventional fuzzy chance-constrained programming by introducing a linear combination of possibility measure and necessity measure with varying preference parameters. To demonstrate its applicability, the model is then applied to a case study in the middle reaches of Heihe River Basin, northwest China. An interval regression analysis method is used to obtain interval crop water production functions in the whole growth period under uncertainty. Therefore, more flexible solutions can be generated for optimal irrigation water allocation. The variation of results can be examined by giving different confidence levels and preference parameters. Besides, it can reflect interrelationships among system benefits, preference parameters, confidence levels and the corresponding risk levels. Comparison between interval crop water production functions and deterministic ones based on the developed INMFCCP model indicates that the former is capable of reflecting more complexities and uncertainties in practical application. These results can provide more reliable scientific basis for supporting irrigation water management in arid areas.
Charge redistribution in QM:QM ONIOM model systems: a constrained density functional theory approach
NASA Astrophysics Data System (ADS)
Beckett, Daniel; Krukau, Aliaksandr; Raghavachari, Krishnan
2017-11-01
The ONIOM hybrid method has found considerable success in QM:QM studies designed to approximate a high level of theory at a significantly reduced cost. This cost reduction is achieved by treating only a small model system with the target level of theory and the rest of the system with a low, inexpensive, level of theory. However, the choice of an appropriate model system is a limiting factor in ONIOM calculations and effects such as charge redistribution across the model system boundary must be considered as a source of error. In an effort to increase the general applicability of the ONIOM model, a method to treat the charge redistribution effect is developed using constrained density functional theory (CDFT) to constrain the charge experienced by the model system in the full calculation to the link atoms in the truncated model system calculations. Two separate CDFT-ONIOM schemes are developed and tested on a set of 20 reactions with eight combinations of levels of theory. It is shown that a scheme using a scaled Lagrange multiplier term obtained from the low-level CDFT model calculation outperforms ONIOM at each combination of levels of theory from 32% to 70%.
NASA Technical Reports Server (NTRS)
Hoflich, P.; Khokhlov, A. M.; Wheeler, J. C.
1995-01-01
We compute optical and infrared light curves of the pulsating class of delayed detonation models for Type Ia supernovae (SN Ia's) using an elaborate treatment of the Local Thermodynamic Equilbrium (LTE) radiation transport, equation of state and ionization balance, expansion opacity including the cooling by CO, Co(+), and SiO, and a Monte Carlo gamma-ray deposition scheme. The models have an amount of Ni-56 in the range from approximately or equal to 0.1 solar mass up to 0.7 solar mass depending on the density at which the transition from a deflagration to a detonation occurs. Models with a large nickel production give light curves comparable to those of typical Type Ia supernovae. Subluminous supernovae can be explained by models with a low nickel production. Multiband light curves are presented in comparison with the normally bright event SN 1992bc and the subluminous events Sn 1991bg and SN 1992bo to establish the principle that the delayed detonation paradigm in Chandrasekhar mass models may give a common explosion mechanism accounting for both normal and subluminous SN Ia's. Secondary IR-maxima are formed in the models of normal SN Ia's as a photospheric effect if the photospheric radius continues to increase well after maximum light. Secondary maxima appear later and stronger in models with moderate expansion velocities and with radioactive material closer to the surface. Model light curves for subluminous SN Ia's tend to show only one 'late' IR-maximum. In some delayed detonation models shell-like envelopes form, which consist of unburned carbon and oxygen. The formation of molecules in these envelopes is addressed. If the model retains a C/O-envelope and is subluminous, strong vibration bands of CO may appear, typically several weeks past maximum light. CO should be very weak or absent in normal Sn Ia's.
Tharakaraman, Kannan; Watanabe, Satoru; Chan, Kuan Rong; Huan, Jia; Subramanian, Vidya; Chionh, Yok Hian; Raguram, Aditya; Quinlan, Devin; McBee, Megan; Ong, Eugenia Z; Gan, Esther S; Tan, Hwee Cheng; Tyagi, Anu; Bhushan, Shashi; Lescar, Julien; Vasudevan, Subhash G; Ooi, Eng Eong; Sasisekharan, Ram
2018-05-09
Following the recent emergence of Zika virus (ZIKV), many murine and human neutralizing anti-ZIKV antibodies have been reported. Given the risk of virus escape mutants, engineering antibodies that target mutationally constrained epitopes with therapeutically relevant potencies can be valuable for combating future outbreaks. Here, we applied computational methods to engineer an antibody, ZAb_FLEP, that targets a highly networked and therefore mutationally constrained surface formed by the envelope protein dimer. ZAb_FLEP neutralized a breadth of ZIKV strains and protected mice in distinct in vivo models, including resolving vertical transmission and fetal mortality in infected pregnant mice. Serial passaging of ZIKV in the presence of ZAb_FLEP failed to generate viral escape mutants, suggesting that its epitope is indeed mutationally constrained. A single-particle cryo-EM reconstruction of the Fab-ZIKV complex validated the structural model and revealed insights into ZAb_FLEP's neutralization mechanism. ZAb_FLEP has potential as a therapeutic in future outbreaks. Copyright © 2018. Published by Elsevier Inc.
Liao, Bolin; Zhang, Yunong; Jin, Long
2016-02-01
In this paper, a new Taylor-type numerical differentiation formula is first presented to discretize the continuous-time Zhang neural network (ZNN), and obtain higher computational accuracy. Based on the Taylor-type formula, two Taylor-type discrete-time ZNN models (termed Taylor-type discrete-time ZNNK and Taylor-type discrete-time ZNNU models) are then proposed and discussed to perform online dynamic equality-constrained quadratic programming. For comparison, Euler-type discrete-time ZNN models (called Euler-type discrete-time ZNNK and Euler-type discrete-time ZNNU models) and Newton iteration, with interesting links being found, are also presented. It is proved herein that the steady-state residual errors of the proposed Taylor-type discrete-time ZNN models, Euler-type discrete-time ZNN models, and Newton iteration have the patterns of O(h(3)), O(h(2)), and O(h), respectively, with h denoting the sampling gap. Numerical experiments, including the application examples, are carried out, of which the results further substantiate the theoretical findings and the efficacy of Taylor-type discrete-time ZNN models. Finally, the comparisons with Taylor-type discrete-time derivative model and other Lagrange-type discrete-time ZNN models for dynamic equality-constrained quadratic programming substantiate the superiority of the proposed Taylor-type discrete-time ZNN models once again.
Wallis, Thomas S. A.; Dorr, Michael; Bex, Peter J.
2015-01-01
Sensitivity to luminance contrast is a prerequisite for all but the simplest visual systems. To examine contrast increment detection performance in a way that approximates the natural environmental input of the human visual system, we presented contrast increments gaze-contingently within naturalistic video freely viewed by observers. A band-limited contrast increment was applied to a local region of the video relative to the observer's current gaze point, and the observer made a forced-choice response to the location of the target (≈25,000 trials across five observers). We present exploratory analyses showing that performance improved as a function of the magnitude of the increment and depended on the direction of eye movements relative to the target location, the timing of eye movements relative to target presentation, and the spatiotemporal image structure at the target location. Contrast discrimination performance can be modeled by assuming that the underlying contrast response is an accelerating nonlinearity (arising from a nonlinear transducer or gain control). We implemented one such model and examined the posterior over model parameters, estimated using Markov-chain Monte Carlo methods. The parameters were poorly constrained by our data; parameters constrained using strong priors taken from previous research showed poor cross-validated prediction performance. Atheoretical logistic regression models were better constrained and provided similar prediction performance to the nonlinear transducer model. Finally, we explored the properties of an extended logistic regression that incorporates both eye movement and image content features. Models of contrast transduction may be better constrained by incorporating data from both artificial and natural contrast perception settings. PMID:26057546
A probabilistic assessment of calcium carbonate export and dissolution in the modern ocean
NASA Astrophysics Data System (ADS)
Battaglia, Gianna; Steinacher, Marco; Joos, Fortunat
2016-05-01
The marine cycle of calcium carbonate (CaCO3) is an important element of the carbon cycle and co-governs the distribution of carbon and alkalinity within the ocean. However, CaCO3 export fluxes and mechanisms governing CaCO3 dissolution are highly uncertain. We present an observationally constrained, probabilistic assessment of the global and regional CaCO3 budgets. Parameters governing pelagic CaCO3 export fluxes and dissolution rates are sampled using a Monte Carlo scheme to construct a 1000-member ensemble with the Bern3D ocean model. Ensemble results are constrained by comparing simulated and observation-based fields of excess dissolved calcium carbonate (TA*). The minerals calcite and aragonite are modelled explicitly and ocean-sediment fluxes are considered. For local dissolution rates, either a strong or a weak dependency on CaCO3 saturation is assumed. In addition, there is the option to have saturation-independent dissolution above the saturation horizon. The median (and 68 % confidence interval) of the constrained model ensemble for global biogenic CaCO3 export is 0.90 (0.72-1.05) Gt C yr-1, that is within the lower half of previously published estimates (0.4-1.8 Gt C yr-1). The spatial pattern of CaCO3 export is broadly consistent with earlier assessments. Export is large in the Southern Ocean, the tropical Indo-Pacific, the northern Pacific and relatively small in the Atlantic. The constrained results are robust across a range of diapycnal mixing coefficients and, thus, ocean circulation strengths. Modelled ocean circulation and transport timescales for the different set-ups were further evaluated with CFC11 and radiocarbon observations. Parameters and mechanisms governing dissolution are hardly constrained by either the TA* data or the current compilation of CaCO3 flux measurements such that model realisations with and without saturation-dependent dissolution achieve skill. We suggest applying saturation-independent dissolution rates in Earth system models to minimise computational costs.
NASA Astrophysics Data System (ADS)
Peylin, P. P.; Bacour, C.; MacBean, N.; Maignan, F.; Bastrikov, V.; Chevallier, F.
2017-12-01
Predicting the fate of carbon stocks and their sensitivity to climate change and land use/management strongly relies on our ability to accurately model net and gross carbon fluxes. However, simulated carbon and water fluxes remain subject to large uncertainties, partly because of unknown or poorly calibrated parameters. Over the past ten years, the carbon cycle data assimilation system at the Laboratoire des Sciences du Climat et de l'Environnement has investigated the benefit of assimilating multiple carbon cycle data streams into the ORCHIDEE LSM, the land surface component of the Institut Pierre Simon Laplace Earth System Model. These datasets have included FLUXNET eddy covariance data (net CO2 flux and latent heat flux) to constrain hourly to seasonal time-scale carbon cycle processes, remote sensing of the vegetation activity (MODIS NDVI) to constrain the leaf phenology, biomass data to constrain "slow" (yearly to decadal) processes of carbon allocation, and atmospheric CO2 concentrations to provide overall large scale constraints on the land carbon sink. Furthermore, we have investigated technical issues related to multiple data stream assimilation and choice of optimization algorithm. This has provided a wide-ranging perspective on the challenges we face in constraining model parameters and thus better quantifying, and reducing, model uncertainty in projections of the future global carbon sink. We review our past studies in terms of the impact of the optimization on key characteristics of the carbon cycle, e.g. the partition of the northern latitudes vs tropical land carbon sink, and compare to the classic atmospheric flux inversion approach. Throughout, we discuss our work in context of the abovementioned challenges, and propose solutions for the community going forward, including the potential of new observations such as atmospheric COS concentrations and satellite-derived Solar Induced Fluorescence to constrain the gross carbon fluxes of the ORCHIDEE model.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gupta, Anuradha; Arun, K. G.; Sathyaprakash, B. S., E-mail: axg645@psu.edu, E-mail: kgarun@cmi.ac.in, E-mail: bss25@psu.edu
We show that the inferred merger rate and chirp masses of binary black holes (BBHs) detected by advanced LIGO (aLIGO) can be used to constrain the rate of double neutron star (DNS) and neutron star–black hole (NSBH) mergers in the universe. We explicitly demonstrate this by considering a set of publicly available population synthesis models of Dominik et al. and show that if all the BBH mergers, GW150914, LVT151012, GW151226, and GW170104, observed by aLIGO arise from isolated binary evolution, the predicted DNS merger rate may be constrained to be 2.3–471.0 Gpc{sup −3} yr{sup −1} and that of NSBH mergersmore » will be constrained to 0.2–48.5 Gpc{sup −3} yr{sup −1}. The DNS merger rates are not constrained much, but the NSBH rates are tightened by a factor of ∼4 as compared to their previous rates. Note that these constrained DNS and NSBH rates are extremely model-dependent and are compared to the unconstrained values 2.3–472.5 Gpc{sup −3} yr{sup −1} and 0.2–218 Gpc{sup −3} yr{sup −1}, respectively, using the same models of Dominik et al. (2012a). These rate estimates may have implications for short Gamma Ray Burst progenitor models assuming they are powered (solely) by DNS or NSBH mergers. While these results are based on a set of open access population synthesis models, which may not necessarily be the representative ones, the proposed method is very general and can be applied to any number of models, thereby yielding more realistic constraints on the DNS and NSBH merger rates from the inferred BBH merger rate and chirp mass.« less
Quantifying How Observations Inform a Numerical Reanalysis of Hawaii
NASA Astrophysics Data System (ADS)
Powell, B. S.
2017-11-01
When assimilating observations into a model via state-estimation, it is possible to quantify how each observation changes the modeled estimate of a chosen oceanic metric. Using an existing 2 year reanalysis of Hawaii that includes more than 31 million observations from satellites, ships, SeaGliders, and autonomous floats, I assess which observations most improve the estimates of the transport and eddy kinetic energy. When the SeaGliders were in the water, they comprised less than 2.5% of the data, but accounted for 23% of the transport adjustment. Because the model physics constrains advanced state-estimation, the prescribed covariances are propagated in time to identify observation-model covariance. I find that observations that constrain the isopycnal tilt across the transport section provide the greatest impact in the analysis. In the case of eddy kinetic energy, observations that constrain the surface-driven upper ocean have more impact. This information can help to identify optimal sampling strategies to improve both state-estimates and forecasts.
Constraining dark sector perturbations I: cosmic shear and CMB lensing
NASA Astrophysics Data System (ADS)
Battye, Richard A.; Moss, Adam; Pearson, Jonathan A.
2015-04-01
We present current and future constraints on equations of state for dark sector perturbations. The equations of state considered are those corresponding to a generalized scalar field model and time-diffeomorphism invariant Script L(g) theories that are equivalent to models of a relativistic elastic medium and also Lorentz violating massive gravity. We develop a theoretical understanding of the observable impact of these models. In order to constrain these models we use CMB temperature data from Planck, BAO measurements, CMB lensing data from Planck and the South Pole Telescope, and weak galaxy lensing data from CFHTLenS. We find non-trivial exclusions on the range of parameters, although the data remains compatible with w=-1. We gauge how future experiments will help to constrain the parameters. This is done via a likelihood analysis for CMB experiments such as CoRE and PRISM, and tomographic galaxy weak lensing surveys, focussing in on the potential discriminatory power of Euclid on mildly non-linear scales.
NASA Technical Reports Server (NTRS)
Carpenter, J. R.; Markley, F. L.; Alfriend, K. T.; Wright, C.; Arcido, J.
2011-01-01
Sequential probability ratio tests explicitly allow decision makers to incorporate false alarm and missed detection risks, and are potentially less sensitive to modeling errors than a procedure that relies solely on a probability of collision threshold. Recent work on constrained Kalman filtering has suggested an approach to formulating such a test for collision avoidance maneuver decisions: a filter bank with two norm-inequality-constrained epoch-state extended Kalman filters. One filter models 1he null hypothesis 1ha1 the miss distance is inside the combined hard body radius at the predicted time of closest approach, and one filter models the alternative hypothesis. The epoch-state filter developed for this method explicitly accounts for any process noise present in the system. The method appears to work well using a realistic example based on an upcoming highly-elliptical orbit formation flying mission.
Constraint-Based Local Search for Constrained Optimum Paths Problems
NASA Astrophysics Data System (ADS)
Pham, Quang Dung; Deville, Yves; van Hentenryck, Pascal
Constrained Optimum Path (COP) problems arise in many real-life applications and are ubiquitous in communication networks. They have been traditionally approached by dedicated algorithms, which are often hard to extend with side constraints and to apply widely. This paper proposes a constraint-based local search (CBLS) framework for COP applications, bringing the compositionality, reuse, and extensibility at the core of CBLS and CP systems. The modeling contribution is the ability to express compositional models for various COP applications at a high level of abstraction, while cleanly separating the model and the search procedure. The main technical contribution is a connected neighborhood based on rooted spanning trees to find high-quality solutions to COP problems. The framework, implemented in COMET, is applied to Resource Constrained Shortest Path (RCSP) problems (with and without side constraints) and to the edge-disjoint paths problem (EDP). Computational results show the potential significance of the approach.
NASA Astrophysics Data System (ADS)
Baisden, W. T.
2011-12-01
Time-series radiocarbon measurements have substantial ability to constrain the size and residence time of the soil C pools commonly represented in ecosystem models. Radiocarbon remains unique in the ability to constrain the large stabilized C pool with decadal residence times. Radiocarbon also contributes usefully to constraining the size and turnover rate of the passive pool, but typically struggles to constrain pools with residence times less than a few years. Overall, the number of pools and associated turnover rates that can be constrained depends upon the number of time-series samples available, the appropriateness of chemical or physical fractions to isolate unequivocal pools, and the utility of additional C flux data to provide additional constraints. In New Zealand pasture soils, we demonstrate the ability to constrain decadal turnover times with in a few years for the stabilized pool and reasonably constrain the passive fraction. Good constraint is obtained with two time-series samples spaced 10 or more years apart after 1970. Three or more time-series samples further improve the level of constraint. Work within this context shows that a two-pool model does explain soil radiocarbon data for the most detailed profiles available (11 time-series samples), and identifies clear and consistent differences in rates of C turnover and passive fraction in Andisols vs Non-Andisols. Furthermore, samples from multiple horizons can commonly be combined, yielding consistent residence times and passive fraction estimates that are stable with, or increase with, depth in different sites. Radiocarbon generally fails to quantify rapid C turnover, however. Given that the strength of radiocarbon is estimating the size and turnover of the stabilized (decadal) and passive (millennial) pools, the magnitude of fast cycling pool(s) can be estimated by subtracting the radiocarbon-based estimates of turnover within stabilized and passive pools from total estimates of NPP. In grazing land, these estimates can be derived primarily from measured aboveground NPP and calculated belowground NPP. Results suggest that only 19-36% of heterotrophic soil respiration is derived from the soil C with rapid turnover times. A final logical step in synthesis is the analysis of temporal variation in NPP, primarily due to climate, as driver of changes in plant inputs and resulting in dynamic changes in rapid and decadal soil C pools. In sites with good time series samples from 1959-1975, we examine the apparent impacts of measured or modelled (Biome-BGC) NPP on soil Δ14C. Ultimately, these approaches have the ability to empirically constrain, and provide limited verification, of the soil C cycle as commonly depicted ecosystem biogeochemistry models.
Phase-field model of domain structures in ferroelectric thin films
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, Y. L.; Hu, S. Y.; Liu, Z. K.
A phase-field model for predicting the coherent microstructure evolution in constrained thin films is developed. It employs an analytical elastic solution derived for a constrained film with arbitrary eigenstrain distributions. The domain structure evolution during a cubic{r_arrow}tetragonal proper ferroelectric phase transition is studied. It is shown that the model is able to simultaneously predict the effects of substrate constraint and temperature on the volume fractions of domain variants, domain-wall orientations, domain shapes, and their temporal evolution. {copyright} 2001 American Institute of Physics.
Flexible Energy Scheduling Tool for Integrating Variable Generation | Grid
, security-constrained economic dispatch, and automatic generation control programs. DOWNLOAD PAPER Electric commitment, security-constrained economic dispatch, and automatic generation control sub-models. Each sub resolutions and operating strategies can be explored. FESTIV produces not only economic metrics but also
Epoch of reionization 21 cm forecasting from MCMC-constrained semi-numerical models
NASA Astrophysics Data System (ADS)
Hassan, Sultan; Davé, Romeel; Finlator, Kristian; Santos, Mario G.
2017-06-01
The recent low value of Planck Collaboration XLVII integrated optical depth to Thomson scattering suggests that the reionization occurred fairly suddenly, disfavouring extended reionization scenarios. This will have a significant impact on the 21 cm power spectrum. Using a semi-numerical framework, we improve our model from instantaneous to include time-integrated ionization and recombination effects, and find that this leads to more sudden reionization. It also yields larger H II bubbles that lead to an order of magnitude more 21 cm power on large scales, while suppressing the small-scale ionization power. Local fluctuations in the neutral hydrogen density play the dominant role in boosting the 21 cm power spectrum on large scales, while recombinations are subdominant. We use a Monte Carlo Markov chain approach to constrain our model to observations of the star formation rate functions at z = 6, 7, 8 from Bouwens et al., the Planck Collaboration XLVII optical depth measurements and the Becker & Bolton ionizing emissivity data at z ˜ 5. We then use this constrained model to perform 21 cm forecasting for Low Frequency Array, Hydrogen Epoch of Reionization Array and Square Kilometre Array in order to determine how well such data can characterize the sources driving reionization. We find that the Mock 21 cm power spectrum alone can somewhat constrain the halo mass dependence of ionizing sources, the photon escape fraction and ionizing amplitude, but combining the Mock 21 cm data with other current observations enables us to separately constrain all these parameters. Our framework illustrates how the future 21 cm data can play a key role in understanding the sources and topology of reionization as observations improve.
Radiolytic and Thermal Process Relevant to Dry Storage of Spent Nuclear Fuels
DOE Office of Scientific and Technical Information (OSTI.GOV)
Marschman, Steven C.; Haustein, Peter E.; Madey, Theodore E.
1999-06-01
This project involves basic research in chemistry and physics aimed at providing information pertinent to the safe long-term dry storage of spent nuclear fuel (SNF), thousands of tons of which remain in water storage across the DOE complex. The Hanford Site K-Basins alone hold 2300 tons of spent fuel, much of it severely corroded, and similar situations exist at Savannah River and Idaho National Engineering and Environmental Laboratory. DOE plans to remove this fuel and seal it in overpack canisters for ''dry'' interim storage for up to 75 years while awaiting permanent disposition. Chemically bound water will remain in thismore » fuel even after the proposed drying steps, leading to possible long-term corrosion of the containers and/or fuel rods themselves, generation of H2 and O2 gas via radiolysis (which could lead to deflagration or detonation), and reactions of pyrophoric uranium hydrides. No thoroughly tested model is now available to predict fuel behavior during preprocessing, processing, or storage. In a collaborative effort among Rutgers University, Pacific Northwest National Laboratory, and Brookhaven National Laboratory, we are studying the radiolytic reaction, drying processes, and corrosion behavior of actual SNF materials and of pure and mixed-phase samples. We propose to determine what is omitted from current models: radiolysis of water adsorbed on or in hydrates or hydroxides, thermodynamics of interfacial phases, and kinetics of drying. A model will be developed and tested against actual fuel rod behavior to ensure validity and applicability to the problems associated with developing dry storage strategies for DOE-owned SNF.« less
Dynamic Parameters of the 2015 Nepal Gorkha Mw7.8 Earthquake Constrained by Multi-observations
NASA Astrophysics Data System (ADS)
Weng, H.; Yang, H.
2017-12-01
Dynamic rupture model can provide much detailed insights into rupture physics that is capable of assessing future seismic risk. Many studies have attempted to constrain the slip-weakening distance, an important parameter controlling friction behavior of rock, for several earthquakes based on dynamic models, kinematic models, and direct estimations from near-field ground motion. However, large uncertainties of the values of the slip-weakening distance still remain, mostly because of the intrinsic trade-offs between the slip-weakening distance and fault strength. Here we use a spontaneously dynamic rupture model to constrain the frictional parameters of the 25 April 2015 Mw7.8 Nepal earthquake, by combining with multiple seismic observations such as high-rate cGPS data, strong motion data, and kinematic source models. With numerous tests we find the trade-off patterns of final slip, rupture speed, static GPS ground displacements, and dynamic ground waveforms are quite different. Combining all the seismic constraints we can conclude a robust solution without a substantial trade-off of average slip-weakening distance, 0.6 m, in contrast to previous kinematical estimation of 5 m. To our best knowledge, this is the first time to robustly determine the slip-weakening distance on seismogenic fault from seismic observations. The well-constrained frictional parameters may be used for future dynamic models to assess seismic hazard, such as estimating the peak ground acceleration (PGA) etc. Similar approach could also be conducted for other great earthquakes, enabling broad estimations of the dynamic parameters in global perspectives that can better reveal the intrinsic physics of earthquakes.
Bassen, David M; Vilkhovoy, Michael; Minot, Mason; Butcher, Jonathan T; Varner, Jeffrey D
2017-01-25
Ensemble modeling is a promising approach for obtaining robust predictions and coarse grained population behavior in deterministic mathematical models. Ensemble approaches address model uncertainty by using parameter or model families instead of single best-fit parameters or fixed model structures. Parameter ensembles can be selected based upon simulation error, along with other criteria such as diversity or steady-state performance. Simulations using parameter ensembles can estimate confidence intervals on model variables, and robustly constrain model predictions, despite having many poorly constrained parameters. In this software note, we present a multiobjective based technique to estimate parameter or models ensembles, the Pareto Optimal Ensemble Technique in the Julia programming language (JuPOETs). JuPOETs integrates simulated annealing with Pareto optimality to estimate ensembles on or near the optimal tradeoff surface between competing training objectives. We demonstrate JuPOETs on a suite of multiobjective problems, including test functions with parameter bounds and system constraints as well as for the identification of a proof-of-concept biochemical model with four conflicting training objectives. JuPOETs identified optimal or near optimal solutions approximately six-fold faster than a corresponding implementation in Octave for the suite of test functions. For the proof-of-concept biochemical model, JuPOETs produced an ensemble of parameters that gave both the mean of the training data for conflicting data sets, while simultaneously estimating parameter sets that performed well on each of the individual objective functions. JuPOETs is a promising approach for the estimation of parameter and model ensembles using multiobjective optimization. JuPOETs can be adapted to solve many problem types, including mixed binary and continuous variable types, bilevel optimization problems and constrained problems without altering the base algorithm. JuPOETs is open source, available under an MIT license, and can be installed using the Julia package manager from the JuPOETs GitHub repository.
The Measurement of Hot-Spots in Granulated Ammonium Nitrate
NASA Astrophysics Data System (ADS)
Proud, W. G.
2002-07-01
Ammonium Nitrate (AN) is one of the components of the most widely used explosive in the world namely, ammonium nitrate: fuel oil mixtures (ANFO). By itself, it is an oxygen positive explosive with a large critical diameter. Hot-spots are produced in explosives by various means including gas space collapse, localised shear or friction. If these hot-spots reach critical conditions of size, temperature and duration reaction can grow. This deflagration stage may eventually transition to detonation. This paper describes high-speed image-intensified photography study in which the number and growth of hot spots in granular AN are monitored for a range of different impact pressures. The results can be used in detonation codes to provide a more accurate and realistic description of the initiation process.
The Measurement of Hot-spots in Granulated Ammonium Nitrate
NASA Astrophysics Data System (ADS)
Proud, William; Field, John
2001-06-01
Ammonium Nitrate (AN) is one of the components of the most widely used explosive in the world ammonium nitrate: fuel oil mixtures (ANFO). By itself, it is an oxygen negative explosive with a large critical diameter. Hot-spots are produced in explosives by various means including gas space collapse, localised shear or friction. If these hot-spots reach critical conditions of size, temperature and duration size reaction can grow. This deflagration stage may eventually transition to detonation. This paper describes a system and presents results where high-speed image intensified photography is used to monitor the number and growth of hot spots in granular AN under a range of different impact pressures. The results can be used in detonation codes to provide a more accurate and realistic description of the initiation process.
Preliminary assessment of combustion modes for internal combustion wave rotors
NASA Technical Reports Server (NTRS)
Nalim, M. Razi
1995-01-01
Combustion within the channels of a wave rotor is examined as a means of obtaining pressure gain during heat addition in a gas turbine engine. Several modes of combustion are considered and the factors that determine the applicability of three modes are evaluated in detail; premixed autoignition/detonation, premixed deflagration, and non-premixed compression ignition. The last two will require strong turbulence for completion of combustion in a reasonable time in the wave rotor. The compression/autoignition modes will require inlet temperatures in excess of 1500 R for reliable ignition with most hydrocarbon fuels; otherwise, a supplementary ignition method must be provided. Examples of combustion mode selection are presented for two core engine applications that had been previously designed with equivalent 4-port wave rotor topping cycles using external combustion.
Characterization of the High-Albedo NEA 3691 Bede
NASA Technical Reports Server (NTRS)
Wooden, Diane H.; Lederer, Susan M.; Jehin, Emmanuel; Rozitis, Benjamin; Jefferson, Jeffrey D.; Nelson, Tyler W.; Dotson, Jessie L.; Ryan, Erin L.; Howell, Ellen S.; Fernandez, Yanga R.;
2016-01-01
Characterization of NEAs provides important inputs to models for atmospheric entry, risk assessment and mitigation. Diameter is a key parameter because diameter translates to kinetic energy in atmospheric entry. Diameters can be derived from the absolute magnitude, H(PA=0deg), and from thermal modeling of observed IR fluxes. For both methods, the albedo (pv) is important - high pv surfaces have cooler temperatures, larger diameters for a given Hmag, and shallower phase curves (larger slope parameter G). Thermal model parameters are coupled, however, so that a higher thermal inertia also results in a cooler surface temperature. Multiple parameters contribute to constraining the diameter. Observations made at multiple observing geometries can contribute to understanding the relationships between and potentially breaking some of the degeneracies between parameters. We present data and analyses on NEA 3691 Bede with the aim of best constraining the diameter and pv from a combination of thermal modeling and light curve analyses. We employ our UKIRT+Michelle mid-IR photometric observations of 3691 Bede's thermal emission at 2 phase angles (27&43 deg 2015-03-19 & 04-13), in addition to WISE data (33deg 2010-05-27, Mainzer+2011). Observing geometries differ by solar phase angles and by moderate changes in heliocentric distance (e.g., further distances produce somewhat cooler surface temperatures). With the NEATM model and for a constant IR beaming parameter (eta=constant), there is a family of solutions for (diameter, pv, G, eta) where G is the slope parameter from the H-G Relation. NEATM models employing Pravec+2012's choice of G=0.43, produce D=1.8 km and pv˜0.4, given that G=0.43 is assumed from studies of main belt asteroids (Warner+2009). We present an analysis of the light curve of 3691 Bede to constrain G from observations. We also investigate fitting thermophysical models (TPM, Rozitis+11) to constrain the coupled parameters of thermal inertia (Gamma) and surface roughness, which in turn affect diameter and pv. Surface composition can be related to pv. This study focuses on understanding and characterizing the dependency of parameters with the aim of constraining diameter, pv and thermal inertia for 3691 Bede.
Characterization of the high-albedo NEA 3691 Bede
NASA Astrophysics Data System (ADS)
Wooden, Diane H.; Lederer, Susan M.; Jehin, Emmanuel; Rozitis, Benjamin; Jefferson, Jeffrey D.; Nelson, Tyler W.; Dotson, Jessie L.; Ryan, Erin L.; Howell, Ellen S.; Fernandez, Yanga R.; Lovell, Amy J.; Woodward, Charles E.; Harker, David Emerson
2016-10-01
Characterization of NEAs provides important inputs to models for atmospheric entry, risk assessment and mitigation. Diameter is a key parameter because diameter translates to kinetic energy in atmospheric entry. Diameters can be derived from the absolute magnitude, H(PA=0deg), and from thermal modeling of observed IR fluxes. For both methods, the albedo (pv) is important - high pv surfaces have cooler temperatures, larger diameters for a given Hmag, and shallower phase curves (larger slope parameter G). Thermal model parameters are coupled, however, so that a higher thermal inertia also results in a cooler surface temperature. Multiple parameters contribute to constraining the diameter.Observations made at multiple observing geometries can contribute to understanding the relationships between and potentially breaking some of the degeneracies between parameters. We present data and analyses on NEA 3691 Bede with the aim of best constraining the diameter and pv from a combination of thermal modeling and light curve analyses. We employ our UKIRT+Michelle mid-IR photometric observations of 3691 Bede's thermal emission at 2 phase angles (27&43 deg 2015-03-19 & 04-13), in addition to WISE data (33deg 2010-05-27, Mainzer+2011).Observing geometries differ by solar phase angles and by moderate changes in heliocentric distance (e.g., further distances produce somewhat cooler surface temperatures). With the NEATM model and for a constant IR beaming parameter (eta=constant), there is a family of solutions for (diameter, pv, G, eta) where G is the slope parameter from the H-G Relation. NEATM models employing Pravec+2012's choice of G=0.43, produce D=1.8 km and pv≈0.4, given that G=0.43 is assumed from studies of main belt asteroids (Warner+2009). We present an analysis of the light curve of 3691 Bede to constrain G from observations. We also investigate fitting thermophysical models (TPM, Rozitis+11) to constrain the coupled parameters of thermal inertia (Gamma) and surface roughness, which in turn affect diameter and pv. Surface composition can be related to pv. This study focuses on understanding and characterizing the dependency of parameters with the aim of constraining diameter, pv and thermal inertia for 3691 Bede.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gigley, H.M.
1982-01-01
An artificial intelligence approach to the simulation of neurolinguistically constrained processes in sentence comprehension is developed using control strategies for simulation of cooperative computation in associative networks. The desirability of this control strategy in contrast to ATN and production system strategies is explained. A first pass implementation of HOPE, an artificial intelligence simulation model of sentence comprehension, constrained by studies of aphasic performance, psycholinguistics, neurolinguistics, and linguistic theory is described. Claims that the model could serve as a basis for sentence production simulation and for a model of language acquisition as associative learning are discussed. HOPE is a model thatmore » performs in a normal state and includes a lesion simulation facility. HOPE is also a research tool. Its modifiability and use as a tool to investigate hypothesized causes of degradation in comprehension performance by aphasic patients are described. Issues of using behavioral constraints in modelling and obtaining appropriate data for simulated process modelling are discussed. Finally, problems of validation of the simulation results are raised; and issues of how to interpret clinical results to define the evolution of the model are discussed. Conclusions with respect to the feasibility of artificial intelligence simulation process modelling are discussed based on the current state of research.« less
NASA Astrophysics Data System (ADS)
Delgado, F.; Kubanek, J.; Anderson, K. R.; Lundgren, P.; Pritchard, M. E.
2017-12-01
The 2011-2012 eruption of Cordón Caulle volcano in Chile is the best scientifically observed rhyodacitic eruption and is thus a key place to understand the dynamics of these rare but powerful explosive rhyodacitic eruptions. Because the volatile phase controls both the eruption temporal evolution and the eruptive style, either explosive or effusive, it is important to constrain the physical parameters that drive these eruptions. The eruption began explosively and after two weeks evolved into a hybrid explosive - lava flow effusion whose volume-time evolution we constrain with a series of TanDEM-X Digital Elevation Models. Our data shows the intrusion of a large volume laccolith or cryptodome during the first 2.5 months of the eruption and lava flow effusion only afterwards, with a total volume of 1.4 km3. InSAR data from the ENVISAT and TerraSAR-X missions shows more than 2 m of subsidence during the effusive eruption phase produced by deflation of a finite spheroidal source at a depth of 5 km. In order to constrain the magma total H2O content, crystal cargo, and reservoir pressure drop we numerically solve the coupled set of equations of a pressurized magma reservoir, magma conduit flow and time dependent density, volatile exsolution and viscosity that we use to invert the InSAR and topographic data time series. We compare the best-fit model parameters with independent estimates of magma viscosity and total gas content measured from lava samples. Preliminary modeling shows that although it is not possible to model both the InSAR and the topographic data during the onset of the laccolith emplacement, it is possible to constrain the magma H2O and crystal content, to 4% wt and 30% which agree well with published literature values.
Constrained Versions of DEDICOM for Use in Unsupervised Part-Of-Speech Tagging
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dunlavy, Daniel; Chew, Peter A.
This reports describes extensions of DEDICOM (DEcomposition into DIrectional COMponents) data models [3] that incorporate bound and linear constraints. The main purpose of these extensions is to investigate the use of improved data models for unsupervised part-of-speech tagging, as described by Chew et al. [2]. In that work, a single domain, two-way DEDICOM model was computed on a matrix of bigram fre- quencies of tokens in a corpus and used to identify parts-of-speech as an unsupervised approach to that problem. An open problem identi ed in that work was the com- putation of a DEDICOM model that more closely resembledmore » the matrices used in a Hidden Markov Model (HMM), speci cally through post-processing of the DEDICOM factor matrices. The work reported here consists of the description of several models that aim to provide a direct solution to that problem and a way to t those models. The approach taken here is to incorporate the model requirements as bound and lin- ear constrains into the DEDICOM model directly and solve the data tting problem as a constrained optimization problem. This is in contrast to the typical approaches in the literature, where the DEDICOM model is t using unconstrained optimization approaches, and model requirements are satis ed as a post-processing step.« less
NASA Astrophysics Data System (ADS)
Martinez, Guillermo F.; Gupta, Hoshin V.
2011-12-01
Methods to select parsimonious and hydrologically consistent model structures are useful for evaluating dominance of hydrologic processes and representativeness of data. While information criteria (appropriately constrained to obey underlying statistical assumptions) can provide a basis for evaluating appropriate model complexity, it is not sufficient to rely upon the principle of maximum likelihood (ML) alone. We suggest that one must also call upon a "principle of hydrologic consistency," meaning that selected ML structures and parameter estimates must be constrained (as well as possible) to reproduce desired hydrological characteristics of the processes under investigation. This argument is demonstrated in the context of evaluating the suitability of candidate model structures for lumped water balance modeling across the continental United States, using data from 307 snow-free catchments. The models are constrained to satisfy several tests of hydrologic consistency, a flow space transformation is used to ensure better consistency with underlying statistical assumptions, and information criteria are used to evaluate model complexity relative to the data. The results clearly demonstrate that the principle of consistency provides a sensible basis for guiding selection of model structures and indicate strong spatial persistence of certain model structures across the continental United States. Further work to untangle reasons for model structure predominance can help to relate conceptual model structures to physical characteristics of the catchments, facilitating the task of prediction in ungaged basins.
On the nullspace of TLS multi-station adjustment
NASA Astrophysics Data System (ADS)
Sterle, Oskar; Kogoj, Dušan; Stopar, Bojan; Kregar, Klemen
2018-07-01
In the article we present an analytic aspect of TLS multi-station least-squares adjustment with the main focus on the datum problem. The datum problem is, compared to previously published researches, theoretically analyzed and solved, where the solution is based on nullspace derivation of the mathematical model. The importance of datum problem solution is seen in a complete description of TLS multi-station adjustment solutions from a set of all minimally constrained least-squares solutions. On a basis of known nullspace, estimable parameters are described and the geometric interpretation of all minimally constrained least squares solutions is presented. At the end a simulated example is used to analyze the results of TLS multi-station minimally constrained and inner constrained least-squares adjustment solutions.
Antunes, J; Debut, V
2017-02-01
Most musical instruments consist of dynamical subsystems connected at a number of constraining points through which energy flows. For physical sound synthesis, one important difficulty deals with enforcing these coupling constraints. While standard techniques include the use of Lagrange multipliers or penalty methods, in this paper, a different approach is explored, the Udwadia-Kalaba (U-K) formulation, which is rooted on analytical dynamics but avoids the use of Lagrange multipliers. This general and elegant formulation has been nearly exclusively used for conceptual systems of discrete masses or articulated rigid bodies, namely, in robotics. However its natural extension to deal with continuous flexible systems is surprisingly absent from the literature. Here, such a modeling strategy is developed and the potential of combining the U-K equation for constrained systems with the modal description is shown, in particular, to simulate musical instruments. Objectives are twofold: (1) Develop the U-K equation for constrained flexible systems with subsystems modelled through unconstrained modes; and (2) apply this framework to compute string/body coupled dynamics. This example complements previous work [Debut, Antunes, Marques, and Carvalho, Appl. Acoust. 108, 3-18 (2016)] on guitar modeling using penalty methods. Simulations show that the proposed technique provides similar results with a significant improvement in computational efficiency.
Macho, Jorge Berzosa; Montón, Luis Gardeazabal; Rodriguez, Roberto Cortiñas
2017-08-01
The Cyber Physical Systems (CPS) paradigm is based on the deployment of interconnected heterogeneous devices and systems, so interoperability is at the heart of any CPS architecture design. In this sense, the adoption of standard and generic data formats for data representation and communication, e.g., XML or JSON, effectively addresses the interoperability problem among heterogeneous systems. Nevertheless, the verbosity of those standard data formats usually demands system resources that might suppose an overload for the resource-constrained devices that are typically deployed in CPS. In this work we present Context- and Template-based Compression (CTC), a data compression approach targeted to resource-constrained devices, which allows reducing the resources needed to transmit, store and process data models. Additionally, we provide a benchmark evaluation and comparison with current implementations of the Efficient XML Interchange (EXI) processor, which is promoted by the World Wide Web Consortium (W3C), and it is the most prominent XML compression mechanism nowadays. Interestingly, the results from the evaluation show that CTC outperforms EXI implementations in terms of memory usage and speed, keeping similar compression rates. As a conclusion, CTC is shown to be a good candidate for managing standard data model representation formats in CPS composed of resource-constrained devices.
Montón, Luis Gardeazabal
2017-01-01
The Cyber Physical Systems (CPS) paradigm is based on the deployment of interconnected heterogeneous devices and systems, so interoperability is at the heart of any CPS architecture design. In this sense, the adoption of standard and generic data formats for data representation and communication, e.g., XML or JSON, effectively addresses the interoperability problem among heterogeneous systems. Nevertheless, the verbosity of those standard data formats usually demands system resources that might suppose an overload for the resource-constrained devices that are typically deployed in CPS. In this work we present Context- and Template-based Compression (CTC), a data compression approach targeted to resource-constrained devices, which allows reducing the resources needed to transmit, store and process data models. Additionally, we provide a benchmark evaluation and comparison with current implementations of the Efficient XML Interchange (EXI) processor, which is promoted by the World Wide Web Consortium (W3C), and it is the most prominent XML compression mechanism nowadays. Interestingly, the results from the evaluation show that CTC outperforms EXI implementations in terms of memory usage and speed, keeping similar compression rates. As a conclusion, CTC is shown to be a good candidate for managing standard data model representation formats in CPS composed of resource-constrained devices. PMID:28763013
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cembranos, Jose A. R.; Diaz-Cruz, J. Lorenzo; Prado, Lilian
Dark Matter direct detection experiments are able to exclude interesting parameter space regions of particle models which predict an important amount of thermal relics. We use recent data to constrain the branon model and to compute the region that is favored by CDMS measurements. Within this work, we also update present colliders constraints with new studies coming from the LHC. Despite the present low luminosity, it is remarkable that for heavy branons, CMS and ATLAS measurements are already more constraining than previous analyses performed with TEVATRON and LEP data.
Influence of exothermic chemical reactions on laser-induced shock waves.
Gottfried, Jennifer L
2014-10-21
Differences in the excitation of non-energetic and energetic residues with a 900 mJ, 6 ns laser pulse (1064 nm) have been investigated. Emission from the laser-induced plasma of energetic materials (e.g. triaminotrinitrobenzene [TATB], cyclotrimethylene trinitramine [RDX], and hexanitrohexaazaisowurtzitane [CL-20]) is significantly reduced compared to non-energetic materials (e.g. sugar, melamine, and l-glutamine). Expansion of the resulting laser-induced shock wave into the air above the sample surface was imaged on a microsecond timescale with a high-speed camera recording multiple frames from each laser shot; the excitation of energetic materials produces larger heat-affected zones in the surrounding atmosphere (facilitating deflagration of particles ejected from the sample surface), results in the formation of additional shock fronts, and generates faster external shock front velocities (>750 m s(-1)) compared to non-energetic materials (550-600 m s(-1)). Non-explosive materials that undergo exothermic chemical reactions in air at high temperatures such as ammonium nitrate and magnesium sulfate produce shock velocities which exceed those of the inert materials but are less than those generated by the exothermic reactions of explosive materials (650-700 m s(-1)). The most powerful explosives produced the highest shock velocities. A comparison to several existing shock models demonstrated that no single model describes the shock propagation for both non-energetic and energetic materials. The influence of the exothermic chemical reactions initiated by the pulsed laser on the velocity of the laser-induced shock waves has thus been demonstrated for the first time.
Central Elemental Abundance Ratios In the Perseus Cluster: Resonant Scattering or SN Ia Enrichment?
NASA Technical Reports Server (NTRS)
Dupke, Renato A.; Arnaud, Keith; White, Nicholas E. (Technical Monitor)
2001-01-01
We have determined abundance ratios in the core of the Perseus Cluster for several elements. These ratios indicate a central dominance of Type 1a supernova (SN Ia) ejects similar to that found for A496, A2199 and A3571. Simultaneous analysis of ASCA spectra from SIS1, GIS2, and GIS3 shows that the ratio of Ni to Fe abundances is approx. 3.4 +/- 1.1 times solar within the central 4'. This ratio is consistent with (and more precise than) that observed in other clusters whose central regions are dominated by SN Ia ejecta. Such a large Ni overabundance is predicted by "convective deflagration" explosion models for SNe Ia such as W7 but is inconsistent with delayed detonation models. We note that with current instrumentation the Ni K(alpha) line is confused with Fe K(beta) and that the Ni overabundance we observe has been interpreted by others as an anomalously large ratio of Fe K(beta) to Fe K(alpha) caused by resonant scattering in the Fe K(alpha) line. We argue that a central enhancement of SN Ia ejecta and hence a high ratio of Ni to Fe abundances are naturally explained by scenarios that include the generation of chemical gradients by suppressed SN Ia winds or ram pressure stripping of cluster galaxies. It is not necessary to suppose that the intracluster gas is optically thick to resonant scattering of the Fe K(alpha) line.
Working Memory in Children: A Time-Constrained Functioning Similar to Adults
ERIC Educational Resources Information Center
Portrat, Sophie; Camos, Valerie; Barrouillet, Pierre
2009-01-01
Within the time-based resource-sharing (TBRS) model, we tested a new conception of the relationships between processing and storage in which the core mechanisms of working memory (WM) are time constrained. However, our previous studies were restricted to adults. The current study aimed at demonstrating that these mechanisms are present and…
NASA Astrophysics Data System (ADS)
Parsons, R. A.; Nimmo, F.
2010-03-01
SHARAD observations constrain the thickness and dust content of lobate debris aprons (LDAs). Simulations of dust-free ice-sheet flow over a flat surface at 205 K for 10-100 m.y. give LDA lengths and thicknesses that are consistent with observations.
Two-Phase Item Selection Procedure for Flexible Content Balancing in CAT
ERIC Educational Resources Information Center
Cheng, Ying; Chang, Hua-Hua; Yi, Qing
2007-01-01
Content balancing is an important issue in the design and implementation of computerized adaptive testing (CAT). Content-balancing techniques that have been applied in fixed content balancing, where the number of items from each content area is fixed, include constrained CAT (CCAT), the modified multinomial model (MMM), modified constrained CAT…
Ammonia (NH(3)has significant impacts on biodiversity, eutrophication, and acidification. Widespread uncertainty in the magnitude and seasonality of NH3 emissions hinders efforts to address these issues. In this work, we constrain U.S. NH3 sources using obse...
Constrained Kalman Filtering Via Density Function Truncation for Turbofan Engine Health Estimation
NASA Technical Reports Server (NTRS)
Simon, Dan; Simon, Donald L.
2006-01-01
Kalman filters are often used to estimate the state variables of a dynamic system. However, in the application of Kalman filters some known signal information is often either ignored or dealt with heuristically. For instance, state variable constraints (which may be based on physical considerations) are often neglected because they do not fit easily into the structure of the Kalman filter. This paper develops an analytic method of incorporating state variable inequality constraints in the Kalman filter. The resultant filter truncates the PDF (probability density function) of the Kalman filter estimate at the known constraints and then computes the constrained filter estimate as the mean of the truncated PDF. The incorporation of state variable constraints increases the computational effort of the filter but significantly improves its estimation accuracy. The improvement is demonstrated via simulation results obtained from a turbofan engine model. The turbofan engine model contains 3 state variables, 11 measurements, and 10 component health parameters. It is also shown that the truncated Kalman filter may be a more accurate way of incorporating inequality constraints than other constrained filters (e.g., the projection approach to constrained filtering).
Ren, Hai-Sheng; Ming, Mei-Jun; Ma, Jian-Yi; Li, Xiang-Yuan
2013-08-22
Within the framework of constrained density functional theory (CDFT), the diabatic or charge localized states of electron transfer (ET) have been constructed. Based on the diabatic states, inner reorganization energy λin has been directly calculated. For solvent reorganization energy λs, a novel and reasonable nonequilibrium solvation model is established by introducing a constrained equilibrium manipulation, and a new expression of λs has been formulated. It is found that λs is actually the cost of maintaining the residual polarization, which equilibrates with the extra electric field. On the basis of diabatic states constructed by CDFT, a numerical algorithm using the new formulations with the dielectric polarizable continuum model (D-PCM) has been implemented. As typical test cases, self-exchange ET reactions between tetracyanoethylene (TCNE) and tetrathiafulvalene (TTF) and their corresponding ionic radicals in acetonitrile are investigated. The calculated reorganization energies λ are 7293 cm(-1) for TCNE/TCNE(-) and 5939 cm(-1) for TTF/TTF(+) reactions, agreeing well with available experimental results of 7250 cm(-1) and 5810 cm(-1), respectively.
NASA Astrophysics Data System (ADS)
Hebert, Philippe; Saint-Amans, Charles
2013-06-01
A detailed description of the reaction rates and mechanisms occurring in shock-induced decomposition of condensed explosives is very important to improve the predictive capabilities of shock-to-detonation transition models. However, direct measurements of such experimental data are difficult to perform during detonation experiments. By coupling pulsed laser ignition of an explosive in a diamond anvil cell (DAC) with time-resolved streak camera recording of transmitted light, it is possible to make direct observations of deflagration phenomena at detonation pressure. We have developed an experimental set-up that allows combustion front propagation rates and time-resolved absorption spectroscopy measurements. The decomposition reactions are initiated using a nanosecond YAG laser and their kinetics is followed by time-resolved absorption spectroscopy. The results obtained for two explosives, nitromethane (NM) and HMX are presented in this paper. For NM, a change in reactivity is clearly seen around 25 GPa. Below this pressure, the reaction products are essentially carbon residues whereas at higher pressure, a transient absorption feature is first observed and is followed by the formation of a white amorphous product. For HMX, the evolution of the absorption as a function of time indicates a multi-step reaction mechanism which is found to depend on both the initial pressure and the laser fluence.
Technical Note: On the use of nudging for aerosol–climate model intercomparison studies
Zhang, K.; Wan, H.; Liu, X.; ...
2014-08-26
Nudging as an assimilation technique has seen increased use in recent years in the development and evaluation of climate models. Constraining the simulated wind and temperature fields using global weather reanalysis facilitates more straightforward comparison between simulation and observation, and reduces uncertainties associated with natural variabilities of the large-scale circulation. On the other hand, the forcing introduced by nudging can be strong enough to change the basic characteristics of the model climate. In the paper we show that for the Community Atmosphere Model version 5 (CAM5), due to the systematic temperature bias in the standard model and the sensitivity ofmore » simulated ice formation to anthropogenic aerosol concentration, nudging towards reanalysis results in substantial reductions in the ice cloud amount and the impact of anthropogenic aerosols on long-wave cloud forcing. In order to reduce discrepancies between the nudged and unconstrained simulations, and meanwhile take the advantages of nudging, two alternative experimentation methods are evaluated. The first one constrains only the horizontal winds. The second method nudges both winds and temperature, but replaces the long-term climatology of the reanalysis by that of the model. Results show that both methods lead to substantially improved agreement with the free-running model in terms of the top-of-atmosphere radiation budget and cloud ice amount. The wind-only nudging is more convenient to apply, and provides higher correlations of the wind fields, geopotential height and specific humidity between simulation and reanalysis. Results from both CAM5 and a second aerosol–climate model ECHAM6-HAM2 also indicate that compared to the wind-and-temperature nudging, constraining only winds leads to better agreement with the free-running model in terms of the estimated shortwave cloud forcing and the simulated convective activities. This suggests nudging the horizontal winds but not temperature is a good strategy for the investigation of aerosol indirect effects since it provides well-constrained meteorology without strongly perturbing the model's mean climate.« less
Technical Note: On the use of nudging for aerosol-climate model intercomparison studies
NASA Astrophysics Data System (ADS)
Zhang, K.; Wan, H.; Liu, X.; Ghan, S. J.; Kooperman, G. J.; Ma, P.-L.; Rasch, P. J.; Neubauer, D.; Lohmann, U.
2014-08-01
Nudging as an assimilation technique has seen increased use in recent years in the development and evaluation of climate models. Constraining the simulated wind and temperature fields using global weather reanalysis facilitates more straightforward comparison between simulation and observation, and reduces uncertainties associated with natural variabilities of the large-scale circulation. On the other hand, the forcing introduced by nudging can be strong enough to change the basic characteristics of the model climate. In the paper we show that for the Community Atmosphere Model version 5 (CAM5), due to the systematic temperature bias in the standard model and the sensitivity of simulated ice formation to anthropogenic aerosol concentration, nudging towards reanalysis results in substantial reductions in the ice cloud amount and the impact of anthropogenic aerosols on long-wave cloud forcing. In order to reduce discrepancies between the nudged and unconstrained simulations, and meanwhile take the advantages of nudging, two alternative experimentation methods are evaluated. The first one constrains only the horizontal winds. The second method nudges both winds and temperature, but replaces the long-term climatology of the reanalysis by that of the model. Results show that both methods lead to substantially improved agreement with the free-running model in terms of the top-of-atmosphere radiation budget and cloud ice amount. The wind-only nudging is more convenient to apply, and provides higher correlations of the wind fields, geopotential height and specific humidity between simulation and reanalysis. Results from both CAM5 and a second aerosol-climate model ECHAM6-HAM2 also indicate that compared to the wind-and-temperature nudging, constraining only winds leads to better agreement with the free-running model in terms of the estimated shortwave cloud forcing and the simulated convective activities. This suggests nudging the horizontal winds but not temperature is a good strategy for the investigation of aerosol indirect effects since it provides well-constrained meteorology without strongly perturbing the model's mean climate.
Numerical modeling of Drangajökull Ice Cap, NW Iceland
NASA Astrophysics Data System (ADS)
Anderson, Leif S.; Jarosch, Alexander H.; Flowers, Gwenn E.; Aðalgeirsdóttir, Guðfinna; Magnússon, Eyjólfur; Pálsson, Finnur; Muñoz-Cobo Belart, Joaquín; Þorsteinsson, Þorsteinn; Jóhannesson, Tómas; Sigurðsson, Oddur; Harning, David; Miller, Gifford H.; Geirsdóttir, Áslaug
2016-04-01
Over the past century the Arctic has warmed twice as fast as the global average. This discrepancy is likely due to feedbacks inherent to the Arctic climate system. These Arctic climate feedbacks are currently poorly quantified, but are essential to future climate predictions based on global circulation modeling. Constraining the magnitude and timing of past Arctic climate changes allows us to test climate feedback parameterizations at different times with different boundary conditions. Because Holocene Arctic summer temperature changes have been largest in the North Atlantic (Kaufman et al., 2004) we focus on constraining the paleoclimate of Iceland. Glaciers are highly sensitive to changes in temperature and precipitation amount. This sensitivity allows for the estimation of paleoclimate using glacier models, modern glacier mass balance data, and past glacier extents. We apply our model to the Drangajökull ice cap (~150 sq. km) in NW Iceland. Our numerical model is resolved in two-dimensions, conserves mass, and applies the shallow-ice-approximation. The bed DEM used in the model runs was constructed from radio echo data surveyed in spring 2014. We constrain the modern surface mass balance of Drangajökull using: 1) ablation and accumulation stakes; 2) ice surface digital elevation models (DEMs) from satellite, airborne LiDAR, and aerial photographs; and 3) full-stokes model-derived vertical ice velocities. The modeled vertical ice velocities and ice surface DEMs are combined to estimate past surface mass balance. We constrain Holocene glacier geometries using moraines and trimlines (e.g., Brynjolfsson, etal, 2014), proglacial-lake cores, and radiocarbon-dated dead vegetation emerging from under the modern glacier. We present a sensitivity analysis of the model to changes in parameters and show the effect of step changes of temperature and precipitation on glacier extent. Our results are placed in context with local lacustrine and marine climate proxies as well as with glacier extent and volume changes across the North Atlantic.
NASA Astrophysics Data System (ADS)
Nerney, E. G.; Bagenal, F.; Yoshioka, K.; Schmidt, C.
2017-12-01
Io emits volcanic gases into space at a rate of about a ton per second. The gases become ionized and trapped in Jupiter's strong magnetic field, forming a torus of plasma that emits 2 terawatts of UV emissions. In recent work re-analyzing UV emissions observed by Voyager, Galileo, & Cassini, we found plasma conditions consistent with a physical chemistry model with a neutral source of dissociated sulfur dioxide from Io (Nerney et al., 2017). In further analysis of UV observations from JAXA's Hisaki mission (using our spectral emission model) we constrain the torus composition with ground based observations. The physical chemistry model (adapted from Delamere et al., 2005) is then used to match derived plasma conditions. We correlate the oxygen to sulfur ratio of the neutral source with volcanic eruptions to understand the change in magnetospheric plasma conditions. Our goal is to better understand and constrain both the temporal and spatial variability of the flow of mass and energy from Io's volcanic atmosphere to Jupiter's dynamic magnetosphere.
Inverse and Predictive Modeling
DOE Office of Scientific and Technical Information (OSTI.GOV)
Syracuse, Ellen Marie
The LANL Seismo-Acoustic team has a strong capability in developing data-driven models that accurately predict a variety of observations. These models range from the simple – one-dimensional models that are constrained by a single dataset and can be used for quick and efficient predictions – to the complex – multidimensional models that are constrained by several types of data and result in more accurate predictions. Team members typically build models of geophysical characteristics of Earth and source distributions at scales of 1 to 1000s of km, the techniques used are applicable for other types of physical characteristics at an evenmore » greater range of scales. The following cases provide a snapshot of some of the modeling work done by the Seismo- Acoustic team at LANL.« less
NASA Astrophysics Data System (ADS)
Panda, Satyajit; Ray, M. C.
2008-04-01
In this paper, a geometrically nonlinear dynamic analysis has been presented for functionally graded (FG) plates integrated with a patch of active constrained layer damping (ACLD) treatment and subjected to a temperature field. The constraining layer of the ACLD treatment is considered to be made of the piezoelectric fiber-reinforced composite (PFRC) material. The temperature field is assumed to be spatially uniform over the substrate plate surfaces and varied through the thickness of the host FG plates. The temperature-dependent material properties of the FG substrate plates are assumed to be graded in the thickness direction of the plates according to a power-law distribution while the Poisson's ratio is assumed to be a constant over the domain of the plate. The constrained viscoelastic layer of the ACLD treatment is modeled using the Golla-Hughes-McTavish (GHM) method. Based on the first-order shear deformation theory, a three-dimensional finite element model has been developed to model the open-loop and closed-loop nonlinear dynamics of the overall FG substrate plates under the thermal environment. The analysis suggests the potential use of the ACLD treatment with its constraining layer made of the PFRC material for active control of geometrically nonlinear vibrations of FG plates in the absence or the presence of the temperature gradient across the thickness of the plates. It is found that the ACLD treatment is more effective in controlling the geometrically nonlinear vibrations of FG plates than in controlling their linear vibrations. The analysis also reveals that the ACLD patch is more effective for controlling the nonlinear vibrations of FG plates when it is attached to the softest surface of the FG plates than when it is bonded to the stiffest surface of the plates. The effect of piezoelectric fiber orientation in the active constraining PFRC layer on the damping characteristics of the overall FG plates is also discussed.
Characterizing biospheric carbon balance using CO2 observations from the OCO-2 satellite
NASA Astrophysics Data System (ADS)
Miller, Scot M.; Michalak, Anna M.; Yadav, Vineet; Tadić, Jovan M.
2018-05-01
NASA's Orbiting Carbon Observatory 2 (OCO-2) satellite launched in summer of 2014. Its observations could allow scientists to constrain CO2 fluxes across regions or continents that were previously difficult to monitor. This study explores an initial step toward that goal; we evaluate the extent to which current OCO-2 observations can detect patterns in biospheric CO2 fluxes and constrain monthly CO2 budgets. Our goal is to guide top-down, inverse modeling studies and identify areas for future improvement. We find that uncertainties and biases in the individual OCO-2 observations are comparable to the atmospheric signal from biospheric fluxes, particularly during Northern Hemisphere winter when biospheric fluxes are small. A series of top-down experiments indicate how these errors affect our ability to constrain monthly biospheric CO2 budgets. We are able to constrain budgets for between two and four global regions using OCO-2 observations, depending on the month, and we can constrain CO2 budgets at the regional level (i.e., smaller than seven global biomes) in only a handful of cases (16 % of all regions and months). The potential of the OCO-2 observations, however, is greater than these results might imply. A set of synthetic data experiments suggests that retrieval errors have a salient effect. Advances in retrieval algorithms and to a lesser extent atmospheric transport modeling will improve the results. In the interim, top-down studies that use current satellite observations are best-equipped to constrain the biospheric carbon balance across only continental or hemispheric regions.
Liu, Qingshan; Wang, Jun
2011-04-01
This paper presents a one-layer recurrent neural network for solving a class of constrained nonsmooth optimization problems with piecewise-linear objective functions. The proposed neural network is guaranteed to be globally convergent in finite time to the optimal solutions under a mild condition on a derived lower bound of a single gain parameter in the model. The number of neurons in the neural network is the same as the number of decision variables of the optimization problem. Compared with existing neural networks for optimization, the proposed neural network has a couple of salient features such as finite-time convergence and a low model complexity. Specific models for two important special cases, namely, linear programming and nonsmooth optimization, are also presented. In addition, applications to the shortest path problem and constrained least absolute deviation problem are discussed with simulation results to demonstrate the effectiveness and characteristics of the proposed neural network.
NASA Astrophysics Data System (ADS)
Ma, Xiaolin; Ma, Chi; Wan, Zhifang; Wang, Kewei
2017-06-01
Effective management of municipal solid waste (MSW) is critical for urban planning and development. This study aims to develop an integrated type 1 and type 2 fuzzy sets chance-constrained programming (ITFCCP) model for tackling regional MSW management problem under a fuzzy environment, where waste generation amounts are supposed to be type 2 fuzzy variables and treated capacities of facilities are assumed to be type 1 fuzzy variables. The evaluation and expression of uncertainty overcome the drawbacks in describing fuzzy possibility distributions as oversimplified forms. The fuzzy constraints are converted to their crisp equivalents through chance-constrained programming under the same or different confidence levels. Regional waste management of the City of Dalian, China, was used as a case study for demonstration. The solutions under various confidence levels reflect the trade-off between system economy and reliability. It is concluded that the ITFCCP model is capable of helping decision makers to generate reasonable waste-allocation alternatives under uncertainties.
Reinterpreting maximum entropy in ecology: a null hypothesis constrained by ecological mechanism.
O'Dwyer, James P; Rominger, Andrew; Xiao, Xiao
2017-07-01
Simplified mechanistic models in ecology have been criticised for the fact that a good fit to data does not imply the mechanism is true: pattern does not equal process. In parallel, the maximum entropy principle (MaxEnt) has been applied in ecology to make predictions constrained by just a handful of state variables, like total abundance or species richness. But an outstanding question remains: what principle tells us which state variables to constrain? Here we attempt to solve both problems simultaneously, by translating a given set of mechanisms into the state variables to be used in MaxEnt, and then using this MaxEnt theory as a null model against which to compare mechanistic predictions. In particular, we identify the sufficient statistics needed to parametrise a given mechanistic model from data and use them as MaxEnt constraints. Our approach isolates exactly what mechanism is telling us over and above the state variables alone. © 2017 John Wiley & Sons Ltd/CNRS.
Constraining Earth's Rheology of the Barents Sea Using Grace Gravity Change Observations
NASA Astrophysics Data System (ADS)
van der Wal, W.; Root, B. C.; Tarasov, L.
2014-12-01
The Barents Sea region was ice covered during last glacial maximum and experiences Glacial Isostatic Adjustment (GIA). Because of the limited amount of relevant geological and geodetic observations, it is difficult to constrain GIA models for this region. With improved ice sheet models and gravity observations from GRACE, it is possible to better constrain Earth rheology. This study aims to constrain the upper mantle viscosity and elastic lithosphere thickness from GRACE data in the Barents Sea region. The GRACE observations are corrected for current ice melting on Svalbard, Novaya Zemlya and Frans Joseph Land. A secular trend in gravity rate trend is estimated from the CSR release 5 GRACE data for the period of February 2003 to July 2013. Furthermore, long wavelength effects from distant large mass balance signals such as Greenland ice melting are filtered out. A new high-variance set of ice loading histories from calibrated glaciological modeling are used in the GIA modeling as it is found that ICE-5G over-estimates the observed GIA gravity change in the region. It is found that the rheology structure represented by VM5a results in over-estimation of the observed gravity change in the region for all ice sheet chronologies investigated. Therefore, other rheological Earth models were investigated. The best fitting upper mantle viscosity and elastic lithosphere thickness in the Barents Sea region are 4 (±0.5)*10^20 Pas and 110 (±20) km, respectively. The GRACE satellite mission proves to be a useful constraint in the Barents Sea Region for improving our knowledge on the upper mantle rheology.
NASA Astrophysics Data System (ADS)
Lundgren, P.; Nikkhoo, M.; Samsonov, S. V.; Milillo, P.; Gil-Cruz, F., Sr.; Lazo, J.
2017-12-01
Copahue volcano straddling the edge of the Agrio-Caviahue caldera along the Chile-Argentinaborder in the southern Andes has been in unrest since inflation began in late 2011. We constrain Copahue'ssource models with satellite and airborne interferometric synthetic aperture radar (InSAR) deformationobservations. InSAR time series from descending track RADARSAT-2 and COSMO-SkyMed data span theentire inflation period from 2011 to 2016, with their initially high rates of 12 and 15 cm/yr, respectively,slowing only slightly despite ongoing small eruptions through 2016. InSAR ascending and descending tracktime series for the 2013-2016 time period constrain a two-source compound dislocation model, with a rate ofvolume increase of 13 × 106 m3/yr. They consist of a shallow, near-vertical, elongated source centered at2.5 km beneath the summit and a deeper, shallowly plunging source centered at 7 km depth connecting theshallow source to the deeper caldera. The deeper source is located directly beneath the volcano tectonicseismicity with the lower bounds of the seismicity parallel to the plunge of the deep source. InSAR time seriesalso show normal fault offsets on the NE flank Copahue faults. Coulomb stress change calculations forright-lateral strike slip (RLSS), thrust, and normal receiver faults show positive values in the north caldera forboth RLSS and normal faults, suggesting that northward trending seismicity and Copahue fault motion withinthe caldera are caused by the modeled sources. Together, the InSAR-constrained source model and theseismicity suggest a deep conduit or transfer zone where magma moves from the central caldera toCopahue's upper edifice.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Tong; Xue, Li; Zhao, Xiao-Hong
Black holes (BHs) hide themselves behind various astronomical phenomena and their properties, i.e., mass and spin, are usually difficult to constrain. One leading candidate for the central engine model of gamma-ray bursts (GRBs) invokes a stellar mass BH and a neutrino-dominated accretion flow (NDAF), with the relativistic jet launched due to neutrino-anti-neutrino annihilations. Such a model gives rise to a matter-dominated fireball, and is suitable to interpret GRBs with a dominant thermal component with a photospheric origin. We propose a method to constrain BH mass and spin within the framework of this model and apply the method to the thermallymore » dominant GRB 101219B, whose initial jet launching radius, r {sub 0}, is constrained from the data. Using our numerical model of NDAF jets, we estimate the following constraints on the central BH: mass M {sub BH} ∼ 5–9 M {sub ⊙}, spin parameter a {sub *} ≳ 0.6, and disk mass 3 M {sub ⊙} ≲ M {sub disk} ≲ 4 M {sub ⊙}. Our results also suggest that the NDAF model is a competitive candidate for the central engine of GRBs with a strong thermal component.« less
Magnetic avalanches in manganese-acetate, "magnetic deflagration"
NASA Astrophysics Data System (ADS)
Suzuki, Yoko
Mn12-acetate, first synthesized in 1980 by Lis, is one example of a class of many molecules called single molecule magnets (SMMs) or molecular nanomagnets. These molecules have several atomic spins strongly coupled together within each molecule. They exhibit interesting quantum mechanical phenomena at low temperatures such as quantum tunneling of magnetization, which was first found with Mn12-acetate in 1996 by Friedman, et al. , and Berry phase oscillations which were measured in Fe8 (another SMM) in 1999 by Wernsdorfer, et al. In addition to possible application as memory storage and qubits for quantum computers, these systems provide the means for studies of mesoscopic physics as well as the interactions of the molecules with their environment, such as phonon, photon, nuclear spin, intermolecular dipole, and exchange interactions. Mn12-acetate has twelve Mn ions magnetically coupled in the center of the molecule yielding a giant spin of S = 10 at low temperature. It also has a large uniaxial anisotropy of 65 K. Below 3 K, magnetization curves show strong hysteresis due to the anisotropy barrier. At thesis temperatures, the spin relaxes through the barrier by quantum tunneling of magnetization, which produces regularly-spaced multiple resonant steps in the hysteresis curve. Magnetic avalanches, first detected by Paulsen et al., also occur for some samples only at low temperature, leading to a very fast single-step reversal of the full magnetization, which clearly differs from relaxation by tunneling. In this thesis, I present the results of detailed experimental studies of two aspects of magnetic avalanche phenomenon: "conditions for the triggering of avalanches" and "propagation of the avalanche front". In the first study, we find the magnetic fields at which avalanches occur are stochastically distributed in a particular range of fields. For the second study, we conducted local time-resolved measurements. The results indicate the magnetization avalanches spread as a narrow interface that propagate through the crystal at a constant velocity which is roughly two orders of magnitude smaller than the speed of sound in solids. We argue this phenomenon is closely analogous to the propagation of a flame front(deflagration) through a flammable chemical substance.
NASA Technical Reports Server (NTRS)
Nicely, Julie M.; Anderson, Daniel C.; Canty, Timothy P.; Salawitch, Ross J.; Wolfe, Glenn M.; Apel, Eric C.; Arnold, Steve R.; Atlas, Elliot L.; Blake, Nicola J.; Bresch, James F.;
2016-01-01
Hydroxyl radical (OH) is the main daytime oxidant in the troposphere and determines the atmospheric lifetimes of many compounds. We use aircraft measurements of O3, H2O, NO, and other species from the Convective Transport of Active Species in the Tropics (CONTRAST) field campaign, which occurred in the tropical western Pacific (TWP) during January-February 2014, to constrain a photochemical box model and estimate concentrations of OH throughout the troposphere. We find that tropospheric column OH (OHCOL) inferred from CONTRAST observations is 12 to 40% higher than found in chemical transport models (CTMs), including CAM-chem-SD run with 2014 meteorology as well as eight models that participated in POLMIP (2008 meteorology). Part of this discrepancy is due to a clear-sky sampling bias that affects CONTRAST observations; accounting for this bias and also for a small difference in chemical mechanism results in our empirically based value of OHCOL being 0 to 20% larger than found within global models. While these global models simulate observed O3 reasonably well, they underestimate NOx (NO +NO2) by a factor of 2, resulting in OHCOL approx.30% lower than box model simulations constrained by observed NO. Underestimations by CTMs of observed CH3CHO throughout the troposphere and of HCHO in the upper troposphere further contribute to differences between our constrained estimates of OH and those calculated by CTMs. Finally, our calculations do not support the prior suggestion of the existence of a tropospheric OH minimum in the TWP, because during January-February 2014 observed levels of O3 and NO were considerably larger than previously reported values in the TWP.
NASA Astrophysics Data System (ADS)
Bapst, J.; Byrne, S.
2016-12-01
The stability of water ice on Mars' surface is determined by its temperature and the density of water vapor at the bottom of the atmosphere. Multiple orbiting instruments have been used to study column-integrated water abundance in the martian atmosphere, resolving the global annual water cycle. However, poor knowledge of the vertical distribution of water makes constraining its abundance near the surface difficult. One must assume a mixing regime to produce surface vapor density estimates. More indirectly, one can use the appearance and disappearance of seasonal water frost, along with ice stability models, to estimate this value. Here, we use derived temperature and surface reflectance data from MGS TES to constrain a 1-D thermal diffusion model, which is coupled to an atmospheric water transport model. TES temperatures are used to constrain thermal properties of our modeled subsurface, while changes in TES albedo can be used to determine the timing of water frost. We tune the density of water vapor in the atmospheric model to match the observed seasonal water frost timing in the northern hemisphere, poleward of 45°N. Thus, we produce a new estimate for the water abundance in the lower atmosphere of Mars and how it varies seasonally and geographically. The timing of water frost can be ambiguous in TES data, especially at lower latitudes where the albedo contrast between frosted and unfrosted surfaces is lower (presumably due to lesser areal coverage of water frost). The uncertainty in frost timing with our approach is <20° LS ( 40 sols), and will be used to define upper and lower bounds in our estimate of vapor density. The implications of our derived vapor densities on the stability of surface and subsurface water ice will be discussed.
NASA Astrophysics Data System (ADS)
Ritzinger, B. T.; Glen, J. M. G.; Athens, N. D.; Denton, K. M.; Bouligand, C.
2015-12-01
Regionally continuous Cenozoic rocks in the Basin and Range that predate the onset of major mid-Miocene extension provide valuable insight into the sequence of faulting and magnitude of extension. An exceptional example of this is Caetano caldera, located in north-central Nevada, that formed during the eruption of the Caetano Tuff at the Eocene-Oligocene transition. The caldera and associated deposits, as well as conformable caldera-filling sedimentary and volcanic units allow for the reconstruction of post Oligocene extensional faulting. Extensive mapping and geochronologic, geochemical and paleomagnetic analyses have been conducted over the last decade to help further constrain the eruptive and extensional history of the Caetano caldera and associated deposits. Gravity and magnetic data, that highlight contrasts in density and magnetic properties (susceptibility and remanence), respectively, are useful for mapping and modeling structural and lithic discontinuities. By combining existing gravity and aeromagnetic data with newly collected high-resolution gravity data, we are performing detailed potential field modeling to better characterize the subsurface within and surrounding the caldera. Modeling is constrained by published geologic map and cross sections and by new rock properties for these units determined from oriented drill core and hand samples collected from outcrops that span all of the major rock units in the study area. These models will enable us to better map the margins of the caldera and more accurately determine subsurface lithic boundaries and complex fault geometries, as well as aid in refining estimates of the magnitude of extension across the caldera. This work highlights the value in combining geologic and geophysical data to build an integrated structural model to help characterize the subsurface and better constrain the extensional tectonic history if this part of the Great Basin.
A finite-temperature Hartree-Fock code for shell-model Hamiltonians
NASA Astrophysics Data System (ADS)
Bertsch, G. F.; Mehlhaff, J. M.
2016-10-01
The codes HFgradZ.py and HFgradT.py find axially symmetric minima of a Hartree-Fock energy functional for a Hamiltonian supplied in a shell model basis. The functional to be minimized is the Hartree-Fock energy for zero-temperature properties or the Hartree-Fock grand potential for finite-temperature properties (thermal energy, entropy). The minimization may be subjected to additional constraints besides axial symmetry and nucleon numbers. A single-particle operator can be used to constrain the minimization by adding it to the single-particle Hamiltonian with a Lagrange multiplier. One can also constrain its expectation value in the zero-temperature code. Also the orbital filling can be constrained in the zero-temperature code, fixing the number of nucleons having given Kπ quantum numbers. This is particularly useful to resolve near-degeneracies among distinct minima.
Constrained inversion as a hypothesis testing tool, what can we learn about the lithosphere?
NASA Astrophysics Data System (ADS)
Moorkamp, Max; Stewart, Fishwick; Jones, Alan G.
2017-04-01
Inversion of geophysical data constrained by a reference model is typically used to guide the inversion of low resolution data towards a geologically plausible solution. For example, a migrated seismic section can provide the location of lithological boundaries for potential field inversions. Here we consider the inversion of long-period magnetotelluric data constrained by models generated through surface wave inversion. In this case, we do not consider the surface wave model inherently better in any sense and want to guide the magnetotelluric inversion towards this model, but we want to test the hypothesis that both datasets can be explained by models with similar structure. If the hypothesis test is successful, i.e. we can fit the observations with a conductivity model with structural similarity to the seismic model, we have found an alternative explanation compared to the individual inversion and can use the differences to learn about the resolution of the magnetotelluric data and can improve our interpretation. Conversely, if the test refutes our hypothesis of coincident structure, we have found features in the models that are sensed fundamentally different by both methods which is potentially instructive on the nature of the anomalies. We use a MT dataset acquired in central Botswana over the Okwa terrane and the adjacent Kaapvaal and Zimbabwe Cratons together with a tomographic model for the region to illustrate and test this approach. Here, various conductive structures have been identified that bridge the Moho. Furthermore, the thickness of the lithosphere inferred from the different methods differs. In both cases the question is in how far this is a result of the ill-posed nature of inversion and in how far these differences can be reconciled. Thus this dataset is an ideal test case for our hypothesis testing approach. Finally, we will demonstrate how we can use the results of the constrained inversion to extract conductivity-velocity relationships in the region and gain further insight into the composition and thermal structure of the lithosphere.
NASA Astrophysics Data System (ADS)
Li, Guang
2017-01-01
This paper presents a fast constrained optimization approach, which is tailored for nonlinear model predictive control of wave energy converters (WEC). The advantage of this approach relies on its exploitation of the differential flatness of the WEC model. This can reduce the dimension of the resulting nonlinear programming problem (NLP) derived from the continuous constrained optimal control of WEC using pseudospectral method. The alleviation of computational burden using this approach helps to promote an economic implementation of nonlinear model predictive control strategy for WEC control problems. The method is applicable to nonlinear WEC models, nonconvex objective functions and nonlinear constraints, which are commonly encountered in WEC control problems. Numerical simulations demonstrate the efficacy of this approach.
The Apollo 16 regolith - A petrographically-constrained chemical mixing model
NASA Technical Reports Server (NTRS)
Kempa, M. J.; Papike, J. J.; White, C.
1980-01-01
A mixing model for Apollo 16 regolith samples has been developed, which differs from other A-16 mixing models in that it is both petrographically constrained and statistically sound. The model was developed using three components representative of rock types present at the A-16 site, plus a representative mare basalt. A linear least-squares fitting program employing the chi-squared test and sum of components was used to determine goodness of fit. Results for surface soils indicate that either there are no significant differences between Cayley and Descartes material at the A-16 site or, if differences do exist, they have been obscured by meteoritic reworking and mixing of the lithologies.
Hierarchical Bayesian Model Averaging for Chance Constrained Remediation Designs
NASA Astrophysics Data System (ADS)
Chitsazan, N.; Tsai, F. T.
2012-12-01
Groundwater remediation designs are heavily relying on simulation models which are subjected to various sources of uncertainty in their predictions. To develop a robust remediation design, it is crucial to understand the effect of uncertainty sources. In this research, we introduce a hierarchical Bayesian model averaging (HBMA) framework to segregate and prioritize sources of uncertainty in a multi-layer frame, where each layer targets a source of uncertainty. The HBMA framework provides an insight to uncertainty priorities and propagation. In addition, HBMA allows evaluating model weights in different hierarchy levels and assessing the relative importance of models in each level. To account for uncertainty, we employ a chance constrained (CC) programming for stochastic remediation design. Chance constrained programming was implemented traditionally to account for parameter uncertainty. Recently, many studies suggested that model structure uncertainty is not negligible compared to parameter uncertainty. Using chance constrained programming along with HBMA can provide a rigorous tool for groundwater remediation designs under uncertainty. In this research, the HBMA-CC was applied to a remediation design in a synthetic aquifer. The design was to develop a scavenger well approach to mitigate saltwater intrusion toward production wells. HBMA was employed to assess uncertainties from model structure, parameter estimation and kriging interpolation. An improved harmony search optimization method was used to find the optimal location of the scavenger well. We evaluated prediction variances of chloride concentration at the production wells through the HBMA framework. The results showed that choosing the single best model may lead to a significant error in evaluating prediction variances for two reasons. First, considering the single best model, variances that stem from uncertainty in the model structure will be ignored. Second, considering the best model with non-dominant model weight may underestimate or overestimate prediction variances by ignoring other plausible propositions. Chance constraints allow developing a remediation design with a desirable reliability. However, considering the single best model, the calculated reliability will be different from the desirable reliability. We calculated the reliability of the design for the models at different levels of HBMA. The results showed that by moving toward the top layers of HBMA, the calculated reliability converges to the chosen reliability. We employed the chance constrained optimization along with the HBMA framework to find the optimal location and pumpage for the scavenger well. The results showed that using models at different levels in the HBMA framework, the optimal location of the scavenger well remained the same, but the optimal extraction rate was altered. Thus, we concluded that the optimal pumping rate was sensitive to the prediction variance. Also, the prediction variance was changed by using different extraction rate. Using very high extraction rate will cause prediction variances of chloride concentration at the production wells to approach zero regardless of which HBMA models used.
Testing a Constrained MPC Controller in a Process Control Laboratory
ERIC Educational Resources Information Center
Ricardez-Sandoval, Luis A.; Blankespoor, Wesley; Budman, Hector M.
2010-01-01
This paper describes an experiment performed by the fourth year chemical engineering students in the process control laboratory at the University of Waterloo. The objective of this experiment is to test the capabilities of a constrained Model Predictive Controller (MPC) to control the operation of a Double Pipe Heat Exchanger (DPHE) in real time.…
Modelling compressible dense and dilute two-phase flows
NASA Astrophysics Data System (ADS)
Saurel, Richard; Chinnayya, Ashwin; Carmouze, Quentin
2017-06-01
Many two-phase flow situations, from engineering science to astrophysics, deal with transition from dense (high concentration of the condensed phase) to dilute concentration (low concentration of the same phase), covering the entire range of volume fractions. Some models are now well accepted at the two limits, but none are able to cover accurately the entire range, in particular regarding waves propagation. In the present work, an alternative to the Baer and Nunziato (BN) model [Baer, M. R. and Nunziato, J. W., "A two-phase mixture theory for the deflagration-to-detonation transition (DDT) in reactive granular materials," Int. J. Multiphase Flow 12(6), 861 (1986)], initially designed for dense flows, is built. The corresponding model is hyperbolic and thermodynamically consistent. Contrarily to the BN model that involves 6 wave speeds, the new formulation involves 4 waves only, in agreement with the Marble model [Marble, F. E., "Dynamics of a gas containing small solid particles," Combustion and Propulsion (5th AGARD Colloquium) (Pergamon Press, 1963), Vol. 175] based on pressureless Euler equations for the dispersed phase, a well-accepted model for low particle volume concentrations. In the new model, the presence of pressure in the momentum equation of the particles and consideration of volume fractions in the two phases render the model valid for large particle concentrations. A symmetric version of the new model is derived as well for liquids containing gas bubbles. This model version involves 4 characteristic wave speeds as well, but with different velocities. Last, the two sub-models with 4 waves are combined in a unique formulation, valid for the full range of volume fractions. It involves the same 6 wave speeds as the BN model, but at a given point of space, 4 waves only emerge, depending on the local volume fractions. The non-linear pressure waves propagate only in the phase with dominant volume fraction. The new model is tested numerically on various test problems ranging from separated phases in a shock tube to shock-particle cloud interaction. Its predictions are compared to BN and Marble models as well as against experimental data showing clear improvements.
Robust model predictive control for constrained continuous-time nonlinear systems
NASA Astrophysics Data System (ADS)
Sun, Tairen; Pan, Yongping; Zhang, Jun; Yu, Haoyong
2018-02-01
In this paper, a robust model predictive control (MPC) is designed for a class of constrained continuous-time nonlinear systems with bounded additive disturbances. The robust MPC consists of a nonlinear feedback control and a continuous-time model-based dual-mode MPC. The nonlinear feedback control guarantees the actual trajectory being contained in a tube centred at the nominal trajectory. The dual-mode MPC is designed to ensure asymptotic convergence of the nominal trajectory to zero. This paper extends current results on discrete-time model-based tube MPC and linear system model-based tube MPC to continuous-time nonlinear model-based tube MPC. The feasibility and robustness of the proposed robust MPC have been demonstrated by theoretical analysis and applications to a cart-damper springer system and a one-link robot manipulator.
Measuring Quasar Spin via X-ray Continuum Fitting
NASA Astrophysics Data System (ADS)
Jenkins, Matthew; Pooley, David; Rappaport, Saul; Steiner, Jack
2018-01-01
We have identified several quasars whose X-ray spectra appear very soft. When fit with power-law models, the best-fit indices are greater than 3. This is very suggestive of thermal disk emission, indicating that the X-ray spectrum is dominated by the disk component. Galactic black hole binaries in such states have been successfully fit with disk-blackbody models to constrain the inner radius, which also constrains the spin of the black hole. We have fit those models to XMM-Newton spectra of several of our identified soft X-ray quasars to place constraints on the spins of the supermassive black holes.
Application of a sparseness constraint in multivariate curve resolution - Alternating least squares.
Hugelier, Siewert; Piqueras, Sara; Bedia, Carmen; de Juan, Anna; Ruckebusch, Cyril
2018-02-13
The use of sparseness in chemometrics is a concept that has increased in popularity. The advantage is, above all, a better interpretability of the results obtained. In this work, sparseness is implemented as a constraint in multivariate curve resolution - alternating least squares (MCR-ALS), which aims at reproducing raw (mixed) data by a bilinear model of chemically meaningful profiles. In many cases, the mixed raw data analyzed are not sparse by nature, but their decomposition profiles can be, as it is the case in some instrumental responses, such as mass spectra, or in concentration profiles linked to scattered distribution maps of powdered samples in hyperspectral images. To induce sparseness in the constrained profiles, one-dimensional and/or two-dimensional numerical arrays can be fitted using a basis of Gaussian functions with a penalty on the coefficients. In this work, a least squares regression framework with L 0 -norm penalty is applied. This L 0 -norm penalty constrains the number of non-null coefficients in the fit of the array constrained without having an a priori on the number and their positions. It has been shown that the sparseness constraint induces the suppression of values linked to uninformative channels and noise in MS spectra and improves the location of scattered compounds in distribution maps, resulting in a better interpretability of the constrained profiles. An additional benefit of the sparseness constraint is a lower ambiguity in the bilinear model, since the major presence of null coefficients in the constrained profiles also helps to limit the solutions for the profiles in the counterpart matrix of the MCR bilinear model. Copyright © 2017 Elsevier B.V. All rights reserved.
An adaptive finite element method for the inequality-constrained Reynolds equation
NASA Astrophysics Data System (ADS)
Gustafsson, Tom; Rajagopal, Kumbakonam R.; Stenberg, Rolf; Videman, Juha
2018-07-01
We present a stabilized finite element method for the numerical solution of cavitation in lubrication, modeled as an inequality-constrained Reynolds equation. The cavitation model is written as a variable coefficient saddle-point problem and approximated by a residual-based stabilized method. Based on our recent results on the classical obstacle problem, we present optimal a priori estimates and derive novel a posteriori error estimators. The method is implemented as a Nitsche-type finite element technique and shown in numerical computations to be superior to the usually applied penalty methods.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jennings, Elise; Wechsler, Risa H.
We present the nonlinear 2D galaxy power spectrum, P(k, µ), in redshift space, measured from the Dark Sky simulations, using galaxy catalogs constructed with both halo occupation distribution and subhalo abundance matching methods, chosen to represent an intermediate redshift sample of luminous red galaxies. We find that the information content in individual µ (cosine of the angle to the line of sight) bins is substantially richer then multipole moments, and show that this can be used to isolate the impact of nonlinear growth and redshift space distortion (RSD) effects. Using the µ < 0.2 simulation data, which we show ismore » not impacted by RSD effects, we can successfully measure the nonlinear bias to an accuracy of ~ 5% at k < 0.6hMpc-1 . This use of individual µ bins to extract the nonlinear bias successfully removes a large parameter degeneracy when constraining the linear growth rate of structure. We carry out a joint parameter estimation, using the low µ simulation data to constrain the nonlinear bias, and µ > 0.2 to constrain the growth rate and show that f can be constrained to ~ 26(22)% to a kmax < 0.4(0.6)hMpc-1 from clustering alone using a simple dispersion model, for a range of galaxy models. Our analysis of individual µ bins also reveals interesting physical effects which arise simply from different methods of populating halos with galaxies. We also find a prominent turnaround scale, at which RSD damping effects are greater then the nonlinear growth, which differs not only for each µ bin but also for each galaxy model. These features may provide unique signatures which could be used to shed light on the galaxy–dark matter connection. Furthermore, the idea of separating nonlinear growth and RSD effects making use of the full information in the 2D galaxy power spectrum yields significant improvements in constraining cosmological parameters and may be a promising probe of galaxy formation models.« less
Disentangling Redshift-Space Distortions and Nonlinear Bias using the 2D Power Spectrum
Jennings, Elise; Wechsler, Risa H.
2015-08-07
We present the nonlinear 2D galaxy power spectrum, P(k, µ), in redshift space, measured from the Dark Sky simulations, using galaxy catalogs constructed with both halo occupation distribution and subhalo abundance matching methods, chosen to represent an intermediate redshift sample of luminous red galaxies. We find that the information content in individual µ (cosine of the angle to the line of sight) bins is substantially richer then multipole moments, and show that this can be used to isolate the impact of nonlinear growth and redshift space distortion (RSD) effects. Using the µ < 0.2 simulation data, which we show ismore » not impacted by RSD effects, we can successfully measure the nonlinear bias to an accuracy of ~ 5% at k < 0.6hMpc-1 . This use of individual µ bins to extract the nonlinear bias successfully removes a large parameter degeneracy when constraining the linear growth rate of structure. We carry out a joint parameter estimation, using the low µ simulation data to constrain the nonlinear bias, and µ > 0.2 to constrain the growth rate and show that f can be constrained to ~ 26(22)% to a kmax < 0.4(0.6)hMpc-1 from clustering alone using a simple dispersion model, for a range of galaxy models. Our analysis of individual µ bins also reveals interesting physical effects which arise simply from different methods of populating halos with galaxies. We also find a prominent turnaround scale, at which RSD damping effects are greater then the nonlinear growth, which differs not only for each µ bin but also for each galaxy model. These features may provide unique signatures which could be used to shed light on the galaxy–dark matter connection. Furthermore, the idea of separating nonlinear growth and RSD effects making use of the full information in the 2D galaxy power spectrum yields significant improvements in constraining cosmological parameters and may be a promising probe of galaxy formation models.« less
Axisymmetric Numerical Modeling of Pulse Detonation Rocket Engines
NASA Technical Reports Server (NTRS)
Morris, Christopher I.
2005-01-01
Pulse detonation rocket engines (PDREs) have generated research interest in recent years as a chemical propulsion system potentially offering improved performance and reduced complexity compared to conventional rocket engines. The detonative mode of combustion employed by these devices offers a thermodynamic advantage over the constant-pressure deflagrative combustion mode used in conventional rocket engines and gas turbines. However, while this theoretical advantage has spurred considerable interest in building PDRE devices, the unsteady blowdown process intrinsic to the PDRE has made realistic estimates of the actual propulsive performance problematic. The recent review article by Kailasanath highlights some of the progress that has been made in comparing the available experimental measurements with analytical and numerical models. In recent work by the author, a quasi-one-dimensional, finite rate chemistry CFD model was utilized to study the gasdynamics and performance characteristics of PDREs over a range of blowdown pressure ratios from 1-1000. Models of this type are computationally inexpensive, and enable first-order parametric studies of the effect of several nozzle and extension geometries on PDRE performance over a wide range of conditions. However, the quasi-one-dimensional approach is limited in that it cannot properly capture the multidimensional blast wave and flow expansion downstream of the PDRE, nor can it resolve nozzle flow separation if present. Moreover, the previous work was limited to single-pulse calculations. In this paper, an axisymmetric finite rate chemistry model is described and utilized to study these issues in greater detail. Example Mach number contour plots showing the multidimensional blast wave and nozzle exhaust plume are shown. The performance results are compared with the quasi-one-dimensional results from the previous paper. Both Euler and Navier-Stokes solutions are calculated in order to determine the effect of viscous effects in the nozzle flowfield. Additionally, comparisons of the model results to performance data from CalTech, as well as experimental flowfield measurements from Stanford University, are also reported.
Spectroscopic Study of a Pulsed High-Energy Plasma Deflagration Accelerator
NASA Astrophysics Data System (ADS)
Loebner, Keith; Underwood, Thomas; Mouratidis, Theodore; Cappelli, Mark
2015-11-01
Observations of broadened Balmer lines emitted by a highly-ionized transient plasma jet are presented. A gated CCD camera coupled to a high-resolution spectrometer is used to obtain chord-averaged broadening data for a complete cross section of the plasma jet, and the data is Abel inverted to derive the radial plasma density distribution. This measurement is performed over narrow gate widths and at multiple axial positions to provide high spatial and temporal resolution. A streak camera coupled to a spectrometer is used to obtain continuous-time broadening data over the entire duration of the discharge event (10-50 microseconds). Analyses of discharge characteristics and comparisons with previous work are discussed. This work is supported by the U.S. Department of Energy Stewardship Science Academic Program, as well as the National Defense Science Engineering Graduate Fellowship.
The Internal Ballistics of an Air Gun
NASA Astrophysics Data System (ADS)
Denny, Mark
2011-02-01
The internal ballistics of a firearm or artillery piece considers the pellet, bullet, or shell motion while it is still inside the barrel. In general, deriving the muzzle speed of a gunpowder firearm from first principles is difficult because powder combustion is fast and it very rapidly raises the temperature of gas (generated by gunpowder deflagration, or burning), which greatly complicates the analysis. A simple case is provided by air guns, for which we can make reasonable approximations that permit a derivation of muzzle speed. It is perhaps surprising that muzzle speed depends upon barrel length (artillerymen debated this dependence for centuries, until it was established experimentally and, later, theoretically ). Here we see that a simple physical analysis, accessible to high school or freshmen undergraduate physics students, not only derives realistic muzzle speed but also shows how it depends upon barrel length.
Constraining the braneworld with gravitational wave observations.
McWilliams, Sean T
2010-04-09
Some braneworld models may have observable consequences that, if detected, would validate a requisite element of string theory. In the infinite Randall-Sundrum model (RS2), the AdS radius of curvature, l, of the extra dimension supports a single bound state of the massless graviton on the brane, thereby reproducing Newtonian gravity in the weak-field limit. However, using the AdS/CFT correspondence, it has been suggested that one possible consequence of RS2 is an enormous increase in Hawking radiation emitted by black holes. We utilize this possibility to derive two novel methods for constraining l via gravitational wave measurements. We show that the EMRI event rate detected by LISA can constrain l at the approximately 1 microm level for optimal cases, while the observation of a single galactic black hole binary with LISA results in an optimal constraint of l < or = 5 microm.
Constraining the Braneworld with Gravitational Wave Observations
NASA Technical Reports Server (NTRS)
McWilliams, Sean T.
2011-01-01
Some braneworld models may have observable consequences that, if detected, would validate a requisite element of string theory. In the infinite Randall-Sundrum model (RS2), the AdS radius of curvature, L, of the extra dimension supports a single bound state of the massless graviton on the brane, thereby reproducing Newtonian gravity in the weak-field limit. However, using the AdS/CFT correspondence, it has been suggested that one possible consequence of RS2 is an enormous increase in Hawking radiation emitted by black holes. We utilize this possibility to derive two novel methods for constraining L via gravitational wave measurements. We show that the EMRI event rate detected by LISA can constrain L at the approximately 1 micron level for optimal cases, while the observation of a single galactic black hole binary with LISA results in an optimal constraint of L less than or equal to 5 microns.
Analytical Dynamics and Nonrigid Spacecraft Simulation
NASA Technical Reports Server (NTRS)
Likins, P. W.
1974-01-01
Application to the simulation of idealized spacecraft are considered both for multiple-rigid-body models and for models consisting of combination of rigid bodies and elastic bodies, with the elastic bodies being defined either as continua, as finite-element systems, or as a collection of given modal data. Several specific examples are developed in detail by alternative methods of analytical mechanics, and results are compared to a Newton-Euler formulation. The following methods are developed from d'Alembert's principle in vector form: (1) Lagrange's form of d'Alembert's principle for independent generalized coordinates; (2) Lagrange's form of d'Alembert's principle for simply constrained systems; (3) Kane's quasi-coordinate formulation of D'Alembert's principle; (4) Lagrange's equations for independent generalized coordinates; (5) Lagrange's equations for simply constrained systems; (6) Lagrangian quasi-coordinate equations (or the Boltzmann-Hamel equations); (7) Hamilton's equations for simply constrained systems; and (8) Hamilton's equations for independent generalized coordinates.
CONORBIT: constrained optimization by radial basis function interpolation in trust regions
Regis, Rommel G.; Wild, Stefan M.
2016-09-26
Here, this paper presents CONORBIT (CONstrained Optimization by Radial Basis function Interpolation in Trust regions), a derivative-free algorithm for constrained black-box optimization where the objective and constraint functions are computationally expensive. CONORBIT employs a trust-region framework that uses interpolating radial basis function (RBF) models for the objective and constraint functions, and is an extension of the ORBIT algorithm. It uses a small margin for the RBF constraint models to facilitate the generation of feasible iterates, and extensive numerical tests confirm that such a margin is helpful in improving performance. CONORBIT is compared with other algorithms on 27 test problems, amore » chemical process optimization problem, and an automotive application. Numerical results show that CONORBIT performs better than COBYLA, a sequential penalty derivative-free method, an augmented Lagrangian method, a direct search method, and another RBF-based algorithm on the test problems and on the automotive application.« less
Structural and parameteric uncertainty quantification in cloud microphysics parameterization schemes
NASA Astrophysics Data System (ADS)
van Lier-Walqui, M.; Morrison, H.; Kumjian, M. R.; Prat, O. P.; Martinkus, C.
2017-12-01
Atmospheric model parameterization schemes employ approximations to represent the effects of unresolved processes. These approximations are a source of error in forecasts, caused in part by considerable uncertainty about the optimal value of parameters within each scheme -- parameteric uncertainty. Furthermore, there is uncertainty regarding the best choice of the overarching structure of the parameterization scheme -- structrual uncertainty. Parameter estimation can constrain the first, but may struggle with the second because structural choices are typically discrete. We address this problem in the context of cloud microphysics parameterization schemes by creating a flexible framework wherein structural and parametric uncertainties can be simultaneously constrained. Our scheme makes no assuptions about drop size distribution shape or the functional form of parametrized process rate terms. Instead, these uncertainties are constrained by observations using a Markov Chain Monte Carlo sampler within a Bayesian inference framework. Our scheme, the Bayesian Observationally-constrained Statistical-physical Scheme (BOSS), has flexibility to predict various sets of prognostic drop size distribution moments as well as varying complexity of process rate formulations. We compare idealized probabilistic forecasts from versions of BOSS with varying levels of structural complexity. This work has applications in ensemble forecasts with model physics uncertainty, data assimilation, and cloud microphysics process studies.
The highest-frequency kHz QPOs in neutron star low mass X-ray binaries
NASA Astrophysics Data System (ADS)
van Doesburgh, Marieke; van der Klis, Michiel; Morsink, Sharon M.
2018-05-01
We investigate the detections with RXTE of the highest-frequency kHz QPOs previously reported in six neutron star (NS) low mass X-ray binaries. We find that the highest-frequency kHz QPO detected in 4U 0614+09 has a 1267 Hz 3σ confidence lower limit on its centroid frequency. This is the highest such limit reported to date, and of direct physical interest as it can be used to constrain QPO models and the supranuclear density equation of state (EoS). We compare our measured frequencies to maximum orbital frequencies predicted in full GR using models of rotating neutron stars with a number of different modern EoS and show that these can accommodate the observed QPO frequencies. Orbital motion constrained by NS and ISCO radii is therefore a viable explanation of these QPOs. In the most constraining case of 4U 0614+09 we find the NS mass must be M<2.1 M⊙. From our measured QPO frequencies we can constrain the NS radii for five of the six sources we studied to narrow ranges (±0.1-0.7 km) different for each source and each EoS.
Liu, Aiqin; Jennings, Louise M; Ingham, Eileen; Fisher, John
2015-09-18
The successful development of early-stage cartilage and meniscus repair interventions in the knee requires biomechanical and biotribological understanding of the design of the therapeutic interventions and their tribological function in the natural joint. The aim of this study was to develop and validate a porcine knee model using a whole joint knee simulator for investigation of the tribological function and biomechanical properties of the natural knee, which could then be used to pre-clinically assess the tribological performance of cartilage and meniscal repair interventions prior to in vivo studies. The tribological performance of standard artificial bearings in terms of anterior-posterior (A/P) shear force was determined in a newly developed six degrees of freedom tribological joint simulator. The porcine knee model was then developed and the tribological properties in terms of shear force measurements were determined for the first time for three levels of biomechanical constraints including A/P constrained, spring force semi-constrained and A/P unconstrained conditions. The shear force measurements showed higher values under the A/P constrained condition (predominantly sliding motion) compared to the A/P unconstrained condition (predominantly rolling motion). This indicated that the shear force simulation model was able to differentiate between tribological behaviours when the femoral and tibial bearing was constrained to slide or/and roll. Therefore, this porcine knee model showed the potential capability to investigate the effect of knee structural, biomechanical and kinematic changes, as well as different cartilage substitution therapies on the tribological function of natural knee joints. Copyright © 2015 The Authors. Published by Elsevier Ltd.. All rights reserved.
NASA Astrophysics Data System (ADS)
Souri, Amir H.; Choi, Yunsoo; Pan, Shuai; Curci, Gabriele; Nowlan, Caroline R.; Janz, Scott J.; Kowalewski, Matthew G.; Liu, Junjie; Herman, Jay R.; Weinheimer, Andrew J.
2018-03-01
A number of satellite-based instruments have become an essential part of monitoring emissions. Despite sound theoretical inversion techniques, the insufficient samples and the footprint size of current observations have introduced an obstacle to narrow the inversion window for regional models. These key limitations can be partially resolved by a set of modest high-quality measurements from airborne remote sensing. This study illustrates the feasibility of nitrogen dioxide (NO2) columns from the Geostationary Coastal and Air Pollution Events Airborne Simulator (GCAS) to constrain anthropogenic NOx emissions in the Houston-Galveston-Brazoria area. We convert slant column densities to vertical columns using a radiative transfer model with (i) NO2 profiles from a high-resolution regional model (1 × 1 km2) constrained by P-3B aircraft measurements, (ii) the consideration of aerosol optical thickness impacts on radiance at NO2 absorption line, and (iii) high-resolution surface albedo constrained by ground-based spectrometers. We characterize errors in the GCAS NO2 columns by comparing them to Pandora measurements and find a striking correlation (r > 0.74) with an uncertainty of 3.5 × 1015 molecules cm-2. On 9 of 10 total days, the constrained anthropogenic emissions by a Kalman filter yield an overall 2-50% reduction in polluted areas, partly counterbalancing the well-documented positive bias of the model. The inversion, however, boosts emissions by 94% in the same areas on a day when an unprecedented local emissions event potentially occurred, significantly mitigating the bias of the model. The capability of GCAS at detecting such an event ensures the significance of forthcoming geostationary satellites for timely estimates of top-down emissions.
Application of High Resolution Air-Borne Remote Sensing Observations for Monitoring NOx Emissions
NASA Astrophysics Data System (ADS)
Souri, A.; Choi, Y.; Pan, S.; Curci, G.; Janz, S. J.; Kowalewski, M. G.; Liu, J.; Herman, J. R.; Weinheimer, A. J.
2017-12-01
Nitrogen oxides (NOx=NO+NO2) are one of the air pollutants, responsible for the formation of tropospheric ozone, acid rain and particulate nitrate. The anthropogenic NOx emissions are commonly estimated based on bottom-up inventories which are complicated by many potential sources of error. One way to improve the emission inventories is to use relevant observations to constrain them. Fortunately, Nitrogen dioxide (NO2) is one of the most successful detected species from remote sensing. Although many studies have shown the capability of using space-borne remote sensing observations for monitoring emissions, the insufficient sample number and footprint of current measurements have introduced a burden to constrain emissions at fine scales. Promisingly, there are several air-borne sensors collected for NASA's campaigns providing high spatial resolution of NO2 columns. Here, we use the well-characterized NO2 columns from the Airborne Compact Atmospheric Mapper (ACAM) onboard NASA's B200 aircraft into a 1×1 km regional model to constrain anthropogenic NOx emissions in the Houston-Galveston-Brazoria area. Firstly, in order to incorporate the data, we convert the NO2 slant column densities to vertical ones using a joint of a radiative transfer model and the 1x1 km regional model constrained by P3-B aircraft measurements. After conducting an inverse modeling method using the Kalman filter, we find the ACAM observations are resourceful at mitigating the overprediction of model in reproducing NO2 on regular days. Moreover, the ACAM provides a unique opportunity to detect an anomaly in emissions leading to strong air quality degradation that is lacking in previous works. Our study provides convincing evidence that future geostationary satellites with high spatial and temporal resolutions will give us insights into uncertainties associated with the emissions at regional scales.
NASA Astrophysics Data System (ADS)
Lundgren, Paul; Nikkhoo, Mehdi; Samsonov, Sergey V.; Milillo, Pietro; Gil-Cruz, Fernando; Lazo, Jonathan
2017-07-01
Copahue volcano straddling the edge of the Agrio-Caviahue caldera along the Chile-Argentina border in the southern Andes has been in unrest since inflation began in late 2011. We constrain Copahue's source models with satellite and airborne interferometric synthetic aperture radar (InSAR) deformation observations. InSAR time series from descending track RADARSAT-2 and COSMO-SkyMed data span the entire inflation period from 2011 to 2016, with their initially high rates of 12 and 15 cm/yr, respectively, slowing only slightly despite ongoing small eruptions through 2016. InSAR ascending and descending track time series for the 2013-2016 time period constrain a two-source compound dislocation model, with a rate of volume increase of 13 × 106 m3/yr. They consist of a shallow, near-vertical, elongated source centered at 2.5 km beneath the summit and a deeper, shallowly plunging source centered at 7 km depth connecting the shallow source to the deeper caldera. The deeper source is located directly beneath the volcano tectonic seismicity with the lower bounds of the seismicity parallel to the plunge of the deep source. InSAR time series also show normal fault offsets on the NE flank Copahue faults. Coulomb stress change calculations for right-lateral strike slip (RLSS), thrust, and normal receiver faults show positive values in the north caldera for both RLSS and normal faults, suggesting that northward trending seismicity and Copahue fault motion within the caldera are caused by the modeled sources. Together, the InSAR-constrained source model and the seismicity suggest a deep conduit or transfer zone where magma moves from the central caldera to Copahue's upper edifice.
Using Parameter Constraints to Choose State Structures in Cost-Effectiveness Modelling.
Thom, Howard; Jackson, Chris; Welton, Nicky; Sharples, Linda
2017-09-01
This article addresses the choice of state structure in a cost-effectiveness multi-state model. Key model outputs, such as treatment recommendations and prioritisation of future research, may be sensitive to state structure choice. For example, it may be uncertain whether to consider similar disease severities or similar clinical events as the same state or as separate states. Standard statistical methods for comparing models require a common reference dataset but merging states in a model aggregates the data, rendering these methods invalid. We propose a method that involves re-expressing a model with merged states as a model on the larger state space in which particular transition probabilities, costs and utilities are constrained to be equal between states. This produces a model that gives identical estimates of cost effectiveness to the model with merged states, while leaving the data unchanged. The comparison of state structures can be achieved by comparing maximised likelihoods or information criteria between constrained and unconstrained models. We can thus test whether the costs and/or health consequences for a patient in two states are the same, and hence if the states can be merged. We note that different structures can be used for rates, costs and utilities, as appropriate. We illustrate our method with applications to two recent models evaluating the cost effectiveness of prescribing anti-depressant medications by depression severity and the cost effectiveness of diagnostic tests for coronary artery disease. State structures in cost-effectiveness models can be compared using standard methods to compare constrained and unconstrained models.
How alive is constrained SUSY really?
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bechtle, Philip; Desch, Klaus; Dreiner, Herbert K.
2016-05-31
Constrained supersymmetric models like the CMSSM might look less attractive nowadays because of fine tuning arguments. They also might look less probable in terms of Bayesian statistics. The question how well the model under study describes the data, however, is answered by frequentist p-values. Thus, for the first time, we calculate a p-value for a supersymmetric model by performing dedicated global toy fits. We combine constraints from low-energy and astrophysical observables, Higgs boson mass and rate measurements as well as the non-observation of new physics in searches for supersymmetry at the LHC. Furthermore, using the framework Fittino, we perform globalmore » fits of the CMSSM to the toy data and find that this model is excluded at the 90% confidence level.« less
Constraining f(T) teleparallel gravity by big bang nucleosynthesis: f(T) cosmology and BBN.
Capozziello, S; Lambiase, G; Saridakis, E N
2017-01-01
We use Big Bang Nucleosynthesis (BBN) observational data on the primordial abundance of light elements to constrain f ( T ) gravity. The three most studied viable f ( T ) models, namely the power law, the exponential and the square-root exponential are considered, and the BBN bounds are adopted in order to extract constraints on their free parameters. For the power-law model, we find that the constraints are in agreement with those obtained using late-time cosmological data. For the exponential and the square-root exponential models, we show that for reliable regions of parameters space they always satisfy the BBN bounds. We conclude that viable f ( T ) models can successfully satisfy the BBN constraints.
Exploring stellar evolution with gravitational-wave observations
NASA Astrophysics Data System (ADS)
Dvorkin, Irina; Uzan, Jean-Philippe; Vangioni, Elisabeth; Silk, Joseph
2018-05-01
Recent detections of gravitational waves from merging binary black holes opened new possibilities to study the evolution of massive stars and black hole formation. In particular, stellar evolution models may be constrained on the basis of the differences in the predicted distribution of black hole masses and redshifts. In this work we propose a framework that combines galaxy and stellar evolution models and use it to predict the detection rates of merging binary black holes for various stellar evolution models. We discuss the prospects of constraining the shape of the time delay distribution of merging binaries using just the observed distribution of chirp masses. Finally, we consider a generic model of primordial black hole formation and discuss the possibility of distinguishing it from stellar-origin black holes.
Missile Guidance Law Based on Robust Model Predictive Control Using Neural-Network Optimization.
Li, Zhijun; Xia, Yuanqing; Su, Chun-Yi; Deng, Jun; Fu, Jun; He, Wei
2015-08-01
In this brief, the utilization of robust model-based predictive control is investigated for the problem of missile interception. Treating the target acceleration as a bounded disturbance, novel guidance law using model predictive control is developed by incorporating missile inside constraints. The combined model predictive approach could be transformed as a constrained quadratic programming (QP) problem, which may be solved using a linear variational inequality-based primal-dual neural network over a finite receding horizon. Online solutions to multiple parametric QP problems are used so that constrained optimal control decisions can be made in real time. Simulation studies are conducted to illustrate the effectiveness and performance of the proposed guidance control law for missile interception.
Probing Models of Dark Matter and the Early Universe
NASA Astrophysics Data System (ADS)
Orlofsky, Nicholas David
This thesis discusses models for dark matter (DM) and their behavior in the early universe. An important question is how phenomenological probes can directly search for signals of DM today. Another topic of investigation is how the DM and other processes in the early universe must evolve. Then, astrophysical bounds on early universe dynamics can constrain DM. We will consider these questions in the context of three classes of DM models--weakly interacting massive particles (WIMPs), axions, and primordial black holes (PBHs). Starting with WIMPs, we consider models where the DM is charged under the electroweak gauge group of the Standard Model. Such WIMPs, if generated by a thermal cosmological history, are constrained by direct detection experiments. To avoid present or near-future bounds, the WIMP model or cosmological history must be altered in some way. This may be accomplished by the inclusion of new states that coannihilate with the WIMP or a period of non-thermal evolution in the early universe. Future experiments are likely to probe some of these altered scenarios, and a non-observation would require a high degree of tuning in some of the model parameters in these scenarios. Next, axions, as light pseudo-Nambu-Goldstone bosons, are susceptible to quantum fluctuations in the early universe that lead to isocurvature perturbations, which are constrained by observations of the cosmic microwave background (CMB). We ask what it would take to allow axion models in the face of these strong CMB bounds. We revisit models where inflationary dynamics modify the axion potential and discuss how isocurvature bounds can be relaxed, elucidating the difficulties in these constructions. Avoiding disruption of inflationary dynamics provides important limits on the parameter space. Finally, PBHs have received interest in part due to observations by LIGO of merging black hole binaries. We ask how these PBHs could arise through inflationary models and investigate the opportunity for corroboration through experimental probes of gravitational waves at pulsar timing arrays. We provide examples of theories that are already ruled out, theories that will soon be probed, and theories that will not be tested in the foreseeable future. The models that are most strongly constrained are those with relatively broad primordial power spectra.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Marschman, S.C.; Cowin, J.P.; Orlando, T.M.
1998-06-01
'This project involves basic research in chemistry and physics aimed at providing information pertinent to the safe long-term dry storage of spent nuclear fuel (SNF), thousands of tons of which remain in water storage across the DOE complex. The Hanford Site K-Basins alone hold 2,300 tons of spent fuel, much of it severely corroded, and similar situations exist at Savannah River and Idaho National Engineering and Environmental Laboratory. The DOE plans to remove this fuel and seal it in overpack canisters for dry interim storage for up to 75 years while awaiting permanent disposition. Chemically-bound water will remain in thismore » fuel even following proposed drying steps, leading to possible long-term corrosion of the containers and/or fuel rods themselves, generation of H{sub 2} and O{sub 2} gas via radiolysis (which could lead to deflagration or detonation), and reactions of pyrophoric uranium hydrides. No thoroughly tested model is currently available to predict fuel behavior during pre-processing, processing, or storage. In a collaboration between Rutgers University, Pacific Northwest National Laboratory, and Brookhaven National Laboratory, the authors are studying the radiolytic reaction, drying processes, and corrosion behavior of actual SNF materials, and of pure and mixed-phase samples. The authors propose to determine what is omitted from current models: radiolysis of water adsorbed on or in hydrates or hydroxides, thermodynamics of interfacial phases, and kinetics of drying. A model will be developed and tested against actual fuel rod behavior to insure validity and applicability to the problems associated with developing dry storage strategies for DOE-owned SNF. This report summarizes work after eight months of a three-year project.'« less
NASA Astrophysics Data System (ADS)
Aryal, Saurav; Finn, Susanna C.; Hewawasam, Kuravi; Maguire, Ryan; Geddes, George; Cook, Timothy; Martel, Jason; Baumgardner, Jeffrey L.; Chakrabarti, Supriya
2018-05-01
Energies and fluxes of precipitating electrons in an aurora over Lowell, MA on 22-23 June 2015 were derived based on simultaneous, high-resolution (≈ 0.02 nm) brightness measurements of N2+ (427.8 nm, blue line), OI (557.7 nm, green line), and OI (630.0 nm, red line) emissions. The electron energies and energy fluxes as a function of time and look direction were derived by nonlinear minimization of model predictions with respect to the measurements. Three different methods were compared; in the first two methods, we constrained the modeled brightnesses and brightness ratios, respectively, with measurements to simultaneously derive energies and fluxes. Then we used a hybrid method where we constrained the individual modeled brightness ratios with measurements to derive energies and then constrained modeled brightnesses with measurements to derive fluxes. Derived energy, assuming Maxwellian distribution, during this storm ranged from 109 to 262 eV and the total energy flux ranged from 0.8 to 2.2 ergs·cm-2·s-1. This approach provides a way to estimate energies and energy fluxes of the precipitating electrons using simultaneous multispectral measurements.
NASA Astrophysics Data System (ADS)
Chan, C. H.; Brown, G.; Rikvold, P. A.
2017-05-01
A generalized approach to Wang-Landau simulations, macroscopically constrained Wang-Landau, is proposed to simulate the density of states of a system with multiple macroscopic order parameters. The method breaks a multidimensional random-walk process in phase space into many separate, one-dimensional random-walk processes in well-defined subspaces. Each of these random walks is constrained to a different set of values of the macroscopic order parameters. When the multivariable density of states is obtained for one set of values of fieldlike model parameters, the density of states for any other values of these parameters can be obtained by a simple transformation of the total system energy. All thermodynamic quantities of the system can then be rapidly calculated at any point in the phase diagram. We demonstrate how to use the multivariable density of states to draw the phase diagram, as well as order-parameter probability distributions at specific phase points, for a model spin-crossover material: an antiferromagnetic Ising model with ferromagnetic long-range interactions. The fieldlike parameters in this model are an effective magnetic field and the strength of the long-range interaction.
Constrained motion model of mobile robots and its applications.
Zhang, Fei; Xi, Yugeng; Lin, Zongli; Chen, Weidong
2009-06-01
Target detecting and dynamic coverage are fundamental tasks in mobile robotics and represent two important features of mobile robots: mobility and perceptivity. This paper establishes the constrained motion model and sensor model of a mobile robot to represent these two features and defines the k -step reachable region to describe the states that the robot may reach. We show that the calculation of the k-step reachable region can be reduced from that of 2(k) reachable regions with the fixed motion styles to k + 1 such regions and provide an algorithm for its calculation. Based on the constrained motion model and the k -step reachable region, the problems associated with target detecting and dynamic coverage are formulated and solved. For target detecting, the k-step detectable region is used to describe the area that the robot may detect, and an algorithm for detecting a target and planning the optimal path is proposed. For dynamic coverage, the k-step detected region is used to represent the area that the robot has detected during its motion, and the dynamic-coverage strategy and algorithm are proposed. Simulation results demonstrate the efficiency of the coverage algorithm in both convex and concave environments.
A synoptic view of the Third Uniform California Earthquake Rupture Forecast (UCERF3)
Field, Edward; Jordan, Thomas H.; Page, Morgan T.; Milner, Kevin R.; Shaw, Bruce E.; Dawson, Timothy E.; Biasi, Glenn; Parsons, Thomas E.; Hardebeck, Jeanne L.; Michael, Andrew J.; Weldon, Ray; Powers, Peter; Johnson, Kaj M.; Zeng, Yuehua; Bird, Peter; Felzer, Karen; van der Elst, Nicholas; Madden, Christopher; Arrowsmith, Ramon; Werner, Maximillan J.; Thatcher, Wayne R.
2017-01-01
Probabilistic forecasting of earthquake‐producing fault ruptures informs all major decisions aimed at reducing seismic risk and improving earthquake resilience. Earthquake forecasting models rely on two scales of hazard evolution: long‐term (decades to centuries) probabilities of fault rupture, constrained by stress renewal statistics, and short‐term (hours to years) probabilities of distributed seismicity, constrained by earthquake‐clustering statistics. Comprehensive datasets on both hazard scales have been integrated into the Uniform California Earthquake Rupture Forecast, Version 3 (UCERF3). UCERF3 is the first model to provide self‐consistent rupture probabilities over forecasting intervals from less than an hour to more than a century, and it is the first capable of evaluating the short‐term hazards that result from multievent sequences of complex faulting. This article gives an overview of UCERF3, illustrates the short‐term probabilities with aftershock scenarios, and draws some valuable scientific conclusions from the modeling results. In particular, seismic, geologic, and geodetic data, when combined in the UCERF3 framework, reject two types of fault‐based models: long‐term forecasts constrained to have local Gutenberg–Richter scaling, and short‐term forecasts that lack stress relaxation by elastic rebound.
Learning to Model in Engineering
ERIC Educational Resources Information Center
Gainsburg, Julie
2013-01-01
Policymakers and education scholars recommend incorporating mathematical modeling into mathematics education. Limited implementation of modeling instruction in schools, however, has constrained research on how students learn to model, leaving unresolved debates about whether modeling should be reified and explicitly taught as a competence, whether…
Experimental Validation of a Thermoelastic Model for SMA Hybrid Composites
NASA Technical Reports Server (NTRS)
Turner, Travis L.
2001-01-01
This study presents results from experimental validation of a recently developed model for predicting the thermomechanical behavior of shape memory alloy hybrid composite (SMAHC) structures, composite structures with an embedded SMA constituent. The model captures the material nonlinearity of the material system with temperature and is capable of modeling constrained, restrained, or free recovery behavior from experimental measurement of fundamental engineering properties. A brief description of the model and analysis procedures is given, followed by an overview of a parallel effort to fabricate and characterize the material system of SMAHC specimens. Static and dynamic experimental configurations for the SMAHC specimens are described and experimental results for thermal post-buckling and random response are presented. Excellent agreement is achieved between the measured and predicted results, fully validating the theoretical model for constrained recovery behavior of SMAHC structures.
Doyle, Jessica M.; Gleeson, Tom; Manning, Andrew H.; Mayer, K. Ulrich
2015-01-01
Environmental tracers provide information on groundwater age, recharge conditions, and flow processes which can be helpful for evaluating groundwater sustainability and vulnerability. Dissolved noble gas data have proven particularly useful in mountainous terrain because they can be used to determine recharge elevation. However, tracer-derived recharge elevations have not been utilized as calibration targets for numerical groundwater flow models. Herein, we constrain and calibrate a regional groundwater flow model with noble-gas-derived recharge elevations for the first time. Tritium and noble gas tracer results improved the site conceptual model by identifying a previously uncertain contribution of mountain block recharge from the Coast Mountains to an alluvial coastal aquifer in humid southwestern British Columbia. The revised conceptual model was integrated into a three-dimensional numerical groundwater flow model and calibrated to hydraulic head data in addition to recharge elevations estimated from noble gas recharge temperatures. Recharge elevations proved to be imperative for constraining hydraulic conductivity, recharge location, and bedrock geometry, and thus minimizing model nonuniqueness. Results indicate that 45% of recharge to the aquifer is mountain block recharge. A similar match between measured and modeled heads was achieved in a second numerical model that excludes the mountain block (no mountain block recharge), demonstrating that hydraulic head data alone are incapable of quantifying mountain block recharge. This result has significant implications for understanding and managing source water protection in recharge areas, potential effects of climate change, the overall water budget, and ultimately ensuring groundwater sustainability.
Constrained Total Energy Expenditure and Metabolic Adaptation to Physical Activity in Adult Humans.
Pontzer, Herman; Durazo-Arvizu, Ramon; Dugas, Lara R; Plange-Rhule, Jacob; Bovet, Pascal; Forrester, Terrence E; Lambert, Estelle V; Cooper, Richard S; Schoeller, Dale A; Luke, Amy
2016-02-08
Current obesity prevention strategies recommend increasing daily physical activity, assuming that increased activity will lead to corresponding increases in total energy expenditure and prevent or reverse energy imbalance and weight gain [1-3]. Such Additive total energy expenditure models are supported by exercise intervention and accelerometry studies reporting positive correlations between physical activity and total energy expenditure [4] but are challenged by ecological studies in humans and other species showing that more active populations do not have higher total energy expenditure [5-8]. Here we tested a Constrained total energy expenditure model, in which total energy expenditure increases with physical activity at low activity levels but plateaus at higher activity levels as the body adapts to maintain total energy expenditure within a narrow range. We compared total energy expenditure, measured using doubly labeled water, against physical activity, measured using accelerometry, for a large (n = 332) sample of adults living in five populations [9]. After adjusting for body size and composition, total energy expenditure was positively correlated with physical activity, but the relationship was markedly stronger over the lower range of physical activity. For subjects in the upper range of physical activity, total energy expenditure plateaued, supporting a Constrained total energy expenditure model. Body fat percentage and activity intensity appear to modulate the metabolic response to physical activity. Models of energy balance employed in public health [1-3] should be revised to better reflect the constrained nature of total energy expenditure and the complex effects of physical activity on metabolic physiology. Copyright © 2016 Elsevier Ltd. All rights reserved.
CPMC-Lab: A MATLAB package for Constrained Path Monte Carlo calculations
NASA Astrophysics Data System (ADS)
Nguyen, Huy; Shi, Hao; Xu, Jie; Zhang, Shiwei
2014-12-01
We describe CPMC-Lab, a MATLAB program for the constrained-path and phaseless auxiliary-field Monte Carlo methods. These methods have allowed applications ranging from the study of strongly correlated models, such as the Hubbard model, to ab initio calculations in molecules and solids. The present package implements the full ground-state constrained-path Monte Carlo (CPMC) method in MATLAB with a graphical interface, using the Hubbard model as an example. The package can perform calculations in finite supercells in any dimensions, under periodic or twist boundary conditions. Importance sampling and all other algorithmic details of a total energy calculation are included and illustrated. This open-source tool allows users to experiment with various model and run parameters and visualize the results. It provides a direct and interactive environment to learn the method and study the code with minimal overhead for setup. Furthermore, the package can be easily generalized for auxiliary-field quantum Monte Carlo (AFQMC) calculations in many other models for correlated electron systems, and can serve as a template for developing a production code for AFQMC total energy calculations in real materials. Several illustrative studies are carried out in one- and two-dimensional lattices on total energy, kinetic energy, potential energy, and charge- and spin-gaps.
Informing Aerosol Transport Models With Satellite Multi-Angle Aerosol Measurements
NASA Technical Reports Server (NTRS)
Limbacher, J.; Patadia, F.; Petrenko, M.; Martin, M. Val; Chin, M.; Gaitley, B.; Garay, M.; Kalashnikova, O.; Nelson, D.; Scollo, S.
2011-01-01
As the aerosol products from the NASA Earth Observing System's Multi-angle Imaging SpectroRadiometer (MISR) mature, we are placing greater focus on ways of using the aerosol amount and type data products, and aerosol plume heights, to constrain aerosol transport models. We have demonstrated the ability to map aerosol air-mass-types regionally, and have identified product upgrades required to apply them globally, including the need for a quality flag indicating the aerosol type information content, that varies depending upon retrieval conditions. We have shown that MISR aerosol type can distinguish smoke from dust, volcanic ash from sulfate and water particles, and can identify qualitative differences in mixtures of smoke, dust, and pollution aerosol components in urban settings. We demonstrated the use of stereo imaging to map smoke, dust, and volcanic effluent plume injection height, and the combination of MISR and MODIS aerosol optical depth maps to constrain wildfire smoke source strength. This talk will briefly highlight where we stand on these application, with emphasis on the steps we are taking toward applying the capabilities toward constraining aerosol transport models, planet-wide.
Wang, Cong; Du, Hua-qiang; Zhou, Guo-mo; Xu, Xiao-jun; Sun, Shao-bo; Gao, Guo-long
2015-05-01
This research focused on the application of remotely sensed imagery from unmanned aerial vehicle (UAV) with high spatial resolution for the estimation of crown closure of moso bamboo forest based on the geometric-optical model, and analyzed the influence of unconstrained and fully constrained linear spectral mixture analysis (SMA) on the accuracy of the estimated results. The results demonstrated that the combination of UAV remotely sensed imagery and geometric-optical model could, to some degrees, achieve the estimation of crown closure. However, the different SMA methods led to significant differentiation in the estimation accuracy. Compared with unconstrained SMA, the fully constrained linear SMA method resulted in higher accuracy of the estimated values, with the coefficient of determination (R2) of 0.63 at 0.01 level, against the measured values acquired during the field survey. Root mean square error (RMSE) of approximate 0.04 was low, indicating that the usage of fully constrained linear SMA could bring about better results in crown closure estimation, which was closer to the actual condition in moso bamboo forest.
Style consistent classification of isogenous patterns.
Sarkar, Prateek; Nagy, George
2005-01-01
In many applications of pattern recognition, patterns appear together in groups (fields) that have a common origin. For example, a printed word is usually a field of character patterns printed in the same font. A common origin induces consistency of style in features measured on patterns. The features of patterns co-occurring in a field are statistically dependent because they share the same, albeit unknown, style. Style constrained classifiers achieve higher classification accuracy by modeling such dependence among patterns in a field. Effects of style consistency on the distributions of field-features (concatenation of pattern features) can be modeled by hierarchical mixtures. Each field derives from a mixture of styles, while, within a field, a pattern derives from a class-style conditional mixture of Gaussians. Based on this model, an optimal style constrained classifier processes entire fields of patterns rendered in a consistent but unknown style. In a laboratory experiment, style constrained classification reduced errors on fields of printed digits by nearly 25 percent over singlet classifiers. Longer fields favor our classification method because they furnish more information about the underlying style.
Constraints on Lunar Structure from Combined Geochemical, Mineralogical, and Geophysical modeling
NASA Astrophysics Data System (ADS)
Bremner, P. M.; Fuqua, H.; Mallik, A.; Diamond, M. R.; Lock, S. J.; Panovska, S.; Nishikawa, Y.; Jiménez-Pérez, H.; Shahar, A.; Panero, W. R.; Lognonne, P. H.; Faul, U.
2016-12-01
The internal physical and geochemical structure of the Moon is still poorly constrained. Here, we take a multidisciplinary approach to attempt to constrain key parameters of the lunar structure. We use an ensemble of 1-D lunar compositional models with chemically and mineralogically distinct layers, and forward calculated physical parameters, in order to constrain the internal structure. We consider both a chemically well-mixed model with uniform bulk composition, and a chemically stratified model that includes a mantle with preserved mineralogical stratigraphy from magma ocean crystallization. Additionally, we use four different lunar temperature profiles that span the range of proposed selenotherms, giving eight separate sets of lunar models. In each set, we employed a grid search and a differential evolution genetic search algorithm to extensively explore model space, where the thickness of individual compositional layers was varied. In total, we forward calculated over one hundred thousand lunar models. It has been proposed that a dense, partially molten layer exists at the CMB to explain the lack of observed far-side deep moonquakes, the observation of reflected seismic phases from deep moonquakes, and enhanced tidal dissipation. However, subsequent models have proposed that these observables can be explained in other ways. In this study, using a variety of modeling techniques, we find that such a layer may have been formed by overturn of an ilmenite-rich layer, formed after the crystallization of a magma ocean. We therefore include a denser layer (modeled as an ilmenite-rich layer) at both the top and bottom of the lunar mantle in our models. For each set of models, we find models that explain the observed lunar mass and moment of inertia. We find that only a narrow range of core radii are consistent with the mass and moment of inertia constraints. Furthermore, in the chemically well-mixed models, we find that a dense layer is required in the upper mantle to meet the moment of inertia requirement. In no set of models is the mass of the lower dense layer well constrained. For the models that fit the observed mass and moment of inertia, we calculated 1-D seismic velocity profiles, the majority of which compare well with those determined by inverting the Apollo seismic data (Garcia et al., 2011 and Weber et al., 2011).
NASA Astrophysics Data System (ADS)
Bloom, A. Anthony; Bowman, Kevin W.; Lee, Meemong; Turner, Alexander J.; Schroeder, Ronny; Worden, John R.; Weidner, Richard; McDonald, Kyle C.; Jacob, Daniel J.
2017-06-01
Wetland emissions remain one of the principal sources of uncertainty in the global atmospheric methane (CH4) budget, largely due to poorly constrained process controls on CH4 production in waterlogged soils. Process-based estimates of global wetland CH4 emissions and their associated uncertainties can provide crucial prior information for model-based top-down CH4 emission estimates. Here we construct a global wetland CH4 emission model ensemble for use in atmospheric chemical transport models (WetCHARTs version 1.0). Our 0.5° × 0.5° resolution model ensemble is based on satellite-derived surface water extent and precipitation reanalyses, nine heterotrophic respiration simulations (eight carbon cycle models and a data-constrained terrestrial carbon cycle analysis) and three temperature dependence parameterizations for the period 2009-2010; an extended ensemble subset based solely on precipitation and the data-constrained terrestrial carbon cycle analysis is derived for the period 2001-2015. We incorporate the mean of the full and extended model ensembles into GEOS-Chem and compare the model against surface measurements of atmospheric CH4; the model performance (site-level and zonal mean anomaly residuals) compares favourably against published wetland CH4 emissions scenarios. We find that uncertainties in carbon decomposition rates and the wetland extent together account for more than 80 % of the dominant uncertainty in the timing, magnitude and seasonal variability in wetland CH4 emissions, although uncertainty in the temperature CH4 : C dependence is a significant contributor to seasonal variations in mid-latitude wetland CH4 emissions. The combination of satellite, carbon cycle models and temperature dependence parameterizations provides a physically informed structural a priori uncertainty that is critical for top-down estimates of wetland CH4 fluxes. Specifically, our ensemble can provide enhanced information on the prior CH4 emission uncertainty and the error covariance structure, as well as a means for using posterior flux estimates and their uncertainties to quantitatively constrain the biogeochemical process controls of global wetland CH4 emissions.
Observational Role of Dark Matter in f(R) Models for Structure Formation
NASA Astrophysics Data System (ADS)
Verma, Murli Manohar; Yadav, Bal Krishna
The fixed points for the dynamical system in the phase space have been calculated with dark matter in the f(R) gravity models. The stability conditions of these fixed points are obtained in the ongoing accelerated phase of the universe, and the values of the Hubble parameter and Ricci scalar are obtained for various evolutionary stages of the universe. We present a range of some modifications of general relativistic action consistent with the ΛCDM model. We elaborate upon the fact that the upcoming cosmological observations would further constrain the bounds on the possible forms of f(R) with greater precision that could in turn constrain the search for dark matter in colliders.
Multiple R&D projects scheduling optimization with improved particle swarm algorithm.
Liu, Mengqi; Shan, Miyuan; Wu, Juan
2014-01-01
For most enterprises, in order to win the initiative in the fierce competition of market, a key step is to improve their R&D ability to meet the various demands of customers more timely and less costly. This paper discusses the features of multiple R&D environments in large make-to-order enterprises under constrained human resource and budget, and puts forward a multi-project scheduling model during a certain period. Furthermore, we make some improvements to existed particle swarm algorithm and apply the one developed here to the resource-constrained multi-project scheduling model for a simulation experiment. Simultaneously, the feasibility of model and the validity of algorithm are proved in the experiment.
A cost-constrained model of strategic service quality emphasis in nursing homes.
Davis, M A; Provan, K G
1996-02-01
This study employed structural equation modeling to test the relationship between three aspects of the environmental context of nursing homes; Medicaid dependence, ownership status, and market demand, and two basic strategic orientations: low cost and differentiation based on service quality emphasis. Hypotheses were proposed and tested against data collected from a sample of nursing homes operating in a single state. Because of the overwhelming importance of cost control in the nursing home industry, a cost constrained strategy perspective was supported. Specifically, while the three contextual variables had no direct effect on service quality emphasis, the entire model was supported when cost control orientation was introduced as a mediating variable.
NASA Astrophysics Data System (ADS)
Liu, Yuan; Wang, Mingqiang; Ning, Xingyao
2018-02-01
Spinning reserve (SR) should be scheduled considering the balance between economy and reliability. To address the computational intractability cursed by the computation of loss of load probability (LOLP), many probabilistic methods use simplified formulations of LOLP to improve the computational efficiency. Two tradeoffs embedded in the SR optimization model are not explicitly analyzed in these methods. In this paper, two tradeoffs including primary tradeoff and secondary tradeoff between economy and reliability in the maximum LOLP constrained unit commitment (UC) model are explored and analyzed in a small system and in IEEE-RTS System. The analysis on the two tradeoffs can help in establishing new efficient simplified LOLP formulations and new SR optimization models.
NASA Astrophysics Data System (ADS)
Sherrington, David; Davison, Lexie; Buhot, Arnaud; Garrahan, Juan P.
2002-02-01
We report a study of a series of simple model systems with only non-interacting Hamiltonians, and hence simple equilibrium thermodynamics, but with constrained dynamics of a type initially suggested by foams and idealized covalent glasses. We demonstrate that macroscopic dynamical features characteristic of real and more complex model glasses, such as two-time decays in energy and auto-correlation functions, arise from the dynamics and we explain them qualitatively and quantitatively in terms of annihilation-diffusion concepts and theory. The comparison is with strong glasses. We also consider fluctuation-dissipation relations and demonstrate subtleties of interpretation. We find no FDT breakdown when the correct normalization is chosen.
Vibration control of beams using stand-off layer damping: finite element modeling and experiments
NASA Astrophysics Data System (ADS)
Chaudry, A.; Baz, A.
2006-03-01
Damping treatments with stand-off layer (SOL) have been widely accepted as an attractive alternative to conventional constrained layer damping (CLD) treatments. Such an acceptance stems from the fact that the SOL, which is simply a slotted spacer layer sandwiched between the viscoelastic layer and the base structure, acts as a strain magnifier that considerably amplifies the shear strain and hence the energy dissipation characteristics of the viscoelastic layer. Accordingly, more effective vibration suppression can be achieved by using SOL as compared to employing CLD. In this paper, a comprehensive finite element model of the stand-off layer constrained damping treatment is developed. The model accounts for the geometrical and physical parameters of the slotted SOL, the viscoelastic, layer the constraining layer, and the base structure. The predictions of the model are validated against the predictions of a distributed transfer function model and a model built using a commercial finite element code (ANSYS). Furthermore, the theoretical predictions are validated experimentally for passive SOL treatments of different configurations. The obtained results indicate a close agreement between theory and experiments. Furthermore, the obtained results demonstrate the effectiveness of the CLD with SOL in enhancing the energy dissipation as compared to the conventional CLD. Extension of the proposed one-dimensional CLD with SOL to more complex structures is a natural extension to the present study.
Shu, Bao; Liu, Hui; Xu, Longwei; Qian, Chuang; Gong, Xiaopeng; An, Xiangdong
2018-04-14
For GPS medium-long baseline real-time kinematic (RTK) positioning, the troposphere parameter is introduced along with coordinates, and the model is ill-conditioned due to its strong correlation with the height parameter. For BeiDou Navigation Satellite System (BDS), additional difficulties occur due to its special satellite constellation. In fact, relative zenith troposphere delay (RZTD) derived from high-precision empirical zenith troposphere models can be introduced. Thus, the model strength can be improved, which is also called the RZTD-constrained RTK model. In this contribution, we first analyze the factors affecting the precision of BDS medium-long baseline RTK; thereafter, 15 baselines ranging from 38 km to 167 km in different troposphere conditions are processed to assess the performance of RZTD-constrained RTK. Results show that the troposphere parameter is difficult to distinguish from the height component, even with long time filtering for BDS-only RTK. Due to the lack of variation in geometry for the BDS geostationary Earth orbit satellite, the long convergence time of ambiguity parameters may reduce the height precision of GPS/BDS-combined RTK in the initial period. When the RZTD-constrained model was used in BDS and GPS/BDS-combined situations compared with the traditional RTK, the standard deviation of the height component for the fixed solution was reduced by 52.4% and 34.0%, respectively.
Liu, Hui; Xu, Longwei; Qian, Chuang; Gong, Xiaopeng; An, Xiangdong
2018-01-01
For GPS medium-long baseline real-time kinematic (RTK) positioning, the troposphere parameter is introduced along with coordinates, and the model is ill-conditioned due to its strong correlation with the height parameter. For BeiDou Navigation Satellite System (BDS), additional difficulties occur due to its special satellite constellation. In fact, relative zenith troposphere delay (RZTD) derived from high-precision empirical zenith troposphere models can be introduced. Thus, the model strength can be improved, which is also called the RZTD-constrained RTK model. In this contribution, we first analyze the factors affecting the precision of BDS medium-long baseline RTK; thereafter, 15 baselines ranging from 38 km to 167 km in different troposphere conditions are processed to assess the performance of RZTD-constrained RTK. Results show that the troposphere parameter is difficult to distinguish from the height component, even with long time filtering for BDS-only RTK. Due to the lack of variation in geometry for the BDS geostationary Earth orbit satellite, the long convergence time of ambiguity parameters may reduce the height precision of GPS/BDS-combined RTK in the initial period. When the RZTD-constrained model was used in BDS and GPS/BDS-combined situations compared with the traditional RTK, the standard deviation of the height component for the fixed solution was reduced by 52.4% and 34.0%, respectively. PMID:29661999
Fault offsets and lateral crustal movement on Europa - Evidence for a mobile ice shell
NASA Technical Reports Server (NTRS)
Schenk, Paul M.; Mckinnon, William B.
1989-01-01
An examination is conducted of Europa's cross-cutting structural relationships between various lineament types, in order to constrain the type of structure involved in each such case and, where possible, to also constrain the degree of extension across the lineaments. Evidence is adduced for significant lateral crustal movement, allowing alternative models and mechanisms for lineament formation to be discussed, as well as plausible lithospheric and crustal models. The question as to whether any of the water-ice layer has been, or currently is, liquid, is also treated in light of the evidence obtained.
NASA Astrophysics Data System (ADS)
Tringe, J. W.; Kercher, J. R.; Springer, H. K.; Glascoe, E. A.; Levie, H. W.; Hsu, P.; Willey, T. M.; Molitoris, J. D.
2013-07-01
We employ in-situ flash x-ray imaging, together with a detailed multiphase convective burn model, to demonstrate how explosives' binder characteristics influence the burning processes in thermal explosions. Our study focuses on the HMX-based explosives LX-10 and PBX 9501. While the HMX (cyclotetramethylene-tetranitramine) crystallite size distributions for these two explosives are nearly identical before heating, our experiments and simulations indicate that after heating, variations result due to differences in binder composition. Post-ignition flash x-ray images reveal that the average density decreases at late times more rapidly in PBX 9501 than LX-10, suggesting a faster conductive burning rate in PBX-9501. Heated permeability measurements in LX-10 and PBX 9501 demonstrate that the binder system characteristics influence the evolution of connected porosity. Once ignited, connected porosity provides pathways for product gas heating ahead of the reaction front and additional surface area for burning, facilitating the transition from conductive to convective burning modes. A multiphase convective burn model implemented in the ALE3D code is used to better understand the influence on burn rates of material properties such as porosity and effective thermally damaged particle size. In this context, particles are defined as gas-impermeable binder-coated crystallites and agglomerations with a set of effective radii reff. Model results demonstrate quantitative agreement with containment wall velocity for confined PBX 9501 and LX-10, and qualitative agreement with density as a function of position in the burning explosive. The model predicts a decrease in post-ignition containment wall velocity with larger radii in reff. These experimental data and model results together provide insight into the initiation and propagation of the reaction wave that defines the convective burn front in HMX-based explosives, a necessary step toward predicting violence under a broad range of conditions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Saide, Pablo E.; Peterson, David A.; de Silva, Arlindo
We couple airborne, ground-based, and satellite observations; conduct regional simulations; and develop and apply an inversion technique to constrain hourly smoke emissions from the Rim Fire, the third largest observed in California, USA. Emissions constrained with multiplatform data show notable nocturnal enhancements (sometimes over a factor of 20), correlate better with daily burned area data, and are a factor of 2–4 higher than a priori estimates, highlighting the need for improved characterization of diurnal profiles and day-to-day variability when modeling extreme fires. Constraining only with satellite data results in smaller enhancements mainly due to missing retrievals near the emissions source,more » suggesting that top-down emission estimates for these events could be underestimated and a multiplatform approach is required to resolve them. Predictions driven by emissions constrained with multiplatform data present significant variations in downwind air quality and in aerosol feedback on meteorology, emphasizing the need for improved emissions estimates during exceptional events.« less
Active/Passive Control of Sound Radiation from Panels using Constrained Layer Damping
NASA Technical Reports Server (NTRS)
Gibbs, Gary P.; Cabell, Randolph H.
2003-01-01
A hybrid passive/active noise control system utilizing constrained layer damping and model predictive feedback control is presented. This system is used to control the sound radiation of panels due to broadband disturbances. To facilitate the hybrid system design, a methodology for placement of constrained layer damping which targets selected modes based on their relative radiated sound power is developed. The placement methodology is utilized to determine two constrained layer damping configurations for experimental evaluation of a hybrid system. The first configuration targets the (4,1) panel mode which is not controllable by the piezoelectric control actuator, and the (2,3) and (5,2) panel modes. The second configuration targets the (1,1) and (3,1) modes. The experimental results demonstrate the improved reduction of radiated sound power using the hybrid passive/active control system as compared to the active control system alone.
NASA Astrophysics Data System (ADS)
Chandran, A.; Schulz, Marc D.; Burnell, F. J.
2016-12-01
Many phases of matter, including superconductors, fractional quantum Hall fluids, and spin liquids, are described by gauge theories with constrained Hilbert spaces. However, thermalization and the applicability of quantum statistical mechanics has primarily been studied in unconstrained Hilbert spaces. In this paper, we investigate whether constrained Hilbert spaces permit local thermalization. Specifically, we explore whether the eigenstate thermalization hypothesis (ETH) holds in a pinned Fibonacci anyon chain, which serves as a representative case study. We first establish that the constrained Hilbert space admits a notion of locality by showing that the influence of a measurement decays exponentially in space. This suggests that the constraints are no impediment to thermalization. We then provide numerical evidence that ETH holds for the diagonal and off-diagonal matrix elements of various local observables in a generic disorder-free nonintegrable model. We also find that certain nonlocal observables obey ETH.
Degree-constrained multicast routing for multimedia communications
NASA Astrophysics Data System (ADS)
Wang, Yanlin; Sun, Yugeng; Li, Guidan
2005-02-01
Multicast services have been increasingly used by many multimedia applications. As one of the key techniques to support multimedia applications, the rational and effective multicast routing algorithms are very important to networks performance. When switch nodes in networks have different multicast capability, multicast routing problem is modeled as the degree-constrained Steiner problem. We presented two heuristic algorithms, named BMSTA and BSPTA, for the degree-constrained case in multimedia communications. Both algorithms are used to generate degree-constrained multicast trees with bandwidth and end to end delay bound. Simulations over random networks were carried out to compare the performance of the two proposed algorithms. Experimental results show that the proposed algorithms have advantages in traffic load balancing, which can avoid link blocking and enhance networks performance efficiently. BMSTA has better ability in finding unsaturated links and (or) unsaturated nodes to generate multicast trees than BSPTA. The performance of BMSTA is affected by the variation of degree constraints.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wright, Warren P.; Nagaraj, Gautam; Kneller, James P.
It has long been recognized that the neutrinos detected from the next core-collapse supernova in the Galaxy have the potential to reveal important information about the dynamics of the explosion and the nucleosynthesis conditions as well as allowing us to probe the properties of the neutrino itself. The neutrinos emitted from thermonuclear—type Ia—supernovae also possess the same potential, although these supernovae are dimmer neutrino sources. For the first time, we calculate the time, energy, line of sight, and neutrino-flavor-dependent features of the neutrino signal expected from a three-dimensional delayed-detonation explosion simulation, where a deflagration-to-detonation transition triggers the complete disruption ofmore » a near-Chandrasekhar mass carbon-oxygen white dwarf. We also calculate the neutrino flavor evolution along eight lines of sight through the simulation as a function of time and energy using an exact three-flavor transformation code. We identify a characteristic spectral peak at ˜10 MeV as a signature of electron captures on copper. This peak is a potentially distinguishing feature of explosion models since it reflects the nucleosynthesis conditions early in the explosion. We simulate the event rates in the Super-K, Hyper-K, JUNO, and DUNE neutrino detectors with the SNOwGLoBES event rate calculation software and also compute the IceCube signal. Hyper-K will be able to detect neutrinos from our model out to a distance of ˜10 kpc. Here, at 1 kpc, JUNO, Super-K, and DUNE would register a few events while IceCube and Hyper-K would register several tens of events.« less
Good initialization model with constrained body structure for scene text recognition
NASA Astrophysics Data System (ADS)
Zhu, Anna; Wang, Guoyou; Dong, Yangbo
2016-09-01
Scene text recognition has gained significant attention in the computer vision community. Character detection and recognition are the promise of text recognition and affect the overall performance to a large extent. We proposed a good initialization model for scene character recognition from cropped text regions. We use constrained character's body structures with deformable part-based models to detect and recognize characters in various backgrounds. The character's body structures are achieved by an unsupervised discriminative clustering approach followed by a statistical model and a self-build minimum spanning tree model. Our method utilizes part appearance and location information, and combines character detection and recognition in cropped text region together. The evaluation results on the benchmark datasets demonstrate that our proposed scheme outperforms the state-of-the-art methods both on scene character recognition and word recognition aspects.
Splines and polynomial tools for flatness-based constrained motion planning
NASA Astrophysics Data System (ADS)
Suryawan, Fajar; De Doná, José; Seron, María
2012-08-01
This article addresses the problem of trajectory planning for flat systems with constraints. Flat systems have the useful property that the input and the state can be completely characterised by the so-called flat output. We propose a spline parametrisation for the flat output, the performance output, the states and the inputs. Using this parametrisation the problem of constrained trajectory planning can be cast into a simple quadratic programming problem. An important result is that the B-spline parametrisation used gives exact results for constrained linear continuous-time system. The result is exact in the sense that the constrained signal can be made arbitrarily close to the boundary without having intersampling issues (as one would have in sampled-data systems). Simulation examples are presented, involving the generation of rest-to-rest trajectories. In addition, an experimental result of the method is also presented, where two methods to generate trajectories for a magnetic-levitation (maglev) system in the presence of constraints are compared and each method's performance is discussed. The first method uses the nonlinear model of the plant, which turns out to belong to the class of flat systems. The second method uses a linearised version of the plant model around an operating point. In every case, a continuous-time description is used. The experimental results on a real maglev system reported here show that, in most scenarios, the nonlinear and linearised models produce almost similar, indistinguishable trajectories.
Constraining the mass of the Local Group
NASA Astrophysics Data System (ADS)
Carlesi, Edoardo; Hoffman, Yehuda; Sorce, Jenny G.; Gottlöber, Stefan
2017-03-01
The mass of the Local Group (LG) is a crucial parameter for galaxy formation theories. However, its observational determination is challenging - its mass budget is dominated by dark matter that cannot be directly observed. To meet this end, the posterior distributions of the LG and its massive constituents have been constructed by means of constrained and random cosmological simulations. Two priors are assumed - the Λ cold dark matter model that is used to set up the simulations, and an LG model that encodes the observational knowledge of the LG and is used to select LG-like objects from the simulations. The constrained simulations are designed to reproduce the local cosmography as it is imprinted on to the Cosmicflows-2 data base of velocities. Several prescriptions are used to define the LG model, focusing in particular on different recent estimates of the tangential velocity of M31. It is found that (a) different vtan choices affect the peak mass values up to a factor of 2, and change mass ratios of MM31 to MMW by up to 20 per cent; (b) constrained simulations yield more sharply peaked posterior distributions compared with the random ones; (c) LG mass estimates are found to be smaller than those found using the timing argument; (d) preferred Milky Way masses lie in the range of (0.6-0.8) × 1012 M⊙; whereas (e) MM31 is found to vary between (1.0-2.0) × 1012 M⊙, with a strong dependence on the vtan values used.
Modeling and query the uncertainty of network constrained moving objects based on RFID data
NASA Astrophysics Data System (ADS)
Han, Liang; Xie, Kunqing; Ma, Xiujun; Song, Guojie
2007-06-01
The management of network constrained moving objects is more and more practical, especially in intelligent transportation system. In the past, the location information of moving objects on network is collected by GPS, which cost high and has the problem of frequent update and privacy. The RFID (Radio Frequency IDentification) devices are used more and more widely to collect the location information. They are cheaper and have less update. And they interfere in the privacy less. They detect the id of the object and the time when moving object passed by the node of the network. They don't detect the objects' exact movement in side the edge, which lead to a problem of uncertainty. How to modeling and query the uncertainty of the network constrained moving objects based on RFID data becomes a research issue. In this paper, a model is proposed to describe the uncertainty of network constrained moving objects. A two level index is presented to provide efficient access to the network and the data of movement. The processing of imprecise time-slice query and spatio-temporal range query are studied in this paper. The processing includes four steps: spatial filter, spatial refinement, temporal filter and probability calculation. Finally, some experiments are done based on the simulated data. In the experiments the performance of the index is studied. The precision and recall of the result set are defined. And how the query arguments affect the precision and recall of the result set is also discussed.
Configuration of the thermal landscape determines thermoregulatory performance of ectotherms
Sears, Michael W.; Angilletta, Michael J.; Schuler, Matthew S.; Borchert, Jason; Dilliplane, Katherine F.; Stegman, Monica; Rusch, Travis W.; Mitchell, William A.
2016-01-01
Although most organisms thermoregulate behaviorally, biologists still cannot easily predict whether mobile animals will thermoregulate in natural environments. Current models fail because they ignore how the spatial distribution of thermal resources constrains thermoregulatory performance over space and time. To overcome this limitation, we modeled the spatially explicit movements of animals constrained by access to thermal resources. Our models predict that ectotherms thermoregulate more accurately when thermal resources are dispersed throughout space than when these resources are clumped. This prediction was supported by thermoregulatory behaviors of lizards in outdoor arenas with known distributions of environmental temperatures. Further, simulations showed how the spatial structure of the landscape qualitatively affects responses of animals to climate. Biologists will need spatially explicit models to predict impacts of climate change on local scales. PMID:27601639
Using Coronal Hole Maps to Constrain MHD Models
NASA Astrophysics Data System (ADS)
Caplan, Ronald M.; Downs, Cooper; Linker, Jon A.; Mikic, Zoran
2017-08-01
In this presentation, we explore the use of coronal hole maps (CHMs) as a constraint for thermodynamic MHD models of the solar corona. Using our EUV2CHM software suite (predsci.com/chd), we construct CHMs from SDO/AIA 193Å and STEREO-A/EUVI 195Å images for multiple Carrington rotations leading up to the August 21st, 2017 total solar eclipse. We then contruct synoptic CHMs from synthetic EUV images generated from global thermodynamic MHD simulations of the corona for each rotation. Comparisons of apparent coronal hole boundaries and estimates of the net open flux are used to benchmark and constrain our MHD model leading up to the eclipse. Specifically, the comparisons are used to find optimal parameterizations of our wave turbulence dissipation (WTD) coronal heating model.
GROWTH AND INEQUALITY: MODEL EVALUATION BASED ON AN ESTIMATION-CALIBRATION STRATEGY
Jeong, Hyeok; Townsend, Robert
2010-01-01
This paper evaluates two well-known models of growth with inequality that have explicit micro underpinnings related to household choice. With incomplete markets or transactions costs, wealth can constrain investment in business and the choice of occupation and also constrain the timing of entry into the formal financial sector. Using the Thai Socio-Economic Survey (SES), we estimate the distribution of wealth and the key parameters that best fit cross-sectional data on household choices and wealth. We then simulate the model economies for two decades at the estimated initial wealth distribution and analyze whether the model economies at those micro-fit parameter estimates can explain the observed macro and sectoral aspects of income growth and inequality change. Both models capture important features of Thai reality. Anomalies and comparisons across the two distinct models yield specific suggestions for improved research on the micro foundations of growth and inequality. PMID:20448833
Constraining viscous dark energy models with the latest cosmological data
NASA Astrophysics Data System (ADS)
Wang, Deng; Yan, Yang-Jie; Meng, Xin-He
2017-10-01
Based on the assumption that the dark energy possessing bulk viscosity is homogeneously and isotropically permeated in the universe, we propose three new viscous dark energy (VDE) models to characterize the accelerating universe. By constraining these three models with the latest cosmological observations, we find that they just deviate very slightly from the standard cosmological model and can alleviate effectively the current H_0 tension between the local observation by the Hubble Space Telescope and the global measurement by the Planck Satellite. Interestingly, we conclude that a spatially flat universe in our VDE model with cosmic curvature is still supported by current data, and the scale invariant primordial power spectrum is strongly excluded at least at the 5.5σ confidence level in the three VDE models as the Planck result. We also give the 95% upper limits of the typical bulk viscosity parameter η in the three VDE scenarios.
Luo, Xiaohui; Wang, Hang; Fan, Yubo
2007-04-01
This study was aimed to develop a 3-D finite element (3-D FE) model of the mental fractured mandible and design the boundary constrains. The CT images from a health volunteer were used as the original information and put into ANSYS program to build a 3-D FE model. The model of the miniplate and screw which were used for the internal fixation was established by Pro/E. The boundary constrains of different muscle loadings were used to simulate the 3 functional conditions of the mandible. A 3-D FE model of mental fractured mandible under the miniplate-screw internal fixation system was constructed. And by the boundary constraints, the 3 biting conditions were simulated and the model could serve as a foundation on which to analyze the biomechanical behavior of the fractured mandible.
NASA Technical Reports Server (NTRS)
Abercromby, Kira J.; Rapp, Jason; Bedard, Donald; Seitzer, Patrick; Cardona, Tommaso; Cowardin, Heather; Barker, Ed; Lederer, Susan
2013-01-01
Spectral reflectance data through the visible regime was collected at Las Campanas Observatory in Chile using an imaging spectrograph on one of the twin 6.5-m Magellan telescopes. The data were obtained on 1-2 May 2012 on the 'Landon Clay' telescope with the LDSS3 (Low Dispersion Survey Spectrograph 3). Five pieces of Geosynchronous Orbit (GEO) or near-GEO debris were identified and observed with an exposure time of 30 seconds on average. In addition, laboratory spectral reflectance data was collected using an Analytical Spectral Device (ASD) field spectrometer at California Polytechnic State University (Cal Poly) in San Luis Obispo on several typical common spacecraft materials including solar cells, circuit boards, various Kapton materials used for multi-layer insulation, and various paints. The remotely collected data and the laboratory-acquired data were then incorporated in a newly developed model that uses a constrained least squares method to unmix the spectrum in specific material components. The results of this model are compared to the previous method of a human-in-the-loop (considered here the traditional method) that identifies possible material components by varying the materials and percentages until a spectral match is obtained. The traditional model was found to match the remotely collected spectral data after it had been divided by the continuum to remove the space weathering effects, or a reddening of the materials. The constrained least-squares model also used the de-reddened spectra as inputs and the results were consistent with those obtained through the traditional method. For comparison, a first-order examination of including reddening effects into the constrained least-squares model will be explored and comparisons to the remotely collected data will be examined. The identification of each object s suspected material component will be discussed herein.
NASA Technical Reports Server (NTRS)
Rapp, Jason; Abercromby, Kira J.; Bedard, Donald; Seitzer, Patrick; Cardona, Tommaso; Cowardin, Heather; Barker, Ed; Lederer, Susan
2012-01-01
Spectral reflectance data through the visible regime was collected at Las Campanas Observatory in Chile using an imaging spectrograph on one of the twin 6.5-m Magellan telescopes. The data were obtained on 1-2 May 2012 on the 'Landon Clay' telescope with the LDSS3 (Low Dispersion Survey Spectrograph 3). Five pieces of Geosynchronous Orbit (GEO) or near-GEO debris were identified and observed with an exposure time of 30 seconds on average. In addition, laboratory spectral reflectance data was collected using an Analytical Spectral Device (ASD) field spectrometer at California Polytechnic State University in San Luis Obispo on several typical common spacecraft materials including solar cells, circuit boards, various Kapton materials used for multi-layer insulation, and various paints. The remotely collected data and the laboratory-acquired data were then incorporated in a newly developed model that uses a constrained least squares method to unmix the spectrum in specific material components. The results of this model are compared to the previous method of a human-in-the-loop (considered here the traditional method) that identifies possible material components by varying the materials and percentages until a spectral match is obtained. The traditional model was found to match the remotely collected spectral data after it had been divided by the continuum to remove the space weathering effects, or a "reddening" of the materials. The constrained least-squares model also used the de-reddened spectra as inputs and the results were consistent with those obtained through the traditional method. For comparison, a first-order examination of including reddening effects into the constrained least-squares model will be explored and comparisons to the remotely collected data will be examined. The identification of each object's suspected material component will be discussed herein.
Section-constrained local geological interface dynamic updating method based on the HRBF surface
NASA Astrophysics Data System (ADS)
Guo, Jiateng; Wu, Lixin; Zhou, Wenhui; Li, Chaoling; Li, Fengdan
2018-02-01
Boundaries, attitudes and sections are the most common data acquired from regional field geological surveys, and they are used for three-dimensional (3D) geological modelling. However, constructing topologically consistent 3D geological models from rapid and automatic regional modelling with convenient local modifications remains unresolved. In previous works, the Hermite radial basis function (HRBF) surface was introduced for the simulation of geological interfaces from geological boundaries and attitudes, which allows 3D geological models to be automatically extracted from the modelling area by the interfaces. However, the reasonability and accuracy of non-supervised subsurface modelling is limited without further modifications generated through explanations and analyses performed by geology experts. In this paper, we provide flexible and convenient manual interactive manipulation tools for geologists to sketch constraint lines, and these tools may help geologists transform and apply their expert knowledge to the models. In the modified modelling workflow, the geological sections were treated as auxiliary constraints to construct more reasonable 3D geological models. The geometric characteristics of section lines were abstracted to coordinates and normal vectors, and along with the transformed coordinates and vectors from boundaries and attitudes, these characteristics were adopted to co-calculate the implicit geological surface function parameters of the HRBF equations and form constrained geological interfaces from topographic (boundaries and attitudes) and subsurface data (sketched sections). Based on this new modelling method, a prototype system was developed, in which the section lines could be imported from databases or interactively sketched, and the models could be immediately updated after the new constraints were added. Experimental comparisons showed that all boundary, attitude and section data are well represented in the constrained models, which are consistent with expert explanations and help improve the quality of the models.
Development and application of theoretical models for Rotating Detonation Engine flowfields
NASA Astrophysics Data System (ADS)
Fievisohn, Robert
As turbine and rocket engine technology matures, performance increases between successive generations of engine development are becoming smaller. One means of accomplishing significant gains in thermodynamic performance and power density is to use detonation-based heat release instead of deflagration. This work is focused on developing and applying theoretical models to aid in the design and understanding of Rotating Detonation Engines (RDEs). In an RDE, a detonation wave travels circumferentially along the bottom of an annular chamber where continuous injection of fresh reactants sustains the detonation wave. RDEs are currently being designed, tested, and studied as a viable option for developing a new generation of turbine and rocket engines that make use of detonation heat release. One of the main challenges in the development of RDEs is to understand the complex flowfield inside the annular chamber. While simplified models are desirable for obtaining timely performance estimates for design analysis, one-dimensional models may not be adequate as they do not provide flow structure information. In this work, a two-dimensional physics-based model is developed, which is capable of modeling the curved oblique shock wave, exit swirl, counter-flow, detonation inclination, and varying pressure along the inflow boundary. This is accomplished by using a combination of shock-expansion theory, Chapman-Jouguet detonation theory, the Method of Characteristics (MOC), and other compressible flow equations to create a shock-fitted numerical algorithm and generate an RDE flowfield. This novel approach provides a numerically efficient model that can provide performance estimates as well as details of the large-scale flow structures in seconds on a personal computer. Results from this model are validated against high-fidelity numerical simulations that may require a high-performance computing framework to provide similar performance estimates. This work provides a designer a new tool to conduct large-scale parametric studies to optimize a design space before conducting computationally-intensive, high-fidelity simulations that may be used to examine additional effects. The work presented in this thesis not only bridges the gap between simple one-dimensional models and high-fidelity full numerical simulations, but it also provides an effective tool for understanding and exploring RDE flow processes.
Chance Constrained Programming Methods in Probabilistic Programming.
1982-03-01
Financial and Quantitative Analysis 2, 1967. Also reproduced in R. F. Byrne et. al., eds.5tudies in Budgeting (Amsterdam: North Holland, 1971 ). [3...Rules for the E-Model of Chance-Constrained Programming," Management Science, 17, 1971 . [23] Garstka, S. J. "The Economic Equivalence of Several...Iowa City: The University of Iowa College of Business Administration, 1981). -3- (29] Kall , P. and A. Prekopa, eds, Recent Results in Stochastic
Magnetohydrodynamical Effects on Nuclear Deflagration Fronts in Type Ia Supernovae
NASA Astrophysics Data System (ADS)
Hristov, Boyan; Collins, David C.; Hoeflich, Peter; Weatherford, Charles A.; Diamond, Tiara R.
2018-05-01
This article presents a study of the effects of magnetic fields on non-distributed nuclear burning fronts as a possible solution to a fundamental problem for the thermonuclear explosion of a Chandrasekhar mass ({M}Ch}) white dwarf (WD), the currently favored scenario for the majority of Type Ia SNe. All existing 3D hydrodynamical simulations predict strong global mixing of the burning products due to Rayleigh–Taylor (RT) instabilities, which contradicts observations. As a first step toward studying the flame physics, we present a set of computational magnet-hydrodynamic models in rectangular flux tubes, resembling a small inner region of a WD. We consider initial magnetic fields up to {10}12 {{G}} of various orientations. We find an increasing suppression of RT instabilities starting at about {10}9 {{G}}. The front speed tends to decrease with increasing magnitude up to about {10}11 {{G}}. For even higher fields new small-scale, finger-like structures develop, which increase the burning speed by a factor of 3 to 4 above the field-free RT-dominated regime. We suggest that the new instability may provide sufficiently accelerated energy production during the distributed burning regime to go over the Chapman–Jougey limit and trigger a detonation. Finally, we discuss the possible origins of high magnetic fields during the final stage of the progenitor evolution or the explosion.
NASA Astrophysics Data System (ADS)
Xavier, Prince K.; Petch, Jon C.; Klingaman, Nicholas P.; Woolnough, Steve J.; Jiang, Xianan; Waliser, Duane E.; Caian, Mihaela; Cole, Jason; Hagos, Samson M.; Hannay, Cecile; Kim, Daehyun; Miyakawa, Tomoki; Pritchard, Michael S.; Roehrig, Romain; Shindo, Eiki; Vitart, Frederic; Wang, Hailan
2015-05-01
An analysis of diabatic heating and moistening processes from 12 to 36 h lead time forecasts from 12 Global Circulation Models are presented as part of the "Vertical structure and physical processes of the Madden-Julian Oscillation (MJO)" project. A lead time of 12-36 h is chosen to constrain the large-scale dynamics and thermodynamics to be close to observations while avoiding being too close to the initial spin-up of the models as they adjust to being driven from the Years of Tropical Convection (YOTC) analysis. A comparison of the vertical velocity and rainfall with the observations and YOTC analysis suggests that the phases of convection associated with the MJO are constrained in most models at this lead time although the rainfall in the suppressed phase is typically overestimated. Although the large-scale dynamics is reasonably constrained, moistening and heating profiles have large intermodel spread. In particular, there are large spreads in convective heating and moistening at midlevels during the transition to active convection. Radiative heating and cloud parameters have the largest relative spread across models at upper levels during the active phase. A detailed analysis of time step behavior shows that some models show strong intermittency in rainfall and differences in the precipitation and dynamics relationship between models. The wealth of model outputs archived during this project is a very valuable resource for model developers beyond the study of the MJO. In addition, the findings of this study can inform the design of process model experiments, and inform the priorities for field experiments and future observing systems.
NASA Astrophysics Data System (ADS)
Kattge, J.; Knorr, W.; Raddatz, T.; Wirth, C.
2009-04-01
Photosynthetic capacity is one of the most sensitive parameters of terrestrial biosphere models whose representation in global scale simulations has been severely hampered by a lack of systematic analyses using a sufficiently broad database. Due to its coupling to stomatal conductance changes in the parameterisation of photosynthetic capacity may potentially influence transpiration rates and vegetation surface temperature. Here, we provide a constrained parameterisation of photosynthetic capacity for different plant functional types in the context of the photosynthesis model proposed by Farquhar et al. (1980), based on a comprehensive compilation of leaf photosynthesis rates and leaf nitrogen content. Mean values of photosynthetic capacity were implemented into the coupled climate-vegetation model ECHAM5/JSBACH and modelled gross primary production (GPP) is compared to a compilation of independent observations on stand scale. Compared to the current standard parameterisation the root-mean-squared difference between modelled and observed GPP is substantially reduced for almost all PFTs by the new parameterisation of photosynthetic capacity. We find a systematic depression of NUE (photosynthetic capacity divided by leaf nitrogen content) on certain tropical soils that are known to be deficient in phosphorus. Photosynthetic capacity of tropical trees derived by this study is substantially lower than standard estimates currently used in terrestrial biosphere models. This causes a decrease of modelled GPP while it significantly increases modelled tropical vegetation surface temperatures, up to 0.8°C. These results emphasise the importance of a constrained parameterisation of photosynthetic capacity not only for the carbon cycle, but also for the climate system.
Fielding-Miller, Rebecca; Dunkle, Kristin
2018-01-01
Women who engage in transactional sex are more likely to experience intimate partner violence (IPV) and are at higher risk of HIV. However, women engage in transactional sex for a variety of reasons and the precise mechanism linking transactional sex and IPV is not fully understood. We conducted a behavioural survey with a cross-sectional sample of 401 women attending 1 rural and 1 urban public antenatal clinic in Swaziland between February and June 2014. We used structural equation modelling to identify and measure constrained relationship agency (CRA) as a latent variable, and then tested the hypothesis that CRA plays a significant role in the pathway between IPV and transactional sex. After controlling for CRA, receiving more material goods from a sexual partner was not associated with higher levels of physical or sexual IPV and was protective against emotional IPV. CRA was the single largest predictor of IPV, and more education was associated with decreased levels of constrained relationship agency. Policies and interventions that target transactional sex as a driver of IPV and HIV may be more successful if they instead target the broader social landscape that constrains women’s agency and drives the harmful aspects of transactional sex. PMID:29132281
Innovative scheme for high-repetition-rate imaging of CN radical.
Satija, Aman; Ruesch, Morgan D; Powell, Michael S; Son, Steven F; Lucht, Robert P
2018-02-01
We have employed, to the best of our knowledge, a novel excitation scheme to perform the first high-repetition-rate planar laser-induced fluorescence (PLIF) measurements of a CN radical in combustion. The third harmonic of a Nd:YVO 4 laser at 355 nm due to its relatively large linewidth overlaps with several R branch transitions in a CN ground electronic state. Therefore, the 355 nm beam was employed to directly excite the CN transitions with good efficiency. The CN measurements were performed in premixed CH 4 -N 2 O flames with varying equivalence ratios. A detailed characterization of the high-speed CN PLIF imaging system is presented via its ability to capture statistical and dynamical information in these premixed flames. Single-shot CN PLIF images obtained over a HMX pellet undergoing self-supported deflagration are presented as an example of the imaging system being applied towards characterizing the flame structure of energetic materials.
NASA Technical Reports Server (NTRS)
Mirels, Harold
1959-01-01
A source distribution method is presented for obtaining flow perturbations due to small unsteady area variations, mass, momentum, and heat additions in a basic uniform (or piecewise uniform) one-dimensional flow. First, the perturbations due to an elemental area variation, mass, momentum, and heat addition are found. The general solution is then represented by a spatial and temporal distribution of these elemental (source) solutions. Emphasis is placed on discussing the physical nature of the flow phenomena. The method is illustrated by several examples. These include the determination of perturbations in basic flows consisting of (1) a shock propagating through a nonuniform tube, (2) a constant-velocity piston driving a shock, (3) ideal shock-tube flows, and (4) deflagrations initiated at a closed end. The method is particularly applicable for finding the perturbations due to relatively thin wall boundary layers.
Universal explosive detection system for homeland security applications
NASA Astrophysics Data System (ADS)
Lee, Vincent Y.; Bromberg, Edward E. A.
2010-04-01
L-3 Communications CyTerra Corporation has developed a high throughput universal explosive detection system (PassPort) to automatically screen the passengers in airports without requiring them to remove their shoes. The technical approach is based on the patented energetic material detection (EMD) technology. By analyzing the results of sample heating with an infrared camera, one can distinguish the deflagration or decomposition of an energetic material from other clutters such as flammables and general background substances. This becomes the basis of a universal explosive detection system that does not require a library and is capable of detecting trace levels of explosives with a low false alarm rate. The PassPort is a simple turnstile type device and integrates a non-intrusive aerodynamic sampling scheme that has been shown capable of detecting trace levels of explosives on shoes. A detailed description of the detection theory and the automated sampling techniques, as well as the field test results, will be presented.
Tritium glovebox stripper system seismic design evaluation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Grinnell, J. J.; Klein, J. E.
2015-09-01
The use of glovebox confinement at US Department of Energy (DOE) tritium facilities has been discussed in numerous publications. Glovebox confinement protects the workers from radioactive material (especially tritium oxide), provides an inert atmosphere for prevention of flammable gas mixtures and deflagrations, and allows recovery of tritium released from the process into the glovebox when a glovebox stripper system (GBSS) is part of the design. Tritium recovery from the glovebox atmosphere reduces emissions from the facility and the radiological dose to the public. Location of US DOE defense programs facilities away from public boundaries also aids in reducing radiological dosesmore » to the public. This is a study based upon design concepts to identify issues and considerations for design of a Seismic GBSS. Safety requirements and analysis should be considered preliminary. Safety requirements for design of GBSS should be developed and finalized as a part of the final design process.« less
Study of thermite mixtures consolidated by cold gas dynamic spray process
NASA Astrophysics Data System (ADS)
Bacciochini, Antoine; Maines, Geoffrey; Poupart, Christian; Radulescu, Matei; Jodoin, Bertrand; Lee, Julian
2013-06-01
The present study focused on the cold gas dynamic spray process for manufacturing finely structured energetic materials with high reactivity, vanishing porosity, as well as structural integrity and arbitrary shape. The experiments have focused the reaction between the aluminum and metal oxides, such as Al-CuO and Al-MoO3 systems. To increase the reactivity, an initial mechanical activation was achieved through interrupted ball milling. The consolidation of the materials used the supersonic cold gas spray technique, where the particles are accelerated to high speeds and consolidated via plastic deformation upon impact, forming activated nano-composites in arbitrary shapes with close to zero porosity. This technique permits to retain the feedstock powder micro-structure and prevents any reactions during the consolidation phase. Reactivity of mixtures has been investigated through flame propagation analysis on cold sprayed samples and compacted powder mixture. Deflagration tests showed the influence of porosity on the reactivity.
Review on factors affecting the performance of pulse detonation engine
NASA Astrophysics Data System (ADS)
Tripathi, Saurabh; Pandey, Krishna Murari
2018-04-01
Now a day's rocket engines (air-breathing type) are being used for aerospace purposes but the studies have shown that these are less efficient, so alternatives are being searched for these. Pulse Detonation Engine (PDE) is one such efficient engine which can replace the rocket engines. In this review paper, different researches have been cited. As can be observed from various researches, insertion of obstacles is better. Deflagration to Detonation(DDT) transition process is found to be most important factor. So a lot of researches are being done considering this DDT chamber. Also, the ignition chamber and ejector were found to improve the effectiveness of PDE. The PDE works with a range of Mach 0-4. Flame acceleration is also found to increase the DDT process. Use of valve and valveless engine has also been compared. Various other factors have been focused in this review paper which is found to boost PDE performance.
Avalanches in Mn12-Acetate: ``Magnetic Burning"
NASA Astrophysics Data System (ADS)
McHugh, Sean; Suzuki, Y.; Graybill, D.; Sarachik, M. P.; Avraham, N.; Myasoedov, Y.; Shtrikman, H.; Zeldov, E.; Bagai, R.; Chakov, N. E.; Christou, G.
2006-03-01
From local time-resolved measurements of fast reversal of the magnetization in single crystals of the molecular magnet Mn12-acetate, we have shown[1] that the magnetization avalanche spreads as a narrow interface that propagates through the crystal at a constant velocity roughly two orders of magnitude smaller than the speed of sound. This phenomenon is closely analogous to the propagation of a flame front (deflagration) through a flammable chemical substance. The propagation speed of the avalanche depends on the energy stored in each molecule, which can be controlled and tuned using an external magnetic field. We report studies of propagation speed with different external fields in Mn12-acetate. [1] Yoko Suzuki, M.P. Sarachik, E.M. Chudnovsky, S. McHugh, R. Gonzalez-Rubio, N. Avraham, Y. Myasoedov, H. Shtrikman, E. Zeldov, N.E. Chakov and G. Christou, Phys. Rev. Lett. 95, 147201 (2005).
NASA Astrophysics Data System (ADS)
Belkina, T. A.; Konyukhova, N. B.; Kurochkin, S. V.
2016-01-01
Previous and new results are used to compare two mathematical insurance models with identical insurance company strategies in a financial market, namely, when the entire current surplus or its constant fraction is invested in risky assets (stocks), while the rest of the surplus is invested in a risk-free asset (bank account). Model I is the classical Cramér-Lundberg risk model with an exponential claim size distribution. Model II is a modification of the classical risk model (risk process with stochastic premiums) with exponential distributions of claim and premium sizes. For the survival probability of an insurance company over infinite time (as a function of its initial surplus), there arise singular problems for second-order linear integrodifferential equations (IDEs) defined on a semiinfinite interval and having nonintegrable singularities at zero: model I leads to a singular constrained initial value problem for an IDE with a Volterra integral operator, while II model leads to a more complicated nonlocal constrained problem for an IDE with a non-Volterra integral operator. A brief overview of previous results for these two problems depending on several positive parameters is given, and new results are presented. Additional results are concerned with the formulation, analysis, and numerical study of "degenerate" problems for both models, i.e., problems in which some of the IDE parameters vanish; moreover, passages to the limit with respect to the parameters through which we proceed from the original problems to the degenerate ones are singular for small and/or large argument values. Such problems are of mathematical and practical interest in themselves. Along with insurance models without investment, they describe the case of surplus completely invested in risk-free assets, as well as some noninsurance models of surplus dynamics, for example, charity-type models.
Constraining the noncommutative spectral action via astrophysical observations.
Nelson, William; Ochoa, Joseph; Sakellariadou, Mairi
2010-09-03
The noncommutative spectral action extends our familiar notion of commutative spaces, using the data encoded in a spectral triple on an almost commutative space. Varying a rather simple action, one can derive all of the standard model of particle physics in this setting, in addition to a modified version of Einstein-Hilbert gravity. In this Letter we use observations of pulsar timings, assuming that no deviation from general relativity has been observed, to constrain the gravitational sector of this theory. While the bounds on the coupling constants remain rather weak, they are comparable to existing bounds on deviations from general relativity in other settings and are likely to be further constrained by future observations.
NASA Astrophysics Data System (ADS)
Stritzinger, M. D.; Valenti, S.; Hoeflich, P.; Baron, E.; Phillips, M. M.; Taddia, F.; Foley, R. J.; Hsiao, E. Y.; Jha, S. W.; McCully, C.; Pandya, V.; Simon, J. D.; Benetti, S.; Brown, P. J.; Burns, C. R.; Campillay, A.; Contreras, C.; Förster, F.; Holmbo, S.; Marion, G. H.; Morrell, N.; Pignata, G.
2015-01-01
We present ultraviolet through near-infrared (NIR) broadband photometry, and visual-wavelength and NIR spectroscopy of the Type Iax supernova (SN) 2012Z. The data set consists of both early- and late-time observations, including the first late phase NIR spectrum obtained for a spectroscopically classified SN Iax. Simple model calculations of its bolometric light curve suggest SN 2012Z produced ~0.3 M⊙ of 56Ni, ejected about a Chandrasekhar mass of material, and had an explosion energy of ~1051 erg, making it one of the brightest (MB = -18.3 mag) and most energetic SN Iax yet observed. The late phase (+269d) NIRspectrum of SN 2012Z is found to broadly resemble similar epoch spectra of normal SNe Ia; however, like other SNe Iax, corresponding visual-wavelength spectra differ substantially from all supernova types. Constraints from the distribution of intermediate mass elements, e.g., silicon and magnesium, indicate that the outer ejecta did not experience significant mixing during or after burning, and the late phase NIR line profiles suggests most of the 56Ni is produced during high density burning. The various observational properties of SN 2012Z are found to be consistent with the theoretical expectations of a Chandrasekhar mass white dwarf progenitor that experiences a pulsational delayed detonation, which produced several tenths of a solar mass of 56Ni during the deflagration burning phase and little (or no) 56Ni during the detonation phase. Within this scenario only a moderate amount of Rayleigh-Taylor mixing occurs both during the deflagration and fallback phase of the pulsation, and the layered structure of the intermediate mass elements is a product of the subsequent denotation phase. The fact that the SNe Iax population does not follow a tight brightness-decline relation similar to SNe Ia can then be understood in the framework of variable amounts of mixing during pulsational rebound and variable amounts of 56Ni production during the early subsonic phase of expansion. Based on observations collected at the European Organization for Astronomical Research in the Southern Hemisphere, Chile (ESO Program 088.D-0222, 184.D-1152), the Magellan 6.5 m telescopes at Las Campanas Observatory, and the Nordic Optical Telescope, operated by the Nordic Optical Telescope Scientific Association at the Observatorio del Roque de los Muchachos, La Palma, Spain, of the Instituto de Astrofisica de Canarias; also partially based on observations made with the Southern African Large Telescope (SALT), and the W. M. Keck Observatory located on the summit of Mauna Kea.Appendix A and Tables 1-5 are available in electronic form at http://www.aanda.orgFITS files of the reduced spectra are only available at the CDS via anonymous ftp to http://cdsarc.u-strasbg.fr (ftp://130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/573/A2
A BRST formulation for the conic constrained particle
NASA Astrophysics Data System (ADS)
Barbosa, Gabriel D.; Thibes, Ronaldo
2018-04-01
We describe the gauge invariant BRST formulation of a particle constrained to move in a general conic. The model considered constitutes an explicit example of an originally second-class system which can be quantized within the BRST framework. We initially impose the conic constraint by means of a Lagrange multiplier leading to a consistent second-class system which generalizes previous models studied in the literature. After calculating the constraint structure and the corresponding Dirac brackets, we introduce a suitable first-order Lagrangian, the resulting modified system is then shown to be gauge invariant. We proceed to the extended phase space introducing fermionic ghost variables, exhibiting the BRST symmetry transformations and writing the Green’s function generating functional for the BRST quantized model.
Spin vectors in the Koronis family: III. (832) Karin
NASA Astrophysics Data System (ADS)
Slivan, Stephen M.; Molnar, Lawrence A.
2012-08-01
Studies of asteroid families constrain models of asteroid collisions and evolution processes, and the Karin cluster within the Koronis family is among the youngest families known (Nesvorný, D., Bottke, Jr., W.F., Dones, L., Levison, H.F. [2002]. Nature 417, 720-722). (832) Karin itself is by far the largest member of the Karin cluster, thus knowledge of Karin's spin vector is important to constrain family formation and evolution models that include spin, and to test whether its spin properties are consistent with the Karin cluster being a very young family. We observed rotation lightcurves of Karin during its four consecutive apparitions in 2006-2009, and combined the new observations with previously published lightcurves to determine its spin vector orientation and preliminary model shape. Karin is a prograde rotator with a period of (18.352 ± 0.003) h, spin obliquity near (42 ± 5)°, and pole ecliptic longitude near either (52 ± 5)° or (230 ± 5)°. The spin vector and shape results for Karin will constrain models of family formation that include spin properties; in the meantime we briefly discuss Karin's own spin in the context of those of other members of the Karin cluster and the parent body's siblings in the Koronis family.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Caprini, Chiara; Tamanini, Nicola, E-mail: chiara.caprini@cea.fr, E-mail: nicola.tamanini@cea.fr
We perform a forecast analysis of the capability of the eLISA space-based interferometer to constrain models of early and interacting dark energy using gravitational wave standard sirens. We employ simulated catalogues of standard sirens given by merging massive black hole binaries visible by eLISA, with an electromagnetic counterpart detectable by future telescopes. We consider three-arms mission designs with arm length of 1, 2 and 5 million km, 5 years of mission duration and the best-level low frequency noise as recently tested by the LISA Pathfinder. Standard sirens with eLISA give access to an intermediate range of redshift 1 ∼< zmore » ∼< 8, and can therefore provide competitive constraints on models where the onset of the deviation from ΛCDM (i.e. the epoch when early dark energy starts to be non-negligible, or when the interaction with dark matter begins) occurs relatively late, at z ∼< 6. If instead early or interacting dark energy is relevant already in the pre-recombination era, current cosmological probes (especially the cosmic microwave background) are more efficient than eLISA in constraining these models, except possibly in the interacting dark energy model if the energy exchange is proportional to the energy density of dark energy.« less
NASA Astrophysics Data System (ADS)
Zhang, Chenglong; Guo, Ping
2017-10-01
The vague and fuzzy parametric information is a challenging issue in irrigation water management problems. In response to this problem, a generalized fuzzy credibility-constrained linear fractional programming (GFCCFP) model is developed for optimal irrigation water allocation under uncertainty. The model can be derived from integrating generalized fuzzy credibility-constrained programming (GFCCP) into a linear fractional programming (LFP) optimization framework. Therefore, it can solve ratio optimization problems associated with fuzzy parameters, and examine the variation of results under different credibility levels and weight coefficients of possibility and necessary. It has advantages in: (1) balancing the economic and resources objectives directly; (2) analyzing system efficiency; (3) generating more flexible decision solutions by giving different credibility levels and weight coefficients of possibility and (4) supporting in-depth analysis of the interrelationships among system efficiency, credibility level and weight coefficient. The model is applied to a case study of irrigation water allocation in the middle reaches of Heihe River Basin, northwest China. Therefore, optimal irrigation water allocation solutions from the GFCCFP model can be obtained. Moreover, factorial analysis on the two parameters (i.e. λ and γ) indicates that the weight coefficient is a main factor compared with credibility level for system efficiency. These results can be effective for support reasonable irrigation water resources management and agricultural production.
Microenvironmental independence associated with tumor progression.
Anderson, Alexander R A; Hassanein, Mohamed; Branch, Kevin M; Lu, Jenny; Lobdell, Nichole A; Maier, Julie; Basanta, David; Weidow, Brandy; Narasanna, Archana; Arteaga, Carlos L; Reynolds, Albert B; Quaranta, Vito; Estrada, Lourdes; Weaver, Alissa M
2009-11-15
Tumor-microenvironment interactions are increasingly recognized to influence tumor progression. To understand the competitive dynamics of tumor cells in diverse microenvironments, we experimentally parameterized a hybrid discrete-continuum mathematical model with phenotypic trait data from a set of related mammary cell lines with normal, transformed, or tumorigenic properties. Surprisingly, in a resource-rich microenvironment, with few limitations on proliferation or migration, transformed (but not tumorigenic) cells were most successful and outcompeted other cell types in heterogeneous tumor simulations. Conversely, constrained microenvironments with limitations on space and/or growth factors gave a selective advantage to phenotypes derived from tumorigenic cell lines. Analysis of the relative performance of each phenotype in constrained versus unconstrained microenvironments revealed that, although all cell types grew more slowly in resource-constrained microenvironments, the most aggressive cells were least affected by microenvironmental constraints. A game theory model testing the relationship between microenvironment resource availability and competitive cellular dynamics supports the concept that microenvironmental independence is an advantageous cellular trait in resource-limited microenvironments.
Low-dimensional recurrent neural network-based Kalman filter for speech enhancement.
Xia, Youshen; Wang, Jun
2015-07-01
This paper proposes a new recurrent neural network-based Kalman filter for speech enhancement, based on a noise-constrained least squares estimate. The parameters of speech signal modeled as autoregressive process are first estimated by using the proposed recurrent neural network and the speech signal is then recovered from Kalman filtering. The proposed recurrent neural network is globally asymptomatically stable to the noise-constrained estimate. Because the noise-constrained estimate has a robust performance against non-Gaussian noise, the proposed recurrent neural network-based speech enhancement algorithm can minimize the estimation error of Kalman filter parameters in non-Gaussian noise. Furthermore, having a low-dimensional model feature, the proposed neural network-based speech enhancement algorithm has a much faster speed than two existing recurrent neural networks-based speech enhancement algorithms. Simulation results show that the proposed recurrent neural network-based speech enhancement algorithm can produce a good performance with fast computation and noise reduction. Copyright © 2015 Elsevier Ltd. All rights reserved.
Tests of gravity with future space-based experiments
NASA Astrophysics Data System (ADS)
Sakstein, Jeremy
2018-03-01
Future space-based tests of relativistic gravitation—laser ranging to Phobos, accelerometers in orbit, and optical networks surrounding Earth—will constrain the theory of gravity with unprecedented precision by testing the inverse-square law, the strong and weak equivalence principles, and the deflection and time delay of light by massive bodies. In this paper, we estimate the bounds that could be obtained on alternative gravity theories that use screening mechanisms to suppress deviations from general relativity in the Solar System: chameleon, symmetron, and Galileon models. We find that space-based tests of the parametrized post-Newtonian parameter γ will constrain chameleon and symmetron theories to new levels, and that tests of the inverse-square law using laser ranging to Phobos will provide the most stringent constraints on Galileon theories to date. We end by discussing the potential for constraining these theories using upcoming tests of the weak equivalence principle, and conclude that further theoretical modeling is required in order to fully utilize the data.
Calculations of Hubbard U from first-principles
NASA Astrophysics Data System (ADS)
Aryasetiawan, F.; Karlsson, K.; Jepsen, O.; Schönberger, U.
2006-09-01
The Hubbard U of the 3d transition metal series as well as SrVO3 , YTiO3 , Ce, and Gd has been estimated using a recently proposed scheme based on the random-phase approximation. The values obtained are generally in good accord with the values often used in model calculations but for some cases the estimated values are somewhat smaller than those used in the literature. We have also calculated the frequency-dependent U for some of the materials. The strong frequency dependence of U in some of the cases considered in this paper suggests that the static value of U may not be the most appropriate one to use in model calculations. We have also made comparison with the constrained local density approximation (LDA) method and found some discrepancies in a number of cases. We emphasize that our scheme and the constrained local density approximation LDA method theoretically ought to give similar results and the discrepancies may be attributed to technical difficulties in performing calculations based on currently implemented constrained LDA schemes.
A Algebraic Approach to the Quantization of Constrained Systems: Finite Dimensional Examples.
NASA Astrophysics Data System (ADS)
Tate, Ranjeet Shekhar
1992-01-01
General relativity has two features in particular, which make it difficult to apply to it existing schemes for the quantization of constrained systems. First, there is no background structure in the theory, which could be used, e.g., to regularize constraint operators, to identify a "time" or to define an inner product on physical states. Second, in the Ashtekar formulation of general relativity, which is a promising avenue to quantum gravity, the natural variables for quantization are not canonical; and, classically, there are algebraic identities between them. Existing schemes are usually not concerned with such identities. Thus, from the point of view of canonical quantum gravity, it has become imperative to find a framework for quantization which provides a general prescription to find the physical inner product, and is flexible enough to accommodate non -canonical variables. In this dissertation I present an algebraic formulation of the Dirac approach to the quantization of constrained systems. The Dirac quantization program is augmented by a general principle to find the inner product on physical states. Essentially, the Hermiticity conditions on physical operators determine this inner product. I also clarify the role in quantum theory of possible algebraic identities between the elementary variables. I use this approach to quantize various finite dimensional systems. Some of these models test the new aspects of the algebraic framework. Others bear qualitative similarities to general relativity, and may give some insight into the pitfalls lurking in quantum gravity. The previous quantizations of one such model had many surprising features. When this model is quantized using the algebraic program, there is no longer any unexpected behaviour. I also construct the complete quantum theory for a previously unsolved relativistic cosmology. All these models indicate that the algebraic formulation provides powerful new tools for quantization. In (spatially compact) general relativity, the Hamiltonian is constrained to vanish. I present various approaches one can take to obtain an interpretation of the quantum theory of such "dynamically constrained" systems. I apply some of these ideas to the Bianchi I cosmology, and analyze the issue of the initial singularity in quantum theory.
NASA Astrophysics Data System (ADS)
Wang, Jun; Xu, Xiaoguang; Henze, Daven K.; Zeng, Jing; Ji, Qiang; Tsay, Si-Chee; Huang, Jianping
2012-04-01
Predicting the influences of dust on atmospheric composition, climate, and human health requires accurate knowledge of dust emissions, but large uncertainties persist in quantifying mineral sources. This study presents a new method for combined use of satellite-measured radiances and inverse modeling to spatially constrain the amount and location of dust emissions. The technique is illustrated with a case study in May 2008; the dust emissions in Taklimakan and Gobi deserts are spatially optimized using the GEOS-Chem chemical transport model and its adjoint constrained by aerosol optical depth (AOD) that are derived over the downwind dark-surface region in China from MODIS (Moderate Resolution Imaging Spectroradiometer) reflectance with the aerosol single scattering properties consistent with GEOS-chem. The adjoint inverse modeling yields an overall 51% decrease in prior dust emissions estimated by GEOS-Chem over the Taklimakan-Gobi area, with more significant reductions south of the Gobi Desert. The model simulation with optimized dust emissions shows much better agreement with independent observations from MISR (Multi-angle Imaging SpectroRadiometer) AOD and MODIS Deep Blue AOD over the dust source region and surface PM10 concentrations. The technique of this study can be applied to global multi-sensor remote sensing data for constraining dust emissions at various temporal and spatial scales, and hence improving the quantification of dust effects on climate, air quality, and human health.
NASA Technical Reports Server (NTRS)
Wang, Jun; Xu, Xiaoguang; Henze, Daven K.; Zeng, Jing; Ji, Qiang; Tsay, Si-Chee; Huang, Jianping
2012-01-01
Predicting the influences of dust on atmospheric composition, climate, and human health requires accurate knowledge of dust emissions, but large uncertainties persist in quantifying mineral sources. This study presents a new method for combined use of satellite-measured radiances and inverse modeling to spatially constrain the amount and location of dust emissions. The technique is illustrated with a case study in May 2008; the dust emissions in Taklimakan and Gobi deserts are spatially optimized using the GEOSChem chemical transport model and its adjoint constrained by aerosol optical depth (AOD) that are derived over the downwind dark-surface region in China from MODIS (Moderate Resolution Imaging Spectroradiometer) reflectance with the aerosol single scattering properties consistent with GEOS-chem. The adjoint inverse modeling yields an overall 51% decrease in prior dust emissions estimated by GEOS-Chem over the Taklimakan-Gobi area, with more significant reductions south of the Gobi Desert. The model simulation with optimized dust emissions shows much better agreement with independent observations from MISR (Multi-angle Imaging SpectroRadiometer) AOD and MODIS Deep Blue AOD over the dust source region and surface PM10 concentrations. The technique of this study can be applied to global multi-sensor remote sensing data for constraining dust emissions at various temporal and spatial scales, and hence improving the quantification of dust effects on climate, air quality, and human health.
Finding viable models in SUSY parameter spaces with signal specific discovery potential
NASA Astrophysics Data System (ADS)
Burgess, Thomas; Lindroos, Jan Øye; Lipniacka, Anna; Sandaker, Heidi
2013-08-01
Recent results from ATLAS giving a Higgs mass of 125.5 GeV, further constrain already highly constrained supersymmetric models such as pMSSM or CMSSM/mSUGRA. As a consequence, finding potentially discoverable and non-excluded regions of model parameter space is becoming increasingly difficult. Several groups have invested large effort in studying the consequences of Higgs mass bounds, upper limits on rare B-meson decays, and limits on relic dark matter density on constrained models, aiming at predicting superpartner masses, and establishing likelihood of SUSY models compared to that of the Standard Model vis-á-vis experimental data. In this paper a framework for efficient search for discoverable, non-excluded regions of different SUSY spaces giving specific experimental signature of interest is presented. The method employs an improved Markov Chain Monte Carlo (MCMC) scheme exploiting an iteratively updated likelihood function to guide search for viable models. Existing experimental and theoretical bounds as well as the LHC discovery potential are taken into account. This includes recent bounds on relic dark matter density, the Higgs sector and rare B-mesons decays. A clustering algorithm is applied to classify selected models according to expected phenomenology enabling automated choice of experimental benchmarks and regions to be used for optimizing searches. The aim is to provide experimentalist with a viable tool helping to target experimental signatures to search for, once a class of models of interest is established. As an example a search for viable CMSSM models with τ-lepton signatures observable with the 2012 LHC data set is presented. In the search 105209 unique models were probed. From these, ten reference benchmark points covering different ranges of phenomenological observables at the LHC were selected.
A Constrained Linear Estimator for Multiple Regression
ERIC Educational Resources Information Center
Davis-Stober, Clintin P.; Dana, Jason; Budescu, David V.
2010-01-01
"Improper linear models" (see Dawes, Am. Psychol. 34:571-582, "1979"), such as equal weighting, have garnered interest as alternatives to standard regression models. We analyze the general circumstances under which these models perform well by recasting a class of "improper" linear models as "proper" statistical models with a single predictor. We…
Modeling of Density-Dependent Flow based on the Thermodynamically Constrained Averaging Theory
NASA Astrophysics Data System (ADS)
Weigand, T. M.; Schultz, P. B.; Kelley, C. T.; Miller, C. T.; Gray, W. G.
2016-12-01
The thermodynamically constrained averaging theory (TCAT) has been used to formulate general classes of porous medium models, including new models for density-dependent flow. The TCAT approach provides advantages that include a firm connection between the microscale, or pore scale, and the macroscale; a thermodynamically consistent basis; explicit inclusion of factors such as a diffusion that arises from gradients associated with pressure and activity and the ability to describe both high and low concentration displacement. The TCAT model is presented and closure relations for the TCAT model are postulated based on microscale averages and a parameter estimation is performed on a subset of the experimental data. Due to the sharpness of the fronts, an adaptive moving mesh technique was used to ensure grid independent solutions within the run time constraints. The optimized parameters are then used for forward simulations and compared to the set of experimental data not used for the parameter estimation.
Feasibility Assessment of a Fine-Grained Access Control Model on Resource Constrained Sensors.
Uriarte Itzazelaia, Mikel; Astorga, Jasone; Jacob, Eduardo; Huarte, Maider; Romaña, Pedro
2018-02-13
Upcoming smart scenarios enabled by the Internet of Things (IoT) envision smart objects that provide services that can adapt to user behavior or be managed to achieve greater productivity. In such environments, smart things are inexpensive and, therefore, constrained devices. However, they are also critical components because of the importance of the information that they provide. Given this, strong security is a requirement, but not all security mechanisms in general and access control models in particular are feasible. In this paper, we present the feasibility assessment of an access control model that utilizes a hybrid architecture and a policy language that provides dynamic fine-grained policy enforcement in the sensors, which requires an efficient message exchange protocol called Hidra. This experimental performance assessment includes a prototype implementation, a performance evaluation model, the measurements and related discussions, which demonstrate the feasibility and adequacy of the analyzed access control model.
Feasibility Assessment of a Fine-Grained Access Control Model on Resource Constrained Sensors
Huarte, Maider; Romaña, Pedro
2018-01-01
Upcoming smart scenarios enabled by the Internet of Things (IoT) envision smart objects that provide services that can adapt to user behavior or be managed to achieve greater productivity. In such environments, smart things are inexpensive and, therefore, constrained devices. However, they are also critical components because of the importance of the information that they provide. Given this, strong security is a requirement, but not all security mechanisms in general and access control models in particular are feasible. In this paper, we present the feasibility assessment of an access control model that utilizes a hybrid architecture and a policy language that provides dynamic fine-grained policy enforcement in the sensors, which requires an efficient message exchange protocol called Hidra. This experimental performance assessment includes a prototype implementation, a performance evaluation model, the measurements and related discussions, which demonstrate the feasibility and adequacy of the analyzed access control model. PMID:29438338
Nonlinear evolution of coarse-grained quantum systems with generalized purity constraints
NASA Astrophysics Data System (ADS)
Burić, Nikola
2010-12-01
Constrained quantum dynamics is used to propose a nonlinear dynamical equation for pure states of a generalized coarse-grained system. The relevant constraint is given either by the generalized purity or by the generalized invariant fluctuation, and the coarse-grained pure states correspond to the generalized coherent, i.e. generalized nonentangled states. Open system model of the coarse-graining is discussed. It is shown that in this model and in the weak coupling limit the constrained dynamical equations coincide with an equation for pointer states, based on Hilbert-Schmidt distance, that was previously suggested in the context of the decoherence theory.
Uncertainty analysis of depth predictions from seismic reflection data using Bayesian statistics
NASA Astrophysics Data System (ADS)
Michelioudakis, Dimitrios G.; Hobbs, Richard W.; Caiado, Camila C. S.
2018-03-01
Estimating the depths of target horizons from seismic reflection data is an important task in exploration geophysics. To constrain these depths we need a reliable and accurate velocity model. Here, we build an optimum 2D seismic reflection data processing flow focused on pre - stack deghosting filters and velocity model building and apply Bayesian methods, including Gaussian process emulation and Bayesian History Matching (BHM), to estimate the uncertainties of the depths of key horizons near the borehole DSDP-258 located in the Mentelle Basin, south west of Australia, and compare the results with the drilled core from that well. Following this strategy, the tie between the modelled and observed depths from DSDP-258 core was in accordance with the ± 2σ posterior credibility intervals and predictions for depths to key horizons were made for the two new drill sites, adjacent the existing borehole of the area. The probabilistic analysis allowed us to generate multiple realizations of pre-stack depth migrated images, these can be directly used to better constrain interpretation and identify potential risk at drill sites. The method will be applied to constrain the drilling targets for the upcoming International Ocean Discovery Program (IODP), leg 369.
Uncertainty analysis of depth predictions from seismic reflection data using Bayesian statistics
NASA Astrophysics Data System (ADS)
Michelioudakis, Dimitrios G.; Hobbs, Richard W.; Caiado, Camila C. S.
2018-06-01
Estimating the depths of target horizons from seismic reflection data is an important task in exploration geophysics. To constrain these depths we need a reliable and accurate velocity model. Here, we build an optimum 2-D seismic reflection data processing flow focused on pre-stack deghosting filters and velocity model building and apply Bayesian methods, including Gaussian process emulation and Bayesian History Matching, to estimate the uncertainties of the depths of key horizons near the Deep Sea Drilling Project (DSDP) borehole 258 (DSDP-258) located in the Mentelle Basin, southwest of Australia, and compare the results with the drilled core from that well. Following this strategy, the tie between the modelled and observed depths from DSDP-258 core was in accordance with the ±2σ posterior credibility intervals and predictions for depths to key horizons were made for the two new drill sites, adjacent to the existing borehole of the area. The probabilistic analysis allowed us to generate multiple realizations of pre-stack depth migrated images, these can be directly used to better constrain interpretation and identify potential risk at drill sites. The method will be applied to constrain the drilling targets for the upcoming International Ocean Discovery Program, leg 369.
Constraining Modern and Historic Mercury Emissions From Gold Mining
NASA Astrophysics Data System (ADS)
Strode, S. A.; Jaeglé, L.; Selin, N. E.; Sunderland, E.
2007-12-01
Mercury emissions from both historic gold and silver mining and modern small-scale gold mining are highly uncertain. Historic mercury emissions can affect the modern atmosphere through reemission from land and ocean, and quantifying mercury emissions from historic gold and silver mining can help constrain modern mining sources. While estimates of mercury emissions during historic gold rushes exceed modern anthropogenic mercury emissions in North America, sediment records in many regions do not show a strong gold rush signal. We use the GEOS-Chem chemical transport model to determine the spatial footprint of mercury emissions from mining and compare model runs from gold rush periods to sediment and ice core records of historic mercury deposition. Based on records of gold and silver production, we include mercury emissions from North and South American mining of 1900 Mg/year in 1880, compared to modern global anthropogenic emissions of 3400 Mg/year. Including this large mining source in GEOS-Chem leads to an overestimate of the modeled 1880 to preindustrial enhancement ratio compared to the sediment core record. We conduct sensitivity studies to constrain the level of mercury emissions from modern and historic mining that is consistent with the deposition records for different regions.
Xie, Y L; Li, Y P; Huang, G H; Li, Y F; Chen, L R
2011-04-15
In this study, an inexact-chance-constrained water quality management (ICC-WQM) model is developed for planning regional environmental management under uncertainty. This method is based on an integration of interval linear programming (ILP) and chance-constrained programming (CCP) techniques. ICC-WQM allows uncertainties presented as both probability distributions and interval values to be incorporated within a general optimization framework. Complexities in environmental management systems can be systematically reflected, thus applicability of the modeling process can be highly enhanced. The developed method is applied to planning chemical-industry development in Binhai New Area of Tianjin, China. Interval solutions associated with different risk levels of constraint violation have been obtained. They can be used for generating decision alternatives and thus help decision makers identify desired policies under various system-reliability constraints of water environmental capacity of pollutant. Tradeoffs between system benefits and constraint-violation risks can also be tackled. They are helpful for supporting (a) decision of wastewater discharge and government investment, (b) formulation of local policies regarding water consumption, economic development and industry structure, and (c) analysis of interactions among economic benefits, system reliability and pollutant discharges. Copyright © 2011 Elsevier B.V. All rights reserved.
Capacity-constrained traffic assignment in networks with residual queues
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lam, W.H.K.; Zhang, Y.
2000-04-01
This paper proposes a capacity-constrained traffic assignment model for strategic transport planning in which the steady-state user equilibrium principle is extended for road networks with residual queues. Therefore, the road-exit capacity and the queuing effects can be incorporated into the strategic transport model for traffic forecasting. The proposed model is applicable to the congested network particularly when the traffic demands exceeds the capacity of the network during the peak period. An efficient solution method is proposed for solving the steady-state traffic assignment problem with residual queues. Then a simple numerical example is employed to demonstrate the application of the proposedmore » model and solution method, while an example of a medium-sized arterial highway network in Sioux Falls, South Dakota, is used to test the applicability of the proposed solution to real problems.« less
González-Ramírez, Laura R.; Ahmed, Omar J.; Cash, Sydney S.; Wayne, C. Eugene; Kramer, Mark A.
2015-01-01
Epilepsy—the condition of recurrent, unprovoked seizures—manifests in brain voltage activity with characteristic spatiotemporal patterns. These patterns include stereotyped semi-rhythmic activity produced by aggregate neuronal populations, and organized spatiotemporal phenomena, including waves. To assess these spatiotemporal patterns, we develop a mathematical model consistent with the observed neuronal population activity and determine analytically the parameter configurations that support traveling wave solutions. We then utilize high-density local field potential data recorded in vivo from human cortex preceding seizure termination from three patients to constrain the model parameters, and propose basic mechanisms that contribute to the observed traveling waves. We conclude that a relatively simple and abstract mathematical model consisting of localized interactions between excitatory cells with slow adaptation captures the quantitative features of wave propagation observed in the human local field potential preceding seizure termination. PMID:25689136
Model selection as a science driver for dark energy surveys
NASA Astrophysics Data System (ADS)
Mukherjee, Pia; Parkinson, David; Corasaniti, Pier Stefano; Liddle, Andrew R.; Kunz, Martin
2006-07-01
A key science goal of upcoming dark energy surveys is to seek time-evolution of the dark energy. This problem is one of model selection, where the aim is to differentiate between cosmological models with different numbers of parameters. However, the power of these surveys is traditionally assessed by estimating their ability to constrain parameters, which is a different statistical problem. In this paper, we use Bayesian model selection techniques, specifically forecasting of the Bayes factors, to compare the abilities of different proposed surveys in discovering dark energy evolution. We consider six experiments - supernova luminosity measurements by the Supernova Legacy Survey, SNAP, JEDI and ALPACA, and baryon acoustic oscillation measurements by WFMOS and JEDI - and use Bayes factor plots to compare their statistical constraining power. The concept of Bayes factor forecasting has much broader applicability than dark energy surveys.
A distance constrained synaptic plasticity model of C. elegans neuronal network
NASA Astrophysics Data System (ADS)
Badhwar, Rahul; Bagler, Ganesh
2017-03-01
Brain research has been driven by enquiry for principles of brain structure organization and its control mechanisms. The neuronal wiring map of C. elegans, the only complete connectome available till date, presents an incredible opportunity to learn basic governing principles that drive structure and function of its neuronal architecture. Despite its apparently simple nervous system, C. elegans is known to possess complex functions. The nervous system forms an important underlying framework which specifies phenotypic features associated to sensation, movement, conditioning and memory. In this study, with the help of graph theoretical models, we investigated the C. elegans neuronal network to identify network features that are critical for its control. The 'driver neurons' are associated with important biological functions such as reproduction, signalling processes and anatomical structural development. We created 1D and 2D network models of C. elegans neuronal system to probe the role of features that confer controllability and small world nature. The simple 1D ring model is critically poised for the number of feed forward motifs, neuronal clustering and characteristic path-length in response to synaptic rewiring, indicating optimal rewiring. Using empirically observed distance constraint in the neuronal network as a guiding principle, we created a distance constrained synaptic plasticity model that simultaneously explains small world nature, saturation of feed forward motifs as well as observed number of driver neurons. The distance constrained model suggests optimum long distance synaptic connections as a key feature specifying control of the network.
The Disk of 48 Lib Revealed by NPOI
NASA Astrophysics Data System (ADS)
Lembryk, Ludwik; Tycner, C.; Sigut, A.; Zavala, R. T.
2013-01-01
We present a study of the disk around the Be star 48 Lib, where NLTE numerical disk models are being compared to the spectral and interferometric data to constrain the physical properties of the inner disk structure. The computational models are generated using the BEDISK code, which accounts for heating and cooling of various atoms in the disk and assumes solar chemical composition. A large set of self-consistent disk models produced with the BEDISK code is in turn used to generate synthetic spectra and images assuming a wide range of inclination angles using the BERAY code. The aim of this project is to constrain the physical properties as well as the inclination angles using both spectroscopic and interferometric data. The interferometric data were obtained using the Naval Precision Optical Interferometer (NPOI), with the focus on Hydrogen Balmer-alpha emission, which is the strongest emission line present due to the circumstellar structure. Because 48 Lib shows clear asymmetric spectral lines, we discuss how we model the asymmetric peaks of the Halpha line by combining two models computed with different density structures. The corresponding synthetic images of these combined density structures are then Fourier transformed and compared to the interferometric data. This numerical strategy has the potential to easily model the commonly observed variation of the ratio of the violet-to-red (V/R ratio) emission peaks and constrain the long-term variability associated with the disk of 48 Lib as well as other emission-line stars that show similar variability.
THE NORTH AMERICAN MERCURY MODEL INTER-COMPARISON STUDY (NAMMIS)
This paper describes the North American Mercury Model Inter-comparison Study (NAMMIS). The NAMMIS is an effort to apply atmospheric Hg models in a tightly constrained testing environment with a focus on North America. With each model using the same input data sets for initial co...
Measurement of Psychological Disorders Using Cognitive Diagnosis Models
ERIC Educational Resources Information Center
Templin, Jonathan L.; Henson, Robert A.
2006-01-01
Cognitive diagnosis models are constrained (multiple classification) latent class models that characterize the relationship of questionnaire responses to a set of dichotomous latent variables. Having emanated from educational measurement, several aspects of such models seem well suited to use in psychological assessment and diagnosis. This article…
NASA Astrophysics Data System (ADS)
Niri, Mohammad Emami; Lumley, David E.
2017-10-01
Integration of 3D and time-lapse 4D seismic data into reservoir modelling and history matching processes poses a significant challenge due to the frequent mismatch between the initial reservoir model, the true reservoir geology, and the pre-production (baseline) seismic data. A fundamental step of a reservoir characterisation and performance study is the preconditioning of the initial reservoir model to equally honour both the geological knowledge and seismic data. In this paper we analyse the issues that have a significant impact on the (mis)match of the initial reservoir model with well logs and inverted 3D seismic data. These issues include the constraining methods for reservoir lithofacies modelling, the sensitivity of the results to the presence of realistic resolution and noise in the seismic data, the geostatistical modelling parameters, and the uncertainties associated with quantitative incorporation of inverted seismic data in reservoir lithofacies modelling. We demonstrate that in a geostatistical lithofacies simulation process, seismic constraining methods based on seismic litho-probability curves and seismic litho-probability cubes yield the best match to the reference model, even when realistic resolution and noise is included in the dataset. In addition, our analyses show that quantitative incorporation of inverted 3D seismic data in static reservoir modelling carries a range of uncertainties and should be cautiously applied in order to minimise the risk of misinterpretation. These uncertainties are due to the limited vertical resolution of the seismic data compared to the scale of the geological heterogeneities, the fundamental instability of the inverse problem, and the non-unique elastic properties of different lithofacies types.
Audebert, M; Oxarango, L; Duquennoi, C; Touze-Foltz, N; Forquet, N; Clément, R
2016-09-01
Leachate recirculation is a key process in the operation of municipal solid waste landfills as bioreactors. To ensure optimal water content distribution, bioreactor operators need tools to design leachate injection systems. Prediction of leachate flow by subsurface flow modelling could provide useful information for the design of such systems. However, hydrodynamic models require additional data to constrain them and to assess hydrodynamic parameters. Electrical resistivity tomography (ERT) is a suitable method to study leachate infiltration at the landfill scale. It can provide spatially distributed information which is useful for constraining hydrodynamic models. However, this geophysical method does not allow ERT users to directly measure water content in waste. The MICS (multiple inversions and clustering strategy) methodology was proposed to delineate the infiltration area precisely during time-lapse ERT survey in order to avoid the use of empirical petrophysical relationships, which are not adapted to a heterogeneous medium such as waste. The infiltration shapes and hydrodynamic information extracted with MICS were used to constrain hydrodynamic models in assessing parameters. The constraint methodology developed in this paper was tested on two hydrodynamic models: an equilibrium model where, flow within the waste medium is estimated using a single continuum approach and a non-equilibrium model where flow is estimated using a dual continuum approach. The latter represents leachate flows into fractures. Finally, this methodology provides insight to identify the advantages and limitations of hydrodynamic models. Furthermore, we suggest an explanation for the large volume detected by MICS when a small volume of leachate is injected. Copyright © 2016 Elsevier Ltd. All rights reserved.
Reddington, C. L.; Carslaw, K. S.; Stier, P.; ...
2017-09-01
The largest uncertainty in the historical radiative forcing of climate is caused by changes in aerosol particles due to anthropogenic activity. Sophisticated aerosol microphysics processes have been included in many climate models in an effort to reduce the uncertainty. However, the models are very challenging to evaluate and constrain because they require extensive in situ measurements of the particle size distribution, number concentration, and chemical composition that are not available from global satellite observations. The Global Aerosol Synthesis and Science Project (GASSP) aims to improve the robustness of global aerosol models by combining new methodologies for quantifying model uncertainty, tomore » create an extensive global dataset of aerosol in situ microphysical and chemical measurements, and to develop new ways to assess the uncertainty associated with comparing sparse point measurements with low-resolution models. GASSP has assembled over 45,000 hours of measurements from ships and aircraft as well as data from over 350 ground stations. The measurements have been harmonized into a standardized format that is easily used by modelers and nonspecialist users. Available measurements are extensive, but they are biased to polluted regions of the Northern Hemisphere, leaving large pristine regions and many continental areas poorly sampled. The aerosol radiative forcing uncertainty can be reduced using a rigorous model–data synthesis approach. Nevertheless, our research highlights significant remaining challenges because of the difficulty of constraining many interwoven model uncertainties simultaneously. Although the physical realism of global aerosol models still needs to be improved, the uncertainty in aerosol radiative forcing will be reduced most effectively by systematically and rigorously constraining the models using extensive syntheses of measurements.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Reddington, C. L.; Carslaw, K. S.; Stier, P.
The largest uncertainty in the historical radiative forcing of climate is caused by changes in aerosol particles due to anthropogenic activity. Sophisticated aerosol microphysics processes have been included in many climate models in an effort to reduce the uncertainty. However, the models are very challenging to evaluate and constrain because they require extensive in situ measurements of the particle size distribution, number concentration, and chemical composition that are not available from global satellite observations. The Global Aerosol Synthesis and Science Project (GASSP) aims to improve the robustness of global aerosol models by combining new methodologies for quantifying model uncertainty, tomore » create an extensive global dataset of aerosol in situ microphysical and chemical measurements, and to develop new ways to assess the uncertainty associated with comparing sparse point measurements with low-resolution models. GASSP has assembled over 45,000 hours of measurements from ships and aircraft as well as data from over 350 ground stations. The measurements have been harmonized into a standardized format that is easily used by modelers and nonspecialist users. Available measurements are extensive, but they are biased to polluted regions of the Northern Hemisphere, leaving large pristine regions and many continental areas poorly sampled. The aerosol radiative forcing uncertainty can be reduced using a rigorous model–data synthesis approach. Nevertheless, our research highlights significant remaining challenges because of the difficulty of constraining many interwoven model uncertainties simultaneously. Although the physical realism of global aerosol models still needs to be improved, the uncertainty in aerosol radiative forcing will be reduced most effectively by systematically and rigorously constraining the models using extensive syntheses of measurements.« less
Evaluating data worth for ground-water management under uncertainty
Wagner, B.J.
1999-01-01
A decision framework is presented for assessing the value of ground-water sampling within the context of ground-water management under uncertainty. The framework couples two optimization models-a chance-constrained ground-water management model and an integer-programing sampling network design model-to identify optimal pumping and sampling strategies. The methodology consists of four steps: (1) The optimal ground-water management strategy for the present level of model uncertainty is determined using the chance-constrained management model; (2) for a specified data collection budget, the monitoring network design model identifies, prior to data collection, the sampling strategy that will minimize model uncertainty; (3) the optimal ground-water management strategy is recalculated on the basis of the projected model uncertainty after sampling; and (4) the worth of the monitoring strategy is assessed by comparing the value of the sample information-i.e., the projected reduction in management costs-with the cost of data collection. Steps 2-4 are repeated for a series of data collection budgets, producing a suite of management/monitoring alternatives, from which the best alternative can be selected. A hypothetical example demonstrates the methodology's ability to identify the ground-water sampling strategy with greatest net economic benefit for ground-water management.A decision framework is presented for assessing the value of ground-water sampling within the context of ground-water management under uncertainty. The framework couples two optimization models - a chance-constrained ground-water management model and an integer-programming sampling network design model - to identify optimal pumping and sampling strategies. The methodology consists of four steps: (1) The optimal ground-water management strategy for the present level of model uncertainty is determined using the chance-constrained management model; (2) for a specified data collection budget, the monitoring network design model identifies, prior to data collection, the sampling strategy that will minimize model uncertainty; (3) the optimal ground-water management strategy is recalculated on the basis of the projected model uncertainty after sampling; and (4) the worth of the monitoring strategy is assessed by comparing the value of the sample information - i.e., the projected reduction in management costs - with the cost of data collection. Steps 2-4 are repeated for a series of data collection budgets, producing a suite of management/monitoring alternatives, from which the best alternative can be selected. A hypothetical example demonstrates the methodology's ability to identify the ground-water sampling strategy with greatest net economic benefit for ground-water management.
Constrained Local UniversE Simulations: a Local Group factory
NASA Astrophysics Data System (ADS)
Carlesi, Edoardo; Sorce, Jenny G.; Hoffman, Yehuda; Gottlöber, Stefan; Yepes, Gustavo; Libeskind, Noam I.; Pilipenko, Sergey V.; Knebe, Alexander; Courtois, Hélène; Tully, R. Brent; Steinmetz, Matthias
2016-05-01
Near-field cosmology is practised by studying the Local Group (LG) and its neighbourhood. This paper describes a framework for simulating the `near field' on the computer. Assuming the Λ cold dark matter (ΛCDM) model as a prior and applying the Bayesian tools of the Wiener filter and constrained realizations of Gaussian fields to the Cosmicflows-2 (CF2) survey of peculiar velocities, constrained simulations of our cosmic environment are performed. The aim of these simulations is to reproduce the LG and its local environment. Our main result is that the LG is likely a robust outcome of the ΛCDMscenario when subjected to the constraint derived from CF2 data, emerging in an environment akin to the observed one. Three levels of criteria are used to define the simulated LGs. At the base level, pairs of haloes must obey specific isolation, mass and separation criteria. At the second level, the orbital angular momentum and energy are constrained, and on the third one the phase of the orbit is constrained. Out of the 300 constrained simulations, 146 LGs obey the first set of criteria, 51 the second and 6 the third. The robustness of our LG `factory' enables the construction of a large ensemble of simulated LGs. Suitable candidates for high-resolution hydrodynamical simulations of the LG can be drawn from this ensemble, which can be used to perform comprehensive studies of the formation of the LG.
Constrained State Estimation for Individual Localization in Wireless Body Sensor Networks
Feng, Xiaoxue; Snoussi, Hichem; Liang, Yan; Jiao, Lianmeng
2014-01-01
Wireless body sensor networks based on ultra-wideband radio have recently received much research attention due to its wide applications in health-care, security, sports and entertainment. Accurate localization is a fundamental problem to realize the development of effective location-aware applications above. In this paper the problem of constrained state estimation for individual localization in wireless body sensor networks is addressed. Priori knowledge about geometry among the on-body nodes as additional constraint is incorporated into the traditional filtering system. The analytical expression of state estimation with linear constraint to exploit the additional information is derived. Furthermore, for nonlinear constraint, first-order and second-order linearizations via Taylor series expansion are proposed to transform the nonlinear constraint to the linear case. Examples between the first-order and second-order nonlinear constrained filters based on interacting multiple model extended kalman filter (IMM-EKF) show that the second-order solution for higher order nonlinearity as present in this paper outperforms the first-order solution, and constrained IMM-EKF obtains superior estimation than IMM-EKF without constraint. Another brownian motion individual localization example also illustrates the effectiveness of constrained nonlinear iterative least square (NILS), which gets better filtering performance than NILS without constraint. PMID:25390408
Double quick, double click reversible peptide "stapling".
Grison, Claire M; Burslem, George M; Miles, Jennifer A; Pilsl, Ludwig K A; Yeo, David J; Imani, Zeynab; Warriner, Stuart L; Webb, Michael E; Wilson, Andrew J
2017-07-01
The development of constrained peptides for inhibition of protein-protein interactions is an emerging strategy in chemical biology and drug discovery. This manuscript introduces a versatile, rapid and reversible approach to constrain peptides in a bioactive helical conformation using BID and RNase S peptides as models. Dibromomaleimide is used to constrain BID and RNase S peptide sequence variants bearing cysteine (Cys) or homocysteine ( h Cys) amino acids spaced at i and i + 4 positions by double substitution. The constraint can be readily removed by displacement of the maleimide using excess thiol. This new constraining methodology results in enhanced α-helical conformation (BID and RNase S peptide) as demonstrated by circular dichroism and molecular dynamics simulations, resistance to proteolysis (BID) as demonstrated by trypsin proteolysis experiments and retained or enhanced potency of inhibition for Bcl-2 family protein-protein interactions (BID), or greater capability to restore the hydrolytic activity of the RNAse S protein (RNase S peptide). Finally, use of a dibromomaleimide functionalized with an alkyne permits further divergent functionalization through alkyne-azide cycloaddition chemistry on the constrained peptide with fluorescein, oligoethylene glycol or biotin groups to facilitate biophysical and cellular analyses. Hence this methodology may extend the scope and accessibility of peptide stapling.
Gaining insight into the T _2^*-T2 relationship in surface NMR free-induction decay measurements
NASA Astrophysics Data System (ADS)
Grombacher, Denys; Auken, Esben
2018-05-01
One of the primary shortcomings of the surface nuclear magnetic resonance (NMR) free-induction decay (FID) measurement is the uncertainty surrounding which mechanism controls the signal's time dependence. Ideally, the FID-estimated relaxation time T_2^* that describes the signal's decay carries an intimate link to the geometry of the pore space. In this limit the parameter T_2^* is closely linked to a related parameter T2, which is more closely linked to pore-geometry. If T_2^* ˜eq {T_2} the FID can provide valuable insight into relative pore-size and can be used to make quantitative permeability estimates. However, given only FID measurements it is difficult to determine whether T_2^* is linked to pore geometry or whether it has been strongly influenced by background magnetic field inhomogeneity. If the link between an observed T_2^* and the underlying T2 could be further constrained the utility of the standard surface NMR FID measurement would be greatly improved. We hypothesize that an approach employing an updated surface NMR forward model that solves the full Bloch equations with appropriately weighted relaxation terms can be used to help constrain the T_2^*-T2 relationship. Weighting the relaxation terms requires estimating the poorly constrained parameters T2 and T1; to deal with this uncertainty we propose to conduct a parameter search involving multiple inversions that employ a suite of forward models each describing a distinct but plausible T_2^*-T2 relationship. We hypothesize that forward models given poor T2 estimates will produce poor data fits when using the complex-inversion, while forward models given reliable T2 estimates will produce satisfactory data fits. By examining the data fits produced by the suite of plausible forward models, the likely T_2^*-T2 can be constrained by identifying the range of T2 estimates that produce reliable data fits. Synthetic and field results are presented to investigate the feasibility of the proposed technique.